code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (KDPMaDiscreteScheduler,) lowercase = 10 def _lowerCamelCase ( self : Tuple , **a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = { 'num_train_timesteps': 1_100, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', } config.update(**a ) return config def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : str ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_0_1, 0.0_0_0_1, 0.0_0_1] , [0.0_0_0_2, 0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : Any = scheduler_class(**a ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : Tuple = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase__ : List[Any] = sample.to(a ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase__ : Tuple = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : int = model(a , a ) lowerCAmelCase__ : Union[str, Any] = scheduler.step(a , a , a ) lowerCAmelCase__ : Optional[int] = output.prev_sample lowerCAmelCase__ : Any = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : int = torch.mean(torch.abs(a ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6934E-07 ) < 1E-2 assert abs(result_mean.item() - 6.1112E-10 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.693428650170972E-07 ) < 1E-2 assert abs(result_mean.item() - 0.0_0_0_2 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' if torch_device == "mps": return lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : str = scheduler_class(**a ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase__ : List[Any] = self.dummy_model() lowerCAmelCase__ : List[Any] = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase__ : List[Any] = sample.to(a ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase__ : List[Any] = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : Optional[int] = model(a , a ) lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a ) lowerCAmelCase__ : Optional[int] = output.prev_sample lowerCAmelCase__ : Tuple = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : str = torch.mean(torch.abs(a ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 2_0.4_1_2_5 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 2_0.4_1_2_5 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3 def _lowerCamelCase ( self : Tuple ): '''simple docstring''' if torch_device == "mps": return lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Tuple = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) scheduler.set_timesteps(self.num_inference_steps , device=a ) lowerCAmelCase__ : List[Any] = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter.to(a ) * scheduler.init_noise_sigma for t in scheduler.timesteps: lowerCAmelCase__ : List[Any] = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : str = model(a , a ) lowerCAmelCase__ : Dict = scheduler.step(a , a , a ) lowerCAmelCase__ : str = output.prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[int] = torch.mean(torch.abs(a ) ) if str(a ).startswith('cpu' ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 2_0.4_1_2_5 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 2_0.4_1_2_5 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3
69
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """google/rembert""": 256, } lowerCamelCase__ = """▁""" class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = RemBertTokenizer def __init__( self : Optional[Any] , a : str=None , a : Any=None , a : List[Any]=True , a : str=True , a : Dict=False , a : Dict="[CLS]" , a : int="[SEP]" , a : Tuple="<unk>" , a : Optional[Any]="[SEP]" , a : Tuple="<pad>" , a : Dict="[CLS]" , a : Optional[Any]="[MASK]" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( a , tokenizer_file=a , do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : int = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : List[Any] = keep_accents lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
69
1
import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = mock.Mock() lowerCAmelCase__ : List[str] = 500 lowerCAmelCase__ : List[str] = {} lowerCAmelCase__ : Union[str, Any] = HTTPError lowerCAmelCase__ : Any = {} # Download this model to make sure it's in the cache. lowerCAmelCase__ : int = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=a ) as mock_head: lowerCAmelCase__ : Any = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = mock.Mock() lowerCAmelCase__ : Dict = 500 lowerCAmelCase__ : str = {} lowerCAmelCase__ : Tuple = HTTPError lowerCAmelCase__ : Optional[int] = {} # Download this model to make sure it's in the cache. lowerCAmelCase__ : int = GPTaTokenizerFast.from_pretrained('gpt2' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=a ) as mock_head: lowerCAmelCase__ : Union[str, Any] = GPTaTokenizerFast.from_pretrained('gpt2' ) # This check we did call the fake head request mock_head.assert_called() def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' try: lowerCAmelCase__ : Dict = tempfile.mktemp() with open(a , 'wb' ) as f: http_get('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' , a ) lowerCAmelCase__ : Any = AlbertTokenizer.from_pretrained(a ) finally: os.remove(a ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile('tokenizer.json' ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open('tokenizer.json' , 'wb' ) as f: http_get('https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json' , a ) lowerCAmelCase__ : int = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1_000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove('tokenizer.json' ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AlbertTokenizer.from_pretrained('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' ) @is_staging_test class A__ ( unittest.TestCase ): lowercase = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] @classmethod def _lowerCamelCase ( cls : str ): '''simple docstring''' lowerCAmelCase__ : str = TOKEN HfFolder.save_token(a ) @classmethod def _lowerCamelCase ( cls : List[str] ): '''simple docstring''' try: delete_repo(token=cls._token , repo_id='test-tokenizer' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-tokenizer-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-tokenizer' ) except HTTPError: pass def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : str = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : List[Any] = BertTokenizer(a ) tokenizer.push_to_hub('test-tokenizer' , use_auth_token=self._token ) lowerCAmelCase__ : Optional[Any] = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='test-tokenizer' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(a , repo_id='test-tokenizer' , push_to_hub=a , use_auth_token=self._token ) lowerCAmelCase__ : List[Any] = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def _lowerCamelCase ( self : int ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : Union[str, Any] = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : Optional[Any] = BertTokenizer(a ) tokenizer.push_to_hub('valid_org/test-tokenizer-org' , use_auth_token=self._token ) lowerCAmelCase__ : Union[str, Any] = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-tokenizer-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( a , repo_id='valid_org/test-tokenizer-org' , push_to_hub=a , use_auth_token=self._token ) lowerCAmelCase__ : Dict = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : Dict = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : List[str] = CustomTokenizer(a ) # No fast custom tokenizer tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) lowerCAmelCase__ : Any = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : int = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : str = BertTokenizerFast.from_pretrained(a ) bert_tokenizer.save_pretrained(a ) lowerCAmelCase__ : Union[str, Any] = CustomTokenizerFast.from_pretrained(a ) tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) lowerCAmelCase__ : Optional[int] = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizerFast' ) lowerCAmelCase__ : List[Any] = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=a , trust_remote_code=a ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = Trie() trie.add('Hello 友達' ) self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {' ': {'友': {'達': {'': 1}}}}}}}}} ) trie.add('Hello' ) trie.data self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {'': 1, ' ': {'友': {'達': {'': 1}}}}}}}}} ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = Trie() self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS] This is a extra_id_100'] ) trie.add('[CLS]' ) trie.add('extra_id_1' ) trie.add('extra_id_100' ) self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS]', ' This is a ', 'extra_id_100'] ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = Trie() trie.add('A' ) self.assertEqual(trie.split('ABC' ) , ['A', 'BC'] ) self.assertEqual(trie.split('BCA' ) , ['BC', 'A'] ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = Trie() trie.add('TOKEN]' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = Trie() trie.add('A' ) trie.add('P' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = Trie() trie.add('AB' ) trie.add('B' ) trie.add('C' ) self.assertEqual(trie.split('ABC' ) , ['AB', 'C'] ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = Trie() trie.add('ABC' ) trie.add('B' ) trie.add('CD' ) self.assertEqual(trie.split('ABCD' ) , ['ABC', 'D'] ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = Trie() lowerCAmelCase__ : str = trie.cut_text('ABC' , [0, 0, 2, 1, 2, 3] ) self.assertEqual(a , ['AB', 'C'] )
69
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : int = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : List[Any] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : int = None lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Dict = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Union[str, Any] = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
69
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_table_transformer""": [ """TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TableTransformerConfig""", """TableTransformerOnnxConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TableTransformerForObjectDetection""", """TableTransformerModel""", """TableTransformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
1
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError('The given input must be positive' ) # get the generated string sequence lowerCAmelCase__ : Any = gray_code_sequence_string(SCREAMING_SNAKE_CASE_ ) # # convert them to integers for i in range(len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ : Any = int(sequence[i] , 2 ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] lowerCAmelCase__ : Optional[Any] = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits lowerCAmelCase__ : Optional[Any] = gray_code_sequence_string(bit_count - 1 ) lowerCAmelCase__ : Union[str, Any] = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): lowerCAmelCase__ : Optional[Any] = '0' + smaller_sequence[i] sequence.append(SCREAMING_SNAKE_CASE_ ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): lowerCAmelCase__ : int = '1' + smaller_sequence[i] sequence.append(SCREAMING_SNAKE_CASE_ ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
69
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
1
import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowerCamelCase__ = get_tests_dir("""fixtures/test_sentencepiece_with_bytefallback.model""") @require_sentencepiece @require_tokenizers class A__ ( __magic_name__ , unittest.TestCase ): lowercase = GPTSwaTokenizer lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Dict ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase__ : List[str] = GPTSwaTokenizer(a , eos_token='<unk>' , bos_token='<unk>' , pad_token='<unk>' ) tokenizer.save_pretrained(self.tmpdirname ) def _lowerCamelCase ( self : List[Any] , a : str ): '''simple docstring''' lowerCAmelCase__ : Dict = 'This is a test' lowerCAmelCase__ : Optional[int] = 'This is a test' return input_text, output_text def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : List[str] = '<s>' lowerCAmelCase__ : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a ) , a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a ) , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , 'j' ) self.assertEqual(len(a ) , 2_000 ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = GPTSwaTokenizer(a ) lowerCAmelCase__ : int = tokenizer.tokenize('This is a test' ) self.assertListEqual(a , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , [465, 287, 265, 631, 842] ) lowerCAmelCase__ : Tuple = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) # fmt: off self.assertListEqual( a , ['▁I', '▁was', '▁bor', 'n', '▁in', '▁', '<0x39>', '2', '0', '0', '0', ',', '▁and', '▁this', '▁is', '▁f', 'al', 's', '<0xC3>', '<0xA9>', '.'] , ) # fmt: on lowerCAmelCase__ : Union[str, Any] = tokenizer.convert_tokens_to_ids(a ) self.assertListEqual( a , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) lowerCAmelCase__ : Dict = tokenizer.convert_ids_to_tokens(a ) # fmt: off self.assertListEqual( a , ['▁I', '▁was', '▁bor', 'n', '▁in', '▁', '<0x39>', '2', '0', '0', '0', ',', '▁and', '▁this', '▁is', '▁f', 'al', 's', '<0xC3>', '<0xA9>', '.'] ) # fmt: on def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : int = GPTSwaTokenizer(a ) lowerCAmelCase__ : Any = ['This is a test', 'I was born in 92000, and this is falsé.'] lowerCAmelCase__ : Optional[int] = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a , a ): self.assertListEqual(tokenizer.encode_fast(a ) , a ) # Test that decode_fast returns the input text for text, token_ids in zip(a , a ): self.assertEqual(tokenizer.decode_fast(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : int = [ '<|python|>def fibonacci(n)\n if n < 0:\n print(\'Incorrect input\')', 'Hey there, how are you doing this fine day?', 'This is a text with a trailing spaces followed by a dot .', 'Häj sväjs lillebrör! =)', 'Det är inget fel på Mr. Cool', ] # fmt: off lowerCAmelCase__ : List[Any] = {'input_ids': [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a , model_name='AI-Sweden/gpt-sw3-126m' , sequences=a , )
69
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
1
import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __magic_name__ , unittest.TestCase ): lowercase = GPTaTokenizer lowercase = GPTaTokenizerFast lowercase = True lowercase = {'add_prefix_space': True} lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : Optional[Any] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', '<|endoftext|>', ] lowerCAmelCase__ : str = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Tuple = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] lowerCAmelCase__ : Optional[int] = {'unk_token': '<unk>'} lowerCAmelCase__ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(a ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : Any , **a : int ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **a ) def _lowerCamelCase ( self : int , **a : str ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **a ) def _lowerCamelCase ( self : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = 'lower newer' lowerCAmelCase__ : Union[str, Any] = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : List[Any] = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) lowerCAmelCase__ : Union[str, Any] = 'lower newer' lowerCAmelCase__ : str = ['\u0120low', 'er', '\u0120', 'n', 'e', 'w', 'er'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a , add_prefix_space=a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Union[str, Any] = tokens + [tokenizer.unk_token] lowerCAmelCase__ : List[str] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' if not self.test_rust_tokenizer: return lowerCAmelCase__ : Any = self.get_tokenizer() lowerCAmelCase__ : List[Any] = self.get_rust_tokenizer(add_prefix_space=a ) lowerCAmelCase__ : List[str] = 'lower newer' # Testing tokenization lowerCAmelCase__ : Union[str, Any] = tokenizer.tokenize(a , add_prefix_space=a ) lowerCAmelCase__ : Optional[Any] = rust_tokenizer.tokenize(a ) self.assertListEqual(a , a ) # Testing conversion to ids without special tokens lowerCAmelCase__ : Tuple = tokenizer.encode(a , add_special_tokens=a , add_prefix_space=a ) lowerCAmelCase__ : Union[str, Any] = rust_tokenizer.encode(a , add_special_tokens=a ) self.assertListEqual(a , a ) # Testing conversion to ids with special tokens lowerCAmelCase__ : List[Any] = self.get_rust_tokenizer(add_prefix_space=a ) lowerCAmelCase__ : Any = tokenizer.encode(a , add_prefix_space=a ) lowerCAmelCase__ : Union[str, Any] = rust_tokenizer.encode(a ) self.assertListEqual(a , a ) # Testing the unknown token lowerCAmelCase__ : Optional[Any] = tokens + [rust_tokenizer.unk_token] lowerCAmelCase__ : Any = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(a ) , a ) def _lowerCamelCase ( self : Any , *a : Optional[int] , **a : Tuple ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple , a : Optional[int]=15 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase__ : int = self.rust_tokenizer_class.from_pretrained(a , **a ) # Simple input lowerCAmelCase__ : str = 'This is a simple input' lowerCAmelCase__ : List[str] = ['This is a simple input 1', 'This is a simple input 2'] lowerCAmelCase__ : int = ('This is a simple input', 'This is a pair') lowerCAmelCase__ : Dict = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(a , tokenizer_r.encode , a , max_length=a , padding='max_length' ) # Simple input self.assertRaises(a , tokenizer_r.encode_plus , a , max_length=a , padding='max_length' ) # Simple input self.assertRaises( a , tokenizer_r.batch_encode_plus , a , max_length=a , padding='max_length' , ) # Pair input self.assertRaises(a , tokenizer_r.encode , a , max_length=a , padding='max_length' ) # Pair input self.assertRaises(a , tokenizer_r.encode_plus , a , max_length=a , padding='max_length' ) # Pair input self.assertRaises( a , tokenizer_r.batch_encode_plus , a , max_length=a , padding='max_length' , ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token='<pad>' ) # Simple input lowerCAmelCase__ : List[Any] = 'This is a simple input' lowerCAmelCase__ : int = ['This is a simple input looooooooong', 'This is a simple input'] lowerCAmelCase__ : str = ('This is a simple input', 'This is a pair') lowerCAmelCase__ : Any = [ ('This is a simple input loooooong', 'This is a simple input'), ('This is a simple pair loooooong', 'This is a simple pair'), ] lowerCAmelCase__ : List[str] = tokenizer.pad_token_id lowerCAmelCase__ : List[Any] = tokenizer(a , padding='max_length' , max_length=30 , return_tensors='np' ) lowerCAmelCase__ : List[str] = tokenizer(a , padding=a , truncate=a , return_tensors='np' ) lowerCAmelCase__ : Optional[int] = tokenizer(*a , padding='max_length' , max_length=60 , return_tensors='np' ) lowerCAmelCase__ : int = tokenizer(a , padding=a , truncate=a , return_tensors='np' ) # s # test single string max_length padding self.assertEqual(out_s['input_ids'].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['input_ids'] ) self.assertTrue(0 in out_s['attention_mask'] ) # s2 # test automatic padding self.assertEqual(out_sa['input_ids'].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['input_ids'][0] ) self.assertFalse(0 in out_sa['attention_mask'][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['input_ids'][1] ) self.assertTrue(0 in out_sa['attention_mask'][1] ) # p # test single pair max_length padding self.assertEqual(out_p['input_ids'].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['input_ids'] ) self.assertTrue(0 in out_p['attention_mask'] ) # p2 # test automatic padding pair self.assertEqual(out_pa['input_ids'].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['input_ids'][0] ) self.assertFalse(0 in out_pa['attention_mask'][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['input_ids'][1] ) self.assertTrue(0 in out_pa['attention_mask'][1] ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : List[str] = '$$$' lowerCAmelCase__ : int = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=a , add_bos_token=a ) lowerCAmelCase__ : Optional[int] = 'This is a simple input' lowerCAmelCase__ : List[Any] = ['This is a simple input 1', 'This is a simple input 2'] lowerCAmelCase__ : List[str] = tokenizer.bos_token_id lowerCAmelCase__ : Any = tokenizer(a ) lowerCAmelCase__ : Any = tokenizer(a ) self.assertEqual(out_s.input_ids[0] , a ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) lowerCAmelCase__ : int = tokenizer.decode(out_s.input_ids ) lowerCAmelCase__ : int = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , a ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : List[Any] = [self.get_tokenizer(do_lower_case=a , add_bos_token=a )] for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): lowerCAmelCase__ : Dict = 'Encode this.' lowerCAmelCase__ : Optional[Any] = 'This one too please.' lowerCAmelCase__ : Union[str, Any] = tokenizer.encode(a , add_special_tokens=a ) encoded_sequence += tokenizer.encode(a , add_special_tokens=a ) lowerCAmelCase__ : Tuple = tokenizer.encode_plus( a , a , add_special_tokens=a , return_special_tokens_mask=a , ) lowerCAmelCase__ : List[Any] = encoded_sequence_dict['input_ids'] lowerCAmelCase__ : Union[str, Any] = encoded_sequence_dict['special_tokens_mask'] self.assertEqual(len(a ) , len(a ) ) lowerCAmelCase__ : Optional[Any] = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(a ) ] lowerCAmelCase__ : Union[str, Any] = [x for x in filtered_sequence if x is not None] self.assertEqual(a , a ) @require_tokenizers class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = AutoTokenizer.from_pretrained('facebook/opt-350m' , from_slow=a ) lowerCAmelCase__ : str = 'A photo of a cat' lowerCAmelCase__ : Dict = tokenizer.encode( a , ) self.assertEqual(a , [2, 250, 1_345, 9, 10, 4_758] ) tokenizer.save_pretrained('test_opt' ) lowerCAmelCase__ : Optional[Any] = AutoTokenizer.from_pretrained('./test_opt' ) lowerCAmelCase__ : str = tokenizer.encode( a , ) self.assertEqual(a , [2, 250, 1_345, 9, 10, 4_758] ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = AutoTokenizer.from_pretrained('facebook/opt-350m' , use_slow=a ) lowerCAmelCase__ : Optional[int] = 'A photo of a cat' lowerCAmelCase__ : Dict = tokenizer.encode( a , ) # Same as above self.assertEqual(a , [2, 250, 1_345, 9, 10, 4_758] ) @unittest.skip('This test is failing because of a bug in the fast tokenizer' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = AutoTokenizer.from_pretrained('facebook/opt-350m' , from_slow=a ) lowerCAmelCase__ : Tuple = 'bos' lowerCAmelCase__ : Optional[int] = tokenizer.get_vocab()['bos'] lowerCAmelCase__ : List[Any] = 'A photo of a cat' lowerCAmelCase__ : Optional[Any] = tokenizer.encode( a , ) # We changed the bos token self.assertEqual(a , [31_957, 250, 1_345, 9, 10, 4_758] ) tokenizer.save_pretrained('./tok' ) lowerCAmelCase__ : Optional[int] = AutoTokenizer.from_pretrained('./tok' ) self.assertTrue(tokenizer.is_fast ) lowerCAmelCase__ : Optional[Any] = tokenizer.encode( a , ) self.assertEqual(a , [31_957, 250, 1_345, 9, 10, 4_758] )
69
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase__ = {"""configuration_xglm""": ["""XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XGLMConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""XGLMTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""XGLMTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """XGLM_PRETRAINED_MODEL_ARCHIVE_LIST""", """XGLMForCausalLM""", """XGLMModel""", """XGLMPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """FlaxXGLMForCausalLM""", """FlaxXGLMModel""", """FlaxXGLMPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFXGLMForCausalLM""", """TFXGLMModel""", """TFXGLMPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
69
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
1
import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class A__ ( __magic_name__ ): lowercase = '' lowercase = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) lowercase = None # compression type in fsspec. ex: "gzip" lowercase = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self : Optional[Any] , a : str = "" , a : Optional[str] = None , a : Optional[dict] = None , **a : Any ): '''simple docstring''' super().__init__(self , **a ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode lowerCAmelCase__ : Tuple = fsspec.open( a , mode='rb' , protocol=a , compression=self.compression , client_kwargs={ 'requote_redirect_url': False, # see https://github.com/huggingface/datasets/pull/5459 'trust_env': True, # Enable reading proxy env variables. **(target_options or {}).pop('client_kwargs' , {} ), # To avoid issues if it was already passed. } , **(target_options or {}) , ) lowerCAmelCase__ : Optional[int] = os.path.basename(self.file.path.split('::' )[0] ) lowerCAmelCase__ : Optional[int] = ( self.compressed_name[: self.compressed_name.rindex('.' )] if '.' in self.compressed_name else self.compressed_name ) lowerCAmelCase__ : Optional[int] = None @classmethod def _lowerCamelCase ( cls : Dict , a : Optional[int] ): '''simple docstring''' return super()._strip_protocol(a ).lstrip('/' ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' if self.dir_cache is None: lowerCAmelCase__ : List[str] = {**self.file.fs.info(self.file.path ), 'name': self.uncompressed_name} lowerCAmelCase__ : Optional[int] = {f['name']: f} def _lowerCamelCase ( self : List[str] , a : str ): '''simple docstring''' return self.file.open().read() def _lowerCamelCase ( self : Optional[int] , a : str , a : str = "rb" , a : Any=None , a : str=True , a : List[Any]=None , **a : List[Any] , ): '''simple docstring''' lowerCAmelCase__ : List[str] = self._strip_protocol(a ) if mode != "rb": raise ValueError(f'''Tried to read with mode {mode} on file {self.file.path} opened with mode \'rb\'''' ) return self.file.open() class A__ ( __magic_name__ ): lowercase = 'bz2' lowercase = 'bz2' lowercase = '.bz2' class A__ ( __magic_name__ ): lowercase = 'gzip' lowercase = 'gzip' lowercase = '.gz' class A__ ( __magic_name__ ): lowercase = 'lz4' lowercase = 'lz4' lowercase = '.lz4' class A__ ( __magic_name__ ): lowercase = 'xz' lowercase = 'xz' lowercase = '.xz' class A__ ( __magic_name__ ): lowercase = 'zstd' lowercase = 'zstd' lowercase = '.zst' def __init__( self : Optional[Any] , a : str , a : str = "rb" , a : Optional[str] = None , a : Optional[dict] = None , a : int = DEFAULT_BLOCK_SIZE , **a : Optional[int] , ): '''simple docstring''' super().__init__( fo=a , mode=a , target_protocol=a , target_options=a , block_size=a , **a , ) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 lowerCAmelCase__ : Dict = self.file.__enter__ class A__ : def __init__( self : Dict , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = file_ def __enter__( self : Any ): '''simple docstring''' self._file.__enter__() return self def __exit__( self : List[Any] , *a : Optional[Any] , **a : str ): '''simple docstring''' self._file.__exit__(*a , **a ) def __iter__( self : Union[str, Any] ): '''simple docstring''' return iter(self._file ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return next(self._file ) def __getattr__( self : List[Any] , a : Optional[int] ): '''simple docstring''' return getattr(self._file , a ) def fixed_enter(*a : str , **a : int ): return WrappedFile(_enter(*a , **a ) ) lowerCAmelCase__ : List[Any] = fixed_enter
69
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
1
from __future__ import annotations import math def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: if depth < 0: raise ValueError('Depth cannot be less than 0' ) if not scores: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , minimax(depth + 1 , node_index * 2 + 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , minimax(depth + 1 , node_index * 2 + 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , ) ) def lowerCAmelCase__ ( ) -> None: lowerCAmelCase__ : int = [90, 23, 6, 33, 21, 65, 123, 34_423] lowerCAmelCase__ : str = math.log(len(SCREAMING_SNAKE_CASE_ ) , 2 ) print(F'''Optimal value : {minimax(0 , 0 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
69
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
1
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> bool: lowerCAmelCase__ : List[Any] = len(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : List[Any] = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): lowerCAmelCase__ : int = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): lowerCAmelCase__ : List[str] = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: lowerCAmelCase__ : List[Any] = subset[i - 1][j] if arr[i - 1] <= j: lowerCAmelCase__ : Union[str, Any] = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
69
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
1
import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = OmegaConf.load(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : List[str] = torch.load(SCREAMING_SNAKE_CASE_ , map_location='cpu' )['model'] lowerCAmelCase__ : str = list(state_dict.keys() ) # extract state_dict for VQVAE lowerCAmelCase__ : Dict = {} lowerCAmelCase__ : Any = 'first_stage_model.' for key in keys: if key.startswith(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Dict = state_dict[key] # extract state_dict for UNetLDM lowerCAmelCase__ : str = {} lowerCAmelCase__ : Optional[int] = 'model.diffusion_model.' for key in keys: if key.startswith(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Union[str, Any] = state_dict[key] lowerCAmelCase__ : Tuple = config.model.params.first_stage_config.params lowerCAmelCase__ : str = config.model.params.unet_config.params lowerCAmelCase__ : int = VQModel(**SCREAMING_SNAKE_CASE_ ).eval() vqvae.load_state_dict(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = UNetLDMModel(**SCREAMING_SNAKE_CASE_ ).eval() unet.load_state_dict(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule='scaled_linear' , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=SCREAMING_SNAKE_CASE_ , ) lowerCAmelCase__ : Tuple = LDMPipeline(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) pipeline.save_pretrained(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", type=str, required=True) parser.add_argument("""--config_path""", type=str, required=True) parser.add_argument("""--output_path""", type=str, required=True) lowerCamelCase__ = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
69
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
1
from __future__ import annotations from collections.abc import MutableSequence class A__ : def __init__( self : int , a : int , a : MutableSequence[float] ): '''simple docstring''' if len(a ) != degree + 1: raise ValueError( 'The number of coefficients should be equal to the degree + 1.' ) lowerCAmelCase__ : list[float] = list(a ) lowerCAmelCase__ : Tuple = degree def __add__( self : Optional[int] , a : Polynomial ): '''simple docstring''' if self.degree > polynomial_a.degree: lowerCAmelCase__ : List[str] = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , a ) else: lowerCAmelCase__ : int = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , a ) def __sub__( self : Optional[Any] , a : Polynomial ): '''simple docstring''' return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : List[Any] ): '''simple docstring''' return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Dict , a : Polynomial ): '''simple docstring''' lowerCAmelCase__ : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , a ) def _lowerCamelCase ( self : Optional[int] , a : int | float ): '''simple docstring''' lowerCAmelCase__ : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = '' for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(a ) return polynomial def __repr__( self : str ): '''simple docstring''' return self.__str__() def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : list[float] = [0] * self.degree for i in range(self.degree ): lowerCAmelCase__ : Dict = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , a ) def _lowerCamelCase ( self : Tuple , a : int | float = 0 ): '''simple docstring''' lowerCAmelCase__ : list[float] = [0] * (self.degree + 2) lowerCAmelCase__ : Optional[Any] = constant for i in range(self.degree + 1 ): lowerCAmelCase__ : Tuple = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , a ) def __eq__( self : Optional[Any] , a : object ): '''simple docstring''' if not isinstance(a , a ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Optional[int] , a : object ): '''simple docstring''' return not self.__eq__(a )
69
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
1
from __future__ import annotations from PIL import Image # Define glider example lowerCamelCase__ = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example lowerCamelCase__ = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list[list[int]]: lowerCAmelCase__ : List[Any] = [] for i in range(len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ : Optional[int] = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours lowerCAmelCase__ : Optional[int] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(SCREAMING_SNAKE_CASE_ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(SCREAMING_SNAKE_CASE_ ) - 1: neighbour_count += cells[i + 1][j] if i < len(SCREAMING_SNAKE_CASE_ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. lowerCAmelCase__ : Dict = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(SCREAMING_SNAKE_CASE_ ) return next_generation def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> list[Image.Image]: lowerCAmelCase__ : Optional[int] = [] for _ in range(SCREAMING_SNAKE_CASE_ ): # Create output image lowerCAmelCase__ : Any = Image.new('RGB' , (len(cells[0] ), len(SCREAMING_SNAKE_CASE_ )) ) lowerCAmelCase__ : str = img.load() # Save cells to image for x in range(len(SCREAMING_SNAKE_CASE_ ) ): for y in range(len(cells[0] ) ): lowerCAmelCase__ : Optional[int] = 255 - cells[y][x] * 255 lowerCAmelCase__ : Optional[int] = (colour, colour, colour) # Save image images.append(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[int] = new_generation(SCREAMING_SNAKE_CASE_ ) return images if __name__ == "__main__": lowerCamelCase__ = generate_images(GLIDER, 16) images[0].save("""out.gif""", save_all=True, append_images=images[1:])
69
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
69
1
from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker lowerCamelCase__ = """CompVis/stable-diffusion-v1-1""" lowerCamelCase__ = """CompVis/stable-diffusion-v1-2""" lowerCamelCase__ = """CompVis/stable-diffusion-v1-3""" lowerCamelCase__ = """CompVis/stable-diffusion-v1-4""" class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : AutoencoderKL , a : CLIPTextModel , a : CLIPTokenizer , a : UNetaDConditionModel , a : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , a : StableDiffusionSafetyChecker , a : CLIPImageProcessor , a : bool = True , ): '''simple docstring''' super()._init_() lowerCAmelCase__ : List[Any] = StableDiffusionPipeline.from_pretrained(a ) lowerCAmelCase__ : Dict = StableDiffusionPipeline.from_pretrained(a ) lowerCAmelCase__ : int = StableDiffusionPipeline.from_pretrained(a ) lowerCAmelCase__ : List[Any] = StableDiffusionPipeline( vae=a , text_encoder=a , tokenizer=a , unet=a , scheduler=a , safety_checker=a , feature_extractor=a , requires_safety_checker=a , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea ) @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return {k: getattr(self , a ) for k in self.config.keys() if not k.startswith('_' )} def _lowerCamelCase ( self : List[Any] , a : Optional[Union[str, int]] = "auto" ): '''simple docstring''' if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowerCAmelCase__ : List[str] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' self.enable_attention_slicing(a ) @torch.no_grad() def _lowerCamelCase ( self : List[Any] , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : Tuple , ): '''simple docstring''' return self.pipea( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) @torch.no_grad() def _lowerCamelCase ( self : Tuple , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : int , ): '''simple docstring''' return self.pipea( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) @torch.no_grad() def _lowerCamelCase ( self : List[Any] , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : List[str] , ): '''simple docstring''' return self.pipea( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) @torch.no_grad() def _lowerCamelCase ( self : List[Any] , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : List[str] , ): '''simple docstring''' return self.pipea( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) @torch.no_grad() def _lowerCamelCase ( self : Tuple , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : List[Any] , ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cuda' if torch.cuda.is_available() else 'cpu' self.to(a ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` must be divisible by 8 but are {height} and {width}.''' ) # Get first result from Stable Diffusion Checkpoint v1.1 lowerCAmelCase__ : Optional[Any] = self.textaimg_sda_a( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) # Get first result from Stable Diffusion Checkpoint v1.2 lowerCAmelCase__ : Dict = self.textaimg_sda_a( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) # Get first result from Stable Diffusion Checkpoint v1.3 lowerCAmelCase__ : str = self.textaimg_sda_a( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) # Get first result from Stable Diffusion Checkpoint v1.4 lowerCAmelCase__ : List[str] = self.textaimg_sda_a( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
69
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __magic_name__ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'LayoutLMv3ImageProcessor' lowercase = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self : Optional[int] , a : Union[str, Any]=None , a : Optional[Any]=None , **a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , a , ) lowerCAmelCase__ : int = kwargs.pop('feature_extractor' ) lowerCAmelCase__ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(a , a ) def __call__( self : List[Any] , a : List[Any] , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a : Union[List[List[int]], List[List[List[int]]]] = None , a : Optional[Union[List[int], List[List[int]]]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : str , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor lowerCAmelCase__ : List[str] = self.image_processor(images=a , return_tensors=a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a , a ): lowerCAmelCase__ : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ : List[str] = features['words'] lowerCAmelCase__ : List[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_token_type_ids=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_length=a , verbose=a , return_tensors=a , **a , ) # add pixel values lowerCAmelCase__ : Tuple = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase__ : List[str] = self.get_overflowing_images(a , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase__ : List[str] = images return encoded_inputs def _lowerCamelCase ( self : Any , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a ) != len(a ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' f''' {len(a )} and {len(a )}''' ) return images_with_overflow def _lowerCamelCase ( self : Union[str, Any] , *a : Optional[Any] , **a : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Tuple , *a : List[str] , **a : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*a , **a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , a , ) return self.image_processor_class @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , a , ) return self.image_processor
69
1
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
69
1
from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['input_features', 'attention_mask'] def __init__( self : Dict , a : List[Any]=80 , a : Tuple=16_000 , a : List[str]=80 , a : int=0.0 , a : str=True , a : List[Any]=True , a : Any=True , **a : int , ): '''simple docstring''' super().__init__(feature_size=a , sampling_rate=a , padding_value=a , **a ) lowerCAmelCase__ : Optional[int] = num_mel_bins lowerCAmelCase__ : Union[str, Any] = do_ceptral_normalize lowerCAmelCase__ : int = normalize_means lowerCAmelCase__ : Tuple = normalize_vars lowerCAmelCase__ : Tuple = True def _lowerCamelCase ( self : List[str] , a : np.ndarray , ): '''simple docstring''' lowerCAmelCase__ : str = waveform * (2**15) # Kaldi compliance: 16-bit signed integers lowerCAmelCase__ : Union[str, Any] = torch.from_numpy(a ).unsqueeze(0 ) lowerCAmelCase__ : Optional[int] = ta_kaldi.fbank(a , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def _lowerCamelCase ( a : np.ndarray , a : int , a : Optional[bool] = True , a : Optional[bool] = True , a : float = 0.0 , ): '''simple docstring''' if normalize_means: lowerCAmelCase__ : int = x[:input_length].mean(axis=0 ) lowerCAmelCase__ : Dict = np.subtract(a , a ) if normalize_vars: lowerCAmelCase__ : Dict = x[:input_length].std(axis=0 ) lowerCAmelCase__ : List[Any] = np.divide(a , a ) if input_length < x.shape[0]: lowerCAmelCase__ : Optional[Any] = padding_value # make sure array is in float32 lowerCAmelCase__ : Any = x.astype(np.floataa ) return x def _lowerCamelCase ( self : Dict , a : List[np.ndarray] , a : Optional[np.ndarray] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(a , a , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(a , a ) ] def __call__( self : Any , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Union[bool, str, PaddingStrategy] = False , a : Optional[int] = None , a : bool = False , a : Optional[int] = None , a : Optional[Union[str, TensorType]] = None , a : Optional[int] = None , a : Optional[bool] = None , **a : Any , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : Optional[int] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : Optional[int] = [np.asarray(a , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Dict = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : Tuple = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Any = [raw_speech] # extract fbank features lowerCAmelCase__ : Tuple = [self._extract_fbank_features(a ) for waveform in raw_speech] # convert into correct format for padding lowerCAmelCase__ : Tuple = BatchFeature({'input_features': features} ) lowerCAmelCase__ : Tuple = self.pad( a , padding=a , max_length=a , truncation=a , pad_to_multiple_of=a , return_attention_mask=a , **a , ) # make sure list is in array format lowerCAmelCase__ : Optional[int] = padded_inputs.get('input_features' ) if isinstance(input_features[0] , a ): lowerCAmelCase__ : int = [np.asarray(a , dtype=np.floataa ) for feature in input_features] lowerCAmelCase__ : int = padded_inputs.get('attention_mask' ) if attention_mask is not None: lowerCAmelCase__ : str = [np.asarray(a , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: lowerCAmelCase__ : Tuple = ( np.array(a , dtype=np.intaa ) if self._get_padding_strategies(a , max_length=a ) is not PaddingStrategy.DO_NOT_PAD else None ) lowerCAmelCase__ : int = self.normalize( padded_inputs['input_features'] , attention_mask=a ) if return_tensors is not None: lowerCAmelCase__ : int = padded_inputs.convert_to_tensors(a ) return padded_inputs
69
lowerCamelCase__ = """Alexander Joslin""" import operator as op from .stack import Stack def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} lowerCAmelCase__ : Stack[int] = Stack() lowerCAmelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE_ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE_ ) elif i == ")": # RULE 4 lowerCAmelCase__ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : List[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : Tuple = operators[opr](SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) operand_stack.push(SCREAMING_SNAKE_CASE_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
69
1
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable lowerCamelCase__ = { """configuration_gpt_neox_japanese""": ["""GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoXJapaneseConfig"""], """tokenization_gpt_neox_japanese""": ["""GPTNeoXJapaneseTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTNeoXJapaneseForCausalLM""", """GPTNeoXJapaneseLayer""", """GPTNeoXJapaneseModel""", """GPTNeoXJapanesePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
import numpy class A__ : def __init__( self : Tuple , a : numpy.ndarray , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : int = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Dict = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : List[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowerCAmelCase__ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : List[Any] = numpy.zeros(output_array.shape ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowerCAmelCase__ : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowerCAmelCase__ : int = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowerCamelCase ( self : Optional[int] , a : numpy.ndarray , a : int , a : bool ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): lowerCAmelCase__ : Any = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Tuple = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def _lowerCamelCase ( self : Optional[Any] , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : Dict = input_arr lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return (value) * (1 - (value)) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : List[str] = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ , output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ , iterations=10 , give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
69
1
import math def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: lowerCAmelCase__ : int = [True] * n lowerCAmelCase__ : Optional[int] = False lowerCAmelCase__ : Dict = False lowerCAmelCase__ : List[Any] = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): lowerCAmelCase__ : Optional[Any] = i * 2 while index < n: lowerCAmelCase__ : Union[str, Any] = False lowerCAmelCase__ : Tuple = index + i lowerCAmelCase__ : str = [2] for i in range(3 , SCREAMING_SNAKE_CASE_ , 2 ): if is_prime[i]: primes.append(SCREAMING_SNAKE_CASE_ ) return primes def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 999_966_663_333 ) -> int: lowerCAmelCase__ : Dict = math.floor(math.sqrt(SCREAMING_SNAKE_CASE_ ) ) + 100 lowerCAmelCase__ : Any = prime_sieve(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : List[Any] = 0 lowerCAmelCase__ : Tuple = primes[prime_index] while (last_prime**2) <= limit: lowerCAmelCase__ : str = primes[prime_index + 1] lowerCAmelCase__ : Union[str, Any] = last_prime**2 lowerCAmelCase__ : Any = next_prime**2 # Get numbers divisible by lps(current) lowerCAmelCase__ : Union[str, Any] = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) lowerCAmelCase__ : str = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps lowerCAmelCase__ : str = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair lowerCAmelCase__ : int = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
69
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A__ : def __init__( self : int , a : str , a : Union[str, Any]=13 , a : int=32 , a : Optional[Any]=2 , a : Tuple=3 , a : List[Any]=16 , a : List[str]=[1, 2, 1] , a : int=[2, 2, 4] , a : int=2 , a : Optional[Any]=2.0 , a : Optional[int]=True , a : Dict=0.0 , a : Any=0.0 , a : int=0.1 , a : List[str]="gelu" , a : Optional[Any]=False , a : str=True , a : Dict=0.0_2 , a : Any=1E-5 , a : Optional[int]=True , a : str=None , a : str=True , a : int=10 , a : str=8 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : Optional[Any] = patch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = embed_dim lowerCAmelCase__ : Tuple = depths lowerCAmelCase__ : List[str] = num_heads lowerCAmelCase__ : List[Any] = window_size lowerCAmelCase__ : Any = mlp_ratio lowerCAmelCase__ : Optional[Any] = qkv_bias lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : int = drop_path_rate lowerCAmelCase__ : Optional[Any] = hidden_act lowerCAmelCase__ : int = use_absolute_embeddings lowerCAmelCase__ : List[str] = patch_norm lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Dict = use_labels lowerCAmelCase__ : List[Any] = type_sequence_label_size lowerCAmelCase__ : Optional[Any] = encoder_stride def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = None if self.use_labels: lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : int = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : List[str] , a : Any , a : str , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = SwinvaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model(a ) lowerCAmelCase__ : str = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase__ : Tuple = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SwinvaForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : Dict = SwinvaForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Union[str, Any] , a : int , a : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.type_sequence_label_size lowerCAmelCase__ : List[Any] = SwinvaForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = config_and_inputs lowerCAmelCase__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = SwinvaModelTester(self ) lowerCAmelCase__ : int = ConfigTester(self , config_class=a , embed_dim=37 ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[int] = model_class(a ) lowerCAmelCase__ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Tuple = [*signature.parameters.keys()] lowerCAmelCase__ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = True for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = True lowerCAmelCase__ : str = False lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Dict = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : int = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Dict = outputs.attentions lowerCAmelCase__ : Dict = len(self.model_tester.depths ) self.assertEqual(len(a ) , a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Optional[int] = config.window_size**2 lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowerCAmelCase__ : Tuple = len(a ) # Check attention is always last and order is fine lowerCAmelCase__ : str = True lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(**self._prepare_for_class(a , a ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): lowerCAmelCase__ : Optional[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase__ : Any = 2 self.assertEqual(out_len + added_hidden_states , len(a ) ) lowerCAmelCase__ : Dict = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _lowerCamelCase ( self : int , a : Optional[int] , a : int , a : Optional[Any] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.hidden_states lowerCAmelCase__ : str = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a ) , a ) # Swinv2 has a different seq_length lowerCAmelCase__ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowerCAmelCase__ : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(a ) , a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = reshaped_hidden_states[0].shape lowerCAmelCase__ : List[str] = ( reshaped_hidden_states[0].view(a , a , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : str = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : List[str] = True self.check_hidden_states_output(a , a , a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Any = 3 lowerCAmelCase__ : int = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase__ : str = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : str = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase__ : Any = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase__ : str = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[str] = SwinvaModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = _config_zero_init(a ) for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(config=a ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( a ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Any = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**a ) # verify the logits lowerCAmelCase__ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Optional[Any] = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) )
69
1
from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar lowerCamelCase__ = TypeVar("""KEY""") lowerCamelCase__ = TypeVar("""VAL""") @dataclass(frozen=__magic_name__ , slots=__magic_name__ ) class A__ ( Generic[KEY, VAL] ): lowercase = 42 lowercase = 42 class A__ ( _Item ): def __init__( self : Optional[int] ): '''simple docstring''' super().__init__(a , a ) def __bool__( self : Any ): '''simple docstring''' return False lowerCamelCase__ = _DeletedItem() class A__ ( MutableMapping[KEY, VAL] ): def __init__( self : Union[str, Any] , a : int = 8 , a : float = 0.7_5 ): '''simple docstring''' lowerCAmelCase__ : List[Any] = initial_block_size lowerCAmelCase__ : list[_Item | None] = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 lowerCAmelCase__ : List[str] = capacity_factor lowerCAmelCase__ : str = 0 def _lowerCamelCase ( self : Optional[int] , a : KEY ): '''simple docstring''' return hash(a ) % len(self._buckets ) def _lowerCamelCase ( self : Any , a : int ): '''simple docstring''' return (ind + 1) % len(self._buckets ) def _lowerCamelCase ( self : Tuple , a : int , a : KEY , a : VAL ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self._buckets[ind] if not stored: lowerCAmelCase__ : Union[str, Any] = _Item(a , a ) self._len += 1 return True elif stored.key == key: lowerCAmelCase__ : Any = _Item(a , a ) return True else: return False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = len(self._buckets ) * self._capacity_factor return len(self ) >= int(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if len(self._buckets ) <= self._initial_block_size: return False lowerCAmelCase__ : Optional[int] = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def _lowerCamelCase ( self : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = self._buckets lowerCAmelCase__ : List[str] = [None] * new_size lowerCAmelCase__ : Optional[Any] = 0 for item in old_buckets: if item: self._add_item(item.key , item.val ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' self._resize(len(self._buckets ) * 2 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self._resize(len(self._buckets ) // 2 ) def _lowerCamelCase ( self : Tuple , a : KEY ): '''simple docstring''' lowerCAmelCase__ : str = self._get_bucket_index(a ) for _ in range(len(self._buckets ) ): yield ind lowerCAmelCase__ : int = self._get_next_ind(a ) def _lowerCamelCase ( self : Tuple , a : KEY , a : VAL ): '''simple docstring''' for ind in self._iterate_buckets(a ): if self._try_set(a , a , a ): break def __setitem__( self : List[Any] , a : KEY , a : VAL ): '''simple docstring''' if self._is_full(): self._size_up() self._add_item(a , a ) def __delitem__( self : List[str] , a : KEY ): '''simple docstring''' for ind in self._iterate_buckets(a ): lowerCAmelCase__ : Optional[int] = self._buckets[ind] if item is None: raise KeyError(a ) if item is _deleted: continue if item.key == key: lowerCAmelCase__ : int = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self : List[Any] , a : KEY ): '''simple docstring''' for ind in self._iterate_buckets(a ): lowerCAmelCase__ : Any = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(a ) def __len__( self : int ): '''simple docstring''' return self._len def __iter__( self : Union[str, Any] ): '''simple docstring''' yield from (item.key for item in self._buckets if item) def __repr__( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[str] = ' ,'.join( f'''{item.key}: {item.val}''' for item in self._buckets if item ) return f'''HashMap({val_string})'''
69
from itertools import permutations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase__ : str = [7, 11, 13, 17] for i, test in enumerate(SCREAMING_SNAKE_CASE_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 10 ) -> int: return sum( int(''.join(map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) ) for num in permutations(range(SCREAMING_SNAKE_CASE_ ) ) if is_substring_divisible(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
69
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['pixel_values'] def __init__( self : Optional[int] , a : bool = True , a : Dict[str, int] = None , a : PILImageResampling = PILImageResampling.BICUBIC , a : bool = True , a : Union[int, float] = 1 / 255 , a : bool = True , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : bool = True , **a : List[str] , ): '''simple docstring''' super().__init__(**a ) lowerCAmelCase__ : Dict = size if size is not None else {'height': 384, 'width': 384} lowerCAmelCase__ : Optional[Any] = get_size_dict(a , default_to_square=a ) lowerCAmelCase__ : List[Any] = do_resize lowerCAmelCase__ : List[Any] = size lowerCAmelCase__ : int = resample lowerCAmelCase__ : Any = do_rescale lowerCAmelCase__ : int = rescale_factor lowerCAmelCase__ : Optional[Any] = do_normalize lowerCAmelCase__ : str = image_mean if image_mean is not None else OPENAI_CLIP_MEAN lowerCAmelCase__ : Optional[Any] = image_std if image_std is not None else OPENAI_CLIP_STD lowerCAmelCase__ : Optional[int] = do_convert_rgb def _lowerCamelCase ( self : Dict , a : np.ndarray , a : Dict[str, int] , a : PILImageResampling = PILImageResampling.BICUBIC , a : Optional[Union[str, ChannelDimension]] = None , **a : List[Any] , ): '''simple docstring''' lowerCAmelCase__ : int = get_size_dict(a , default_to_square=a ) if "height" not in size or "width" not in size: raise ValueError(f'''The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}''' ) lowerCAmelCase__ : Optional[int] = (size['height'], size['width']) return resize(a , size=a , resample=a , data_format=a , **a ) def _lowerCamelCase ( self : Any , a : np.ndarray , a : Union[int, float] , a : Optional[Union[str, ChannelDimension]] = None , **a : Dict , ): '''simple docstring''' return rescale(a , scale=a , data_format=a , **a ) def _lowerCamelCase ( self : str , a : np.ndarray , a : Union[float, List[float]] , a : Union[float, List[float]] , a : Optional[Union[str, ChannelDimension]] = None , **a : str , ): '''simple docstring''' return normalize(a , mean=a , std=a , data_format=a , **a ) def _lowerCamelCase ( self : int , a : ImageInput , a : Optional[bool] = None , a : Optional[Dict[str, int]] = None , a : PILImageResampling = None , a : Optional[bool] = None , a : Optional[float] = None , a : Optional[bool] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[str, TensorType]] = None , a : bool = None , a : ChannelDimension = ChannelDimension.FIRST , **a : Dict , ): '''simple docstring''' lowerCAmelCase__ : List[Any] = do_resize if do_resize is not None else self.do_resize lowerCAmelCase__ : List[str] = resample if resample is not None else self.resample lowerCAmelCase__ : Any = do_rescale if do_rescale is not None else self.do_rescale lowerCAmelCase__ : str = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCAmelCase__ : Dict = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase__ : List[Any] = image_mean if image_mean is not None else self.image_mean lowerCAmelCase__ : str = image_std if image_std is not None else self.image_std lowerCAmelCase__ : List[Any] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowerCAmelCase__ : List[str] = size if size is not None else self.size lowerCAmelCase__ : str = get_size_dict(a , default_to_square=a ) lowerCAmelCase__ : Optional[int] = make_list_of_images(a ) if not valid_images(a ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowerCAmelCase__ : Tuple = [convert_to_rgb(a ) for image in images] # All transformations expect numpy arrays. lowerCAmelCase__ : Union[str, Any] = [to_numpy_array(a ) for image in images] if do_resize: lowerCAmelCase__ : Dict = [self.resize(image=a , size=a , resample=a ) for image in images] if do_rescale: lowerCAmelCase__ : int = [self.rescale(image=a , scale=a ) for image in images] if do_normalize: lowerCAmelCase__ : List[Any] = [self.normalize(image=a , mean=a , std=a ) for image in images] lowerCAmelCase__ : List[Any] = [to_channel_dimension_format(a , a ) for image in images] lowerCAmelCase__ : int = BatchFeature(data={'pixel_values': images} , tensor_type=a ) return encoded_outputs
69
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
69
1
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: if len(SCREAMING_SNAKE_CASE_ ) != len(SCREAMING_SNAKE_CASE_ ): raise ValueError('String lengths must match!' ) lowerCAmelCase__ : Tuple = 0 for chara, chara in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
69
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ , lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
69
1
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase__ : List[str] = BlipImageProcessor() lowerCAmelCase__ : Dict = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) lowerCAmelCase__ : Tuple = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) lowerCAmelCase__ : str = InstructBlipProcessor(a , a , a ) processor.save_pretrained(self.tmpdirname ) def _lowerCamelCase ( self : List[Any] , **a : List[Any] ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **a ).tokenizer def _lowerCamelCase ( self : List[Any] , **a : Dict ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **a ).image_processor def _lowerCamelCase ( self : Dict , **a : List[str] ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **a ).qformer_tokenizer def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCAmelCase__ : Any = [Image.fromarray(np.moveaxis(a , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) lowerCAmelCase__ : str = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) lowerCAmelCase__ : Tuple = self.get_image_processor(do_normalize=a , padding_value=1.0 ) lowerCAmelCase__ : str = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=a , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , a ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , a ) self.assertIsInstance(processor.qformer_tokenizer , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = self.get_image_processor() lowerCAmelCase__ : Optional[int] = self.get_tokenizer() lowerCAmelCase__ : Tuple = self.get_qformer_tokenizer() lowerCAmelCase__ : Optional[Any] = InstructBlipProcessor( tokenizer=a , image_processor=a , qformer_tokenizer=a ) lowerCAmelCase__ : Any = self.prepare_image_inputs() lowerCAmelCase__ : Tuple = image_processor(a , return_tensors='np' ) lowerCAmelCase__ : Union[str, Any] = processor(images=a , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.get_image_processor() lowerCAmelCase__ : Tuple = self.get_tokenizer() lowerCAmelCase__ : Dict = self.get_qformer_tokenizer() lowerCAmelCase__ : Any = InstructBlipProcessor( tokenizer=a , image_processor=a , qformer_tokenizer=a ) lowerCAmelCase__ : str = 'lower newer' lowerCAmelCase__ : Any = processor(text=a ) lowerCAmelCase__ : int = tokenizer(a , return_token_type_ids=a ) lowerCAmelCase__ : Any = qformer_tokenizer(a , return_token_type_ids=a ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['qformer_' + key] ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.get_image_processor() lowerCAmelCase__ : Any = self.get_tokenizer() lowerCAmelCase__ : int = self.get_qformer_tokenizer() lowerCAmelCase__ : int = InstructBlipProcessor( tokenizer=a , image_processor=a , qformer_tokenizer=a ) lowerCAmelCase__ : Optional[int] = 'lower newer' lowerCAmelCase__ : Any = self.prepare_image_inputs() lowerCAmelCase__ : Union[str, Any] = processor(text=a , images=a ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , ) # test if it raises when no input is passed with pytest.raises(a ): processor() def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.get_image_processor() lowerCAmelCase__ : str = self.get_tokenizer() lowerCAmelCase__ : Any = self.get_qformer_tokenizer() lowerCAmelCase__ : List[str] = InstructBlipProcessor( tokenizer=a , image_processor=a , qformer_tokenizer=a ) lowerCAmelCase__ : Tuple = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCAmelCase__ : List[Any] = processor.batch_decode(a ) lowerCAmelCase__ : Dict = tokenizer.batch_decode(a ) self.assertListEqual(a , a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = self.get_image_processor() lowerCAmelCase__ : Optional[Any] = self.get_tokenizer() lowerCAmelCase__ : Optional[int] = self.get_qformer_tokenizer() lowerCAmelCase__ : Tuple = InstructBlipProcessor( tokenizer=a , image_processor=a , qformer_tokenizer=a ) lowerCAmelCase__ : int = 'lower newer' lowerCAmelCase__ : List[str] = self.prepare_image_inputs() lowerCAmelCase__ : List[Any] = processor(text=a , images=a ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , )
69
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
1
import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class A__ : def __init__( self : Union[str, Any] , a : List[Any] , a : Tuple=13 , a : List[Any]=7 , a : Dict=True , a : int=True , a : Optional[Any]=False , a : Union[str, Any]=True , a : Optional[int]=99 , a : Union[str, Any]=64 , a : Optional[int]=5 , a : Dict=4 , a : Dict=64 , a : str="gelu" , a : int=0.1 , a : List[str]=0.1 , a : Tuple=512 , a : Optional[Any]=16 , a : Tuple=2 , a : List[Any]=0.0_2 , a : List[Any]=3 , a : List[Any]=4 , a : Any=None , ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = parent lowerCAmelCase__ : List[Any] = batch_size lowerCAmelCase__ : Optional[int] = seq_length lowerCAmelCase__ : Optional[int] = is_training lowerCAmelCase__ : str = use_input_mask lowerCAmelCase__ : List[Any] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : Optional[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : Dict = num_hidden_layers lowerCAmelCase__ : Optional[Any] = num_attention_heads lowerCAmelCase__ : Union[str, Any] = intermediate_size lowerCAmelCase__ : Dict = hidden_act lowerCAmelCase__ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase__ : List[str] = attention_probs_dropout_prob lowerCAmelCase__ : List[str] = max_position_embeddings lowerCAmelCase__ : List[str] = type_vocab_size lowerCAmelCase__ : Any = type_sequence_label_size lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : str = num_labels lowerCAmelCase__ : Union[str, Any] = num_choices lowerCAmelCase__ : Optional[Any] = scope def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return MPNetConfig.from_pretrained('microsoft/mpnet-base' ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Union[str, Any] = None if self.use_input_mask: lowerCAmelCase__ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : Optional[int] = None lowerCAmelCase__ : Optional[int] = None lowerCAmelCase__ : Optional[Any] = None if self.use_labels: lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : Any = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def _lowerCamelCase ( self : Optional[Any] , a : Any , a : Union[str, Any] , a : int , a : Tuple , a : Any , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = MPNetModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : int = model(a , a ) lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowerCamelCase ( self : Any , a : Optional[int] , a : Dict , a : int , a : Dict , a : Optional[Any] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = MPNetForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : int = model( a , attention_mask=a , start_positions=a , end_positions=a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Any , a : Any , a : Union[str, Any] , a : List[str] , a : List[str] , a : Tuple , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = self.num_labels lowerCAmelCase__ : Union[str, Any] = MPNetForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Optional[Any] , a : Tuple , a : int , a : List[Any] , a : Dict , a : Union[str, Any] , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Dict = MPNetForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : int = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : Dict = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Optional[Any] , a : Optional[int] , a : Optional[int] , a : Optional[int] , a : Any , a : int , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.num_labels lowerCAmelCase__ : Optional[int] = MPNetForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : str = config_and_inputs lowerCAmelCase__ : Tuple = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) lowercase = ( { 'feature-extraction': MPNetModel, 'fill-mask': MPNetForMaskedLM, 'question-answering': MPNetForQuestionAnswering, 'text-classification': MPNetForSequenceClassification, 'token-classification': MPNetForTokenClassification, 'zero-shot': MPNetForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = MPNetModelTester(self ) lowerCAmelCase__ : Any = ConfigTester(self , config_class=a , hidden_size=37 ) def _lowerCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*a ) @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = MPNetModel.from_pretrained('microsoft/mpnet-base' ) lowerCAmelCase__ : Union[str, Any] = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) lowerCAmelCase__ : Union[str, Any] = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : str = torch.tensor( [[[-0.0_5_5_0, 0.1_9_4_3, -0.0_7_4_0], [-0.0_5_6_2, 0.2_2_1_1, -0.0_5_7_9], [-0.0_4_3_7, 0.3_3_3_7, -0.0_6_4_1]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , a , atol=1E-4 ) )
69
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """google/rembert""": 256, } lowerCamelCase__ = """▁""" class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = RemBertTokenizer def __init__( self : Optional[Any] , a : str=None , a : Any=None , a : List[Any]=True , a : str=True , a : Dict=False , a : Dict="[CLS]" , a : int="[SEP]" , a : Tuple="<unk>" , a : Optional[Any]="[SEP]" , a : Tuple="<pad>" , a : Dict="[CLS]" , a : Optional[Any]="[MASK]" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( a , tokenizer_file=a , do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : int = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : List[Any] = keep_accents lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
69
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase__ = { """configuration_graphormer""": ["""GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GraphormerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """GraphormerForGraphClassification""", """GraphormerModel""", """GraphormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : int = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : List[Any] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : int = None lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Dict = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Union[str, Any] = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
69
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCamelCase__ = {"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
1
from __future__ import annotations from typing import Any class A__ : def __init__( self : List[str] , a : int , a : int , a : float = 0 ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = row, column lowerCAmelCase__ : int = [[default_value for c in range(a )] for r in range(a )] def __str__( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = f'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier lowerCAmelCase__ : Tuple = 0 for row_vector in self.array: for obj in row_vector: lowerCAmelCase__ : Any = max(a , len(str(a ) ) ) lowerCAmelCase__ : int = f'''%{max_element_length}s''' # Make string and return def single_line(a : list[float] ) -> str: nonlocal string_format_identifier lowerCAmelCase__ : int = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(a ) for row_vector in self.array ) return s def __repr__( self : int ): '''simple docstring''' return str(self ) def _lowerCamelCase ( self : Optional[Any] , a : tuple[int, int] ): '''simple docstring''' if not (isinstance(a , (list, tuple) ) and len(a ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : List[str] , a : tuple[int, int] ): '''simple docstring''' assert self.validate_indicies(a ) return self.array[loc[0]][loc[1]] def __setitem__( self : Any , a : tuple[int, int] , a : float ): '''simple docstring''' assert self.validate_indicies(a ) lowerCAmelCase__ : Optional[Any] = value def __add__( self : Tuple , a : Matrix ): '''simple docstring''' assert isinstance(a , a ) assert self.row == another.row and self.column == another.column # Add lowerCAmelCase__ : List[Any] = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase__ : Tuple = self[r, c] + another[r, c] return result def __neg__( self : Dict ): '''simple docstring''' lowerCAmelCase__ : str = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase__ : List[Any] = -self[r, c] return result def __sub__( self : Dict , a : Matrix ): '''simple docstring''' return self + (-another) def __mul__( self : Tuple , a : int | float | Matrix ): '''simple docstring''' if isinstance(a , (int, float) ): # Scalar multiplication lowerCAmelCase__ : Any = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase__ : Optional[Any] = self[r, c] * another return result elif isinstance(a , a ): # Matrix multiplication assert self.column == another.row lowerCAmelCase__ : List[str] = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: lowerCAmelCase__ : Union[str, Any] = f'''Unsupported type given for another ({type(a )})''' raise TypeError(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase__ : Optional[int] = self[r, c] return result def _lowerCamelCase ( self : List[str] , a : Matrix , a : Matrix ): '''simple docstring''' assert isinstance(a , a ) and isinstance(a , a ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate lowerCAmelCase__ : Any = v.transpose() lowerCAmelCase__ : Any = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def lowerCAmelCase__ ( ) -> None: # a^(-1) lowerCAmelCase__ : Any = Matrix(3 , 3 , 0 ) for i in range(3 ): lowerCAmelCase__ : List[str] = 1 print(F'''a^(-1) is {ainv}''' ) # u, v lowerCAmelCase__ : Dict = Matrix(3 , 1 , 0 ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = 1, 2, -3 lowerCAmelCase__ : Any = Matrix(3 , 1 , 0 ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = 4, -2, 5 print(F'''u is {u}''' ) print(F'''v is {v}''' ) print(F'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}''' ) def lowerCAmelCase__ ( ) -> None: import doctest doctest.testmod() testa()
69
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
1
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("""1.6"""): lowerCamelCase__ = True from torch.cuda.amp import autocast lowerCamelCase__ = logging.getLogger(__name__) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None ) -> Dict: return field(default_factory=lambda: default , metadata=SCREAMING_SNAKE_CASE_ ) @dataclass class A__ : lowercase = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) lowercase = field( default=__magic_name__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) lowercase = field( default=0.1 , metadata={'help': 'The dropout ratio for the attention probabilities.'} ) lowercase = field( default=0.1 , metadata={'help': 'The dropout ratio for activations inside the fully connected layer.'} ) lowercase = field( default=0.1 , metadata={ 'help': 'The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.' } , ) lowercase = field( default=0.1 , metadata={'help': 'The dropout probabilitiy for all 1D convolutional layers in feature extractor.'} , ) lowercase = field( default=0.05 , metadata={ 'help': ( 'Propability of each feature vector along the time axis to be chosen as the start of the vector' 'span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature' 'vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.' ) } , ) lowercase = field(default=0.0 , metadata={'help': 'The LayerDrop probability.'} ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) lowercase = field( default='train+validation' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) lowercase = field( default=__magic_name__ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) lowercase = field( default=__magic_name__ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) lowercase = field( default=__magic_name__ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of validation examples to this ' 'value if set.' ) } , ) lowercase = list_field( default=[',', '?', '.', '!', '-', ';', ':', '""', '%', '\'', '"', '�'] , metadata={'help': 'A list of characters to remove from the transcripts.'} , ) @dataclass class A__ : lowercase = 42 lowercase = True lowercase = None lowercase = None lowercase = None lowercase = None def __call__( self : Tuple , a : List[Dict[str, Union[List[int], torch.Tensor]]] ): '''simple docstring''' lowerCAmelCase__ : Dict = [{'input_values': feature['input_values']} for feature in features] lowerCAmelCase__ : int = [{'input_ids': feature['labels']} for feature in features] lowerCAmelCase__ : Dict = self.processor.pad( a , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) lowerCAmelCase__ : Dict = self.processor.pad( labels=a , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors='pt' , ) # replace padding with -100 to ignore loss correctly lowerCAmelCase__ : List[Any] = labels_batch['input_ids'].masked_fill(labels_batch.attention_mask.ne(1 ) , -100 ) lowerCAmelCase__ : Any = labels return batch class A__ ( __magic_name__ ): def _lowerCamelCase ( self : Tuple , a : nn.Module , a : Dict[str, Union[torch.Tensor, Any]] ): '''simple docstring''' model.train() lowerCAmelCase__ : Optional[Any] = self._prepare_inputs(a ) if self.use_amp: with autocast(): lowerCAmelCase__ : Optional[int] = self.compute_loss(a , a ) else: lowerCAmelCase__ : str = self.compute_loss(a , a ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": lowerCAmelCase__ : Union[str, Any] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowerCAmelCase__ : List[str] = loss.sum() / (inputs['labels'] >= 0).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowerCAmelCase__ : Tuple = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(a ).backward() elif self.use_apex: with amp.scale_loss(a , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(a ) else: loss.backward() return loss.detach() def lowerCAmelCase__ ( ) -> Union[str, Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase__ : Union[str, Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = parser.parse_args_into_dataclasses() # Detecting last checkpoint. lowerCAmelCase__ : Optional[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCAmelCase__ : Union[str, Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info('Training/evaluation parameters %s' , SCREAMING_SNAKE_CASE_ ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: lowerCAmelCase__ : List[Any] = datasets.load_dataset( 'common_voice' , data_args.dataset_config_name , split=data_args.train_split_name ) lowerCAmelCase__ : List[Any] = datasets.load_dataset('common_voice' , data_args.dataset_config_name , split='test' ) # Create and save tokenizer lowerCAmelCase__ : Dict = F'''[{"".join(data_args.chars_to_ignore )}]''' def remove_special_characters(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Optional[int] = re.sub(SCREAMING_SNAKE_CASE_ , '' , batch['sentence'] ).lower() + ' ' return batch lowerCAmelCase__ : Any = train_dataset.map(SCREAMING_SNAKE_CASE_ , remove_columns=['sentence'] ) lowerCAmelCase__ : Any = eval_dataset.map(SCREAMING_SNAKE_CASE_ , remove_columns=['sentence'] ) def extract_all_chars(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Any = ' '.join(batch['text'] ) lowerCAmelCase__ : int = list(set(SCREAMING_SNAKE_CASE_ ) ) return {"vocab": [vocab], "all_text": [all_text]} lowerCAmelCase__ : List[str] = train_dataset.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , batch_size=-1 , keep_in_memory=SCREAMING_SNAKE_CASE_ , remove_columns=train_dataset.column_names , ) lowerCAmelCase__ : str = train_dataset.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , batch_size=-1 , keep_in_memory=SCREAMING_SNAKE_CASE_ , remove_columns=eval_dataset.column_names , ) lowerCAmelCase__ : Optional[Any] = list(set(vocab_train['vocab'][0] ) | set(vocab_test['vocab'][0] ) ) lowerCAmelCase__ : Tuple = {v: k for k, v in enumerate(SCREAMING_SNAKE_CASE_ )} lowerCAmelCase__ : Tuple = vocab_dict[' '] del vocab_dict[" "] lowerCAmelCase__ : Optional[Any] = len(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = len(SCREAMING_SNAKE_CASE_ ) with open('vocab.json' , 'w' ) as vocab_file: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCAmelCase__ : str = WavaVecaCTCTokenizer( 'vocab.json' , unk_token='[UNK]' , pad_token='[PAD]' , word_delimiter_token='|' , ) lowerCAmelCase__ : List[Any] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0.0 , do_normalize=SCREAMING_SNAKE_CASE_ , return_attention_mask=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = WavaVecaProcessor(feature_extractor=SCREAMING_SNAKE_CASE_ , tokenizer=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction='mean' , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: lowerCAmelCase__ : Union[str, Any] = min(len(SCREAMING_SNAKE_CASE_ ) , data_args.max_train_samples ) lowerCAmelCase__ : List[str] = train_dataset.select(range(SCREAMING_SNAKE_CASE_ ) ) if data_args.max_val_samples is not None: lowerCAmelCase__ : List[Any] = eval_dataset.select(range(data_args.max_val_samples ) ) lowerCAmelCase__ : Union[str, Any] = torchaudio.transforms.Resample(48_000 , 16_000 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = torchaudio.load(batch['path'] ) lowerCAmelCase__ : Tuple = resampler(SCREAMING_SNAKE_CASE_ ).squeeze().numpy() lowerCAmelCase__ : Dict = 16_000 lowerCAmelCase__ : int = batch['text'] return batch lowerCAmelCase__ : Union[str, Any] = train_dataset.map( SCREAMING_SNAKE_CASE_ , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) lowerCAmelCase__ : int = eval_dataset.map( SCREAMING_SNAKE_CASE_ , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(SCREAMING_SNAKE_CASE_ ): # check that all files have the correct sampling rate assert ( len(set(batch['sampling_rate'] ) ) == 1 ), F'''Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.''' lowerCAmelCase__ : Tuple = processor( audio=batch['speech'] , text=batch['target_text'] , sampling_rate=batch['sampling_rate'][0] ) batch.update(SCREAMING_SNAKE_CASE_ ) return batch lowerCAmelCase__ : List[Any] = train_dataset.map( SCREAMING_SNAKE_CASE_ , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=SCREAMING_SNAKE_CASE_ , num_proc=data_args.preprocessing_num_workers , ) lowerCAmelCase__ : Tuple = eval_dataset.map( SCREAMING_SNAKE_CASE_ , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=SCREAMING_SNAKE_CASE_ , num_proc=data_args.preprocessing_num_workers , ) # Metric lowerCAmelCase__ : List[Any] = datasets.load_metric('wer' ) def compute_metrics(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Optional[int] = pred.predictions lowerCAmelCase__ : Optional[Any] = np.argmax(SCREAMING_SNAKE_CASE_ , axis=-1 ) lowerCAmelCase__ : str = processor.tokenizer.pad_token_id lowerCAmelCase__ : Optional[int] = processor.batch_decode(SCREAMING_SNAKE_CASE_ ) # we do not want to group tokens when computing the metrics lowerCAmelCase__ : Optional[int] = processor.batch_decode(pred.label_ids , group_tokens=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = wer_metric.compute(predictions=SCREAMING_SNAKE_CASE_ , references=SCREAMING_SNAKE_CASE_ ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator lowerCAmelCase__ : Optional[Any] = DataCollatorCTCWithPadding(processor=SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ ) # Initialize our Trainer lowerCAmelCase__ : Dict = CTCTrainer( model=SCREAMING_SNAKE_CASE_ , data_collator=SCREAMING_SNAKE_CASE_ , args=SCREAMING_SNAKE_CASE_ , compute_metrics=SCREAMING_SNAKE_CASE_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: lowerCAmelCase__ : Union[str, Any] = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): lowerCAmelCase__ : Tuple = model_args.model_name_or_path else: lowerCAmelCase__ : List[Any] = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) lowerCAmelCase__ : str = trainer.train(resume_from_checkpoint=SCREAMING_SNAKE_CASE_ ) trainer.save_model() lowerCAmelCase__ : Optional[int] = train_result.metrics lowerCAmelCase__ : int = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase__ : List[Any] = min(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ) trainer.log_metrics('train' , SCREAMING_SNAKE_CASE_ ) trainer.save_metrics('train' , SCREAMING_SNAKE_CASE_ ) trainer.save_state() # Evaluation lowerCAmelCase__ : int = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) lowerCAmelCase__ : Any = trainer.evaluate() lowerCAmelCase__ : Optional[int] = data_args.max_val_samples if data_args.max_val_samples is not None else len(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Tuple = min(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ) trainer.log_metrics('eval' , SCREAMING_SNAKE_CASE_ ) trainer.save_metrics('eval' , SCREAMING_SNAKE_CASE_ ) return results if __name__ == "__main__": main()
69
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
1
from __future__ import annotations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> set[str]: lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = set(SCREAMING_SNAKE_CASE_ ), [start] while stack: lowerCAmelCase__ : Optional[int] = stack.pop() explored.add(SCREAMING_SNAKE_CASE_ ) # Differences from BFS: # 1) pop last element instead of first one # 2) add adjacent elements to stack without exploring them for adj in reversed(graph[v] ): if adj not in explored: stack.append(SCREAMING_SNAKE_CASE_ ) return explored lowerCamelCase__ = { """A""": ["""B""", """C""", """D"""], """B""": ["""A""", """D""", """E"""], """C""": ["""A""", """F"""], """D""": ["""B""", """D"""], """E""": ["""B""", """F"""], """F""": ["""C""", """E""", """G"""], """G""": ["""F"""], } if __name__ == "__main__": import doctest doctest.testmod() print(depth_first_search(G, """A"""))
69
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = StableDiffusionInpaintPipeline lowercase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS lowercase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowercase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowercase = frozenset([] ) def _lowerCamelCase ( self : str ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=a , ) lowerCAmelCase__ : Dict = PNDMScheduler(skip_prk_steps=a ) torch.manual_seed(0 ) lowerCAmelCase__ : int = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) lowerCAmelCase__ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='gelu' , projection_dim=512 , ) lowerCAmelCase__ : Optional[int] = CLIPTextModel(a ) lowerCAmelCase__ : List[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) lowerCAmelCase__ : Tuple = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def _lowerCamelCase ( self : Optional[Any] , a : Dict , a : Union[str, Any]=0 ): '''simple docstring''' lowerCAmelCase__ : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(a ) ).to(a ) lowerCAmelCase__ : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase__ : Dict = Image.fromarray(np.uinta(a ) ).convert('RGB' ).resize((64, 64) ) lowerCAmelCase__ : Optional[Any] = Image.fromarray(np.uinta(image + 4 ) ).convert('RGB' ).resize((64, 64) ) if str(a ).startswith('mps' ): lowerCAmelCase__ : Optional[int] = torch.manual_seed(a ) else: lowerCAmelCase__ : Union[str, Any] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Optional[int] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': init_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[str] = StableDiffusionInpaintPipeline(**a ) lowerCAmelCase__ : Tuple = sd_pipe.to(a ) sd_pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[int] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Tuple = sd_pipe(**a ).images lowerCAmelCase__ : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Union[str, Any] = np.array([0.4_7_2_7, 0.5_7_3_5, 0.3_9_4_1, 0.5_4_4_6, 0.5_9_2_6, 0.4_3_9_4, 0.5_0_6_2, 0.4_6_5_4, 0.4_4_7_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) lowerCAmelCase__ : List[str] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) lowerCAmelCase__ : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint' '/yellow_cat_sitting_on_a_park_bench.npy' ) lowerCAmelCase__ : List[str] = 'stabilityai/stable-diffusion-2-inpainting' lowerCAmelCase__ : str = StableDiffusionInpaintPipeline.from_pretrained(a , safety_checker=a ) pipe.to(a ) pipe.set_progress_bar_config(disable=a ) pipe.enable_attention_slicing() lowerCAmelCase__ : Dict = 'Face of a yellow cat, high resolution, sitting on a park bench' lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) lowerCAmelCase__ : List[str] = pipe( prompt=a , image=a , mask_image=a , generator=a , output_type='np' , ) lowerCAmelCase__ : List[str] = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : str = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) lowerCAmelCase__ : Union[str, Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) lowerCAmelCase__ : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint' '/yellow_cat_sitting_on_a_park_bench_fp16.npy' ) lowerCAmelCase__ : Tuple = 'stabilityai/stable-diffusion-2-inpainting' lowerCAmelCase__ : Tuple = StableDiffusionInpaintPipeline.from_pretrained( a , torch_dtype=torch.floataa , safety_checker=a , ) pipe.to(a ) pipe.set_progress_bar_config(disable=a ) pipe.enable_attention_slicing() lowerCAmelCase__ : Optional[int] = 'Face of a yellow cat, high resolution, sitting on a park bench' lowerCAmelCase__ : int = torch.manual_seed(0 ) lowerCAmelCase__ : List[Any] = pipe( prompt=a , image=a , mask_image=a , generator=a , output_type='np' , ) lowerCAmelCase__ : Tuple = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def _lowerCamelCase ( self : Any ): '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCAmelCase__ : List[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) lowerCAmelCase__ : List[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) lowerCAmelCase__ : Tuple = 'stabilityai/stable-diffusion-2-inpainting' lowerCAmelCase__ : List[str] = PNDMScheduler.from_pretrained(a , subfolder='scheduler' ) lowerCAmelCase__ : List[Any] = StableDiffusionInpaintPipeline.from_pretrained( a , safety_checker=a , scheduler=a , torch_dtype=torch.floataa , ) pipe.to(a ) pipe.set_progress_bar_config(disable=a ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() lowerCAmelCase__ : List[str] = 'Face of a yellow cat, high resolution, sitting on a park bench' lowerCAmelCase__ : List[str] = torch.manual_seed(0 ) lowerCAmelCase__ : Union[str, Any] = pipe( prompt=a , image=a , mask_image=a , generator=a , num_inference_steps=2 , output_type='np' , ) lowerCAmelCase__ : str = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.6_5 * 10**9
69
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
1
import os import jsonlines import numpy as np from tqdm import tqdm lowerCamelCase__ = 2048 lowerCamelCase__ = 4096 lowerCamelCase__ = 42 lowerCamelCase__ = os.environ.pop("""PROCESS_TRAIN""", """false""") lowerCamelCase__ = {"""null""": 0, """short""": 1, """long""": 2, """yes""": 3, """no""": 4} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: def choose_first(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): assert isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) == 1: lowerCAmelCase__ : Optional[Any] = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: lowerCAmelCase__ : Optional[int] = {k: [a[k]] for k in a} if len(a['start_token'] ) > 0: break return a lowerCAmelCase__ : Dict = {'id': example['id']} lowerCAmelCase__ : Tuple = example['annotations'] lowerCAmelCase__ : int = annotation['yes_no_answer'] if 0 in yes_no_answer or 1 in yes_no_answer: lowerCAmelCase__ : int = ['yes'] if 1 in yes_no_answer else ['no'] lowerCAmelCase__ : Dict = [] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Any = ['<cls>'] else: lowerCAmelCase__ : Any = ['short'] lowerCAmelCase__ : Dict = choose_first(annotation['short_answers'] ) if len(out['start_token'] ) == 0: # answer will be long if short is not available lowerCAmelCase__ : List[str] = ['long'] lowerCAmelCase__ : List[str] = choose_first(annotation['long_answer'] , is_long_answer=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = [] answer.update(SCREAMING_SNAKE_CASE_ ) # disregard some samples if len(answer['start_token'] ) > 1 or answer["start_token"] == answer["end_token"]: lowerCAmelCase__ : Any = True else: lowerCAmelCase__ : Tuple = False lowerCAmelCase__ : Tuple = ['start_token', 'end_token', 'start_byte', 'end_byte', 'text'] if not all(isinstance(answer[k] , SCREAMING_SNAKE_CASE_ ) for k in cols ): raise ValueError('Issue in ID' , example['id'] ) return answer def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ) -> Union[str, Any]: lowerCAmelCase__ : int = _get_single_answer(SCREAMING_SNAKE_CASE_ ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element lowerCAmelCase__ : List[Any] = example['document']['tokens'] lowerCAmelCase__ : Tuple = [] for i in range(len(doc['token'] ) ): if not doc["is_html"][i]: context.append(doc['token'][i] ) return { "context": " ".join(SCREAMING_SNAKE_CASE_ ), "answer": { "start_token": -100, # ignore index in cross-entropy "end_token": -100, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples lowerCAmelCase__ : int = ['start_token', 'end_token'] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 lowerCAmelCase__ : Tuple = example['document']['tokens'] lowerCAmelCase__ : int = answer['start_token'] lowerCAmelCase__ : List[Any] = answer['end_token'] lowerCAmelCase__ : Tuple = [] for i in range(len(doc['token'] ) ): if not doc["is_html"][i]: context.append(doc['token'][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 lowerCAmelCase__ : Tuple = ' '.join(context[start_token:end_token] ) # checking above code if assertion: lowerCAmelCase__ : Union[str, Any] = doc['is_html'][answer['start_token'] : answer['end_token']] lowerCAmelCase__ : List[Any] = doc['token'][answer['start_token'] : answer['end_token']] lowerCAmelCase__ : Any = ' '.join([old[i] for i in range(len(SCREAMING_SNAKE_CASE_ ) ) if not is_html[i]] ) if new != old: print('ID:' , example['id'] ) print('New:' , SCREAMING_SNAKE_CASE_ , end='\n' ) print('Old:' , SCREAMING_SNAKE_CASE_ , end='\n\n' ) return { "context": " ".join(SCREAMING_SNAKE_CASE_ ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=2_048 , SCREAMING_SNAKE_CASE_=4_096 , SCREAMING_SNAKE_CASE_=True ) -> Dict: # overlap will be of doc_stride - q_len lowerCAmelCase__ : Optional[int] = get_context_and_ans(SCREAMING_SNAKE_CASE_ , assertion=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = out['answer'] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } lowerCAmelCase__ : int = tokenizer(example['question']['text'] , out['context'] ).input_ids lowerCAmelCase__ : List[str] = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : Dict = input_ids[:q_len] lowerCAmelCase__ : List[Any] = range(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) , max_length - doc_stride ) for i in doc_start_indices: lowerCAmelCase__ : Tuple = i + max_length - q_len lowerCAmelCase__ : Optional[Any] = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer['category'][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-100] * len(SCREAMING_SNAKE_CASE_ ), "end_token": [-100] * len(SCREAMING_SNAKE_CASE_ ), "category": category, }, } lowerCAmelCase__ : List[str] = out['context'].split() lowerCAmelCase__ : int = splitted_context[answer['end_token']] lowerCAmelCase__ : List[str] = len( tokenizer( ' '.join(splitted_context[: answer['start_token']] ) , add_special_tokens=SCREAMING_SNAKE_CASE_ , ).input_ids ) lowerCAmelCase__ : int = len( tokenizer(' '.join(splitted_context[: answer['end_token']] ) , add_special_tokens=SCREAMING_SNAKE_CASE_ ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token lowerCAmelCase__ : List[Any] = len(tokenizer(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 lowerCAmelCase__ : Optional[Any] = input_ids[answer['start_token'] : answer['end_token'] + 1] # right & left are inclusive lowerCAmelCase__ : Optional[int] = answer['start_token'] lowerCAmelCase__ : Dict = answer['end_token'] if assertion: lowerCAmelCase__ : Union[str, Any] = tokenizer.decode(SCREAMING_SNAKE_CASE_ ) if answer["span"] != new: print('ISSUE IN TOKENIZATION' ) print('OLD:' , answer['span'] ) print('NEW:' , SCREAMING_SNAKE_CASE_ , end='\n\n' ) if len(SCREAMING_SNAKE_CASE_ ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } lowerCAmelCase__ : Optional[Any] = input_ids[:q_len] lowerCAmelCase__ : Union[str, Any] = range(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) , max_length - doc_stride ) lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Any = [] lowerCAmelCase__ : Dict = [] lowerCAmelCase__ : int = [] # null, yes, no, long, short for i in doc_start_indices: lowerCAmelCase__ : Optional[int] = i + max_length - q_len lowerCAmelCase__ : int = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: lowerCAmelCase__ : Optional[int] = start_token - i + q_len lowerCAmelCase__ : List[Any] = end_token - i + q_len answers_category.append(answer['category'][0] ) # ["short"] -> "short" else: lowerCAmelCase__ : int = -100 lowerCAmelCase__ : Union[str, Any] = -100 answers_category.append('null' ) lowerCAmelCase__ : List[str] = inputs[-1][start_token : end_token + 1] answers_start_token.append(SCREAMING_SNAKE_CASE_ ) answers_end_token.append(SCREAMING_SNAKE_CASE_ ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print('ISSUE in strided for ID:' , example['id'] ) print('New:' , tokenizer.decode(SCREAMING_SNAKE_CASE_ ) ) print('Old:' , tokenizer.decode(SCREAMING_SNAKE_CASE_ ) , end='\n\n' ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=2_048 , SCREAMING_SNAKE_CASE_=4_096 , SCREAMING_SNAKE_CASE_=False ) -> Union[str, Any]: lowerCAmelCase__ : Dict = get_strided_contexts_and_ans( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , doc_stride=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , assertion=SCREAMING_SNAKE_CASE_ , ) return example def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: with jsonlines.open(SCREAMING_SNAKE_CASE_ , 'a' ) as writer: for example in tqdm(SCREAMING_SNAKE_CASE_ , total=len(SCREAMING_SNAKE_CASE_ ) , desc='Saving samples ... ' ): lowerCAmelCase__ : Union[str, Any] = example['labels'] for ids, start, end, cat in zip( example['input_ids'] , labels['start_token'] , labels['end_token'] , labels['category'] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { 'input_ids': ids, 'start_token': start, 'end_token': end, 'category': CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer lowerCamelCase__ = load_dataset("""natural_questions""") lowerCamelCase__ = BigBirdTokenizer.from_pretrained("""google/bigbird-roberta-base""") lowerCamelCase__ = data["""train""" if PROCESS_TRAIN == """true""" else """validation"""] lowerCamelCase__ = { """tokenizer""": tokenizer, """doc_stride""": DOC_STRIDE, """max_length""": MAX_LENGTH, """assertion""": False, } lowerCamelCase__ = data.map(prepare_inputs, fn_kwargs=fn_kwargs) lowerCamelCase__ = data.remove_columns(["""annotations""", """document""", """id""", """question"""]) print(data) np.random.seed(SEED) lowerCamelCase__ = """nq-training.jsonl""" if PROCESS_TRAIN == """true""" else """nq-validation.jsonl""" save_to_disk(data, file_name=cache_file_name)
69
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
1
from __future__ import annotations import unittest from transformers import RoFormerConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerModel, ) from transformers.models.roformer.modeling_tf_roformer import ( TFRoFormerSelfAttention, TFRoFormerSinusoidalPositionalEmbedding, ) class A__ : def __init__( self : Union[str, Any] , a : Tuple , a : Optional[int]=13 , a : Any=7 , a : Optional[int]=True , a : Optional[int]=True , a : Optional[int]=True , a : str=True , a : Dict=99 , a : List[Any]=32 , a : Tuple=2 , a : List[str]=4 , a : int=37 , a : Union[str, Any]="gelu" , a : List[str]=0.1 , a : List[Any]=0.1 , a : List[Any]=512 , a : List[str]=16 , a : Optional[Any]=2 , a : Tuple=0.0_2 , a : str=3 , a : Optional[int]=4 , a : Any=None , ): '''simple docstring''' lowerCAmelCase__ : Any = parent lowerCAmelCase__ : Any = 13 lowerCAmelCase__ : Optional[Any] = 7 lowerCAmelCase__ : Any = True lowerCAmelCase__ : str = True lowerCAmelCase__ : Dict = True lowerCAmelCase__ : Any = True lowerCAmelCase__ : Tuple = 99 lowerCAmelCase__ : Any = 32 lowerCAmelCase__ : Optional[int] = 2 lowerCAmelCase__ : List[Any] = 4 lowerCAmelCase__ : int = 37 lowerCAmelCase__ : str = 'gelu' lowerCAmelCase__ : Optional[int] = 0.1 lowerCAmelCase__ : List[str] = 0.1 lowerCAmelCase__ : Tuple = 512 lowerCAmelCase__ : Tuple = 16 lowerCAmelCase__ : int = 2 lowerCAmelCase__ : Dict = 0.0_2 lowerCAmelCase__ : str = 3 lowerCAmelCase__ : Optional[int] = 4 lowerCAmelCase__ : Optional[int] = None def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Dict = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : Optional[Any] = None if self.use_token_type_ids: lowerCAmelCase__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase__ : Dict = None lowerCAmelCase__ : Dict = None lowerCAmelCase__ : str = None if self.use_labels: lowerCAmelCase__ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : Union[str, Any] = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=a , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : Tuple , a : int , a : str , a : int , a : Optional[int] , a : List[Any] , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = TFRoFormerModel(config=a ) lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowerCAmelCase__ : List[Any] = [input_ids, input_mask] lowerCAmelCase__ : Dict = model(a ) lowerCAmelCase__ : Optional[Any] = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : Optional[int] , a : Optional[Any] , a : List[Any] , a : int , a : Tuple , a : Optional[int] , a : List[Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = True lowerCAmelCase__ : Optional[int] = TFRoFormerForCausalLM(config=a ) lowerCAmelCase__ : List[Any] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } lowerCAmelCase__ : int = model(a )['logits'] self.parent.assertListEqual( list(prediction_scores.numpy().shape ) , [self.batch_size, self.seq_length, self.vocab_size] ) def _lowerCamelCase ( self : Any , a : int , a : str , a : int , a : List[str] , a : str , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : str = TFRoFormerForMaskedLM(config=a ) lowerCAmelCase__ : Any = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } lowerCAmelCase__ : Dict = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[Any] , a : Tuple , a : int , a : List[str] , a : Dict , a : Optional[Any] , a : Optional[int] , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Tuple = TFRoFormerForSequenceClassification(config=a ) lowerCAmelCase__ : Union[str, Any] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } lowerCAmelCase__ : Tuple = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Optional[Any] , a : Tuple , a : Union[str, Any] , a : Tuple , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : str = self.num_choices lowerCAmelCase__ : int = TFRoFormerForMultipleChoice(config=a ) lowerCAmelCase__ : List[str] = tf.tile(tf.expand_dims(a , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase__ : List[str] = tf.tile(tf.expand_dims(a , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase__ : int = tf.tile(tf.expand_dims(a , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase__ : int = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, 'token_type_ids': multiple_choice_token_type_ids, } lowerCAmelCase__ : int = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Optional[Any] , a : Tuple , a : Union[str, Any] , a : List[Any] , a : Union[str, Any] , a : Any , a : int , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.num_labels lowerCAmelCase__ : Tuple = TFRoFormerForTokenClassification(config=a ) lowerCAmelCase__ : Union[str, Any] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } lowerCAmelCase__ : str = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : Any , a : Any , a : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : Union[str, Any] , a : Optional[int] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = TFRoFormerForQuestionAnswering(config=a ) lowerCAmelCase__ : Any = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } lowerCAmelCase__ : Tuple = model(a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : Union[str, Any] = config_and_inputs lowerCAmelCase__ : List[str] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( TFRoFormerModel, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerForMultipleChoice, ) if is_tf_available() else () ) lowercase = ( { 'feature-extraction': TFRoFormerModel, 'fill-mask': TFRoFormerForMaskedLM, 'question-answering': TFRoFormerForQuestionAnswering, 'text-classification': TFRoFormerForSequenceClassification, 'text-generation': TFRoFormerForCausalLM, 'token-classification': TFRoFormerForTokenClassification, 'zero-shot': TFRoFormerForSequenceClassification, } if is_tf_available() else {} ) lowercase = False lowercase = False def _lowerCamelCase ( self : Optional[Any] , a : Optional[int] , a : List[str] , a : str , a : Optional[int] , a : List[Any] ): '''simple docstring''' if pipeline_test_casse_name == "TextGenerationPipelineTests": return True return False def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = TFRoFormerModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , hidden_size=37 ) def _lowerCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*a ) def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*a ) def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*a ) @slow def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Dict = TFRoFormerModel.from_pretrained('junnyu/roformer_chinese_base' ) self.assertIsNotNone(a ) @require_tf class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = TFRoFormerForMaskedLM.from_pretrained('junnyu/roformer_chinese_base' ) lowerCAmelCase__ : int = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowerCAmelCase__ : int = model(a )[0] # TODO Replace vocab size lowerCAmelCase__ : Dict = 50_000 lowerCAmelCase__ : List[Any] = [1, 6, vocab_size] self.assertEqual(output.shape , a ) print(output[:, :3, :3] ) # TODO Replace values below with what was printed above. lowerCAmelCase__ : List[str] = tf.constant( [ [ [-0.1_2_0_5_3_3_4_1, -1.0_2_6_4_9_0_1, 0.2_9_2_2_1_9_4_6], [-1.5_1_3_3_7_8_3, 0.1_9_7_4_3_3, 0.1_5_1_9_0_6_0_7], [-5.0_1_3_5_4_0_3, -3.9_0_0_2_5_6, -0.8_4_0_3_8_7_6_4], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , a , atol=1E-4 ) @require_tf class A__ ( unittest.TestCase ): lowercase = 1E-4 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = tf.constant([[4, 10]] ) lowerCAmelCase__ : Tuple = TFRoFormerSinusoidalPositionalEmbedding(num_positions=6 , embedding_dim=6 ) lowerCAmelCase__ : Optional[Any] = emba(input_ids.shape ) lowerCAmelCase__ : Dict = tf.constant( [[0.0_0_0_0, 0.0_0_0_0, 0.0_0_0_0, 1.0_0_0_0, 1.0_0_0_0, 1.0_0_0_0], [0.8_4_1_5, 0.0_4_6_4, 0.0_0_2_2, 0.5_4_0_3, 0.9_9_8_9, 1.0_0_0_0]] ) tf.debugging.assert_near(a , a , atol=self.tolerance ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : int = tf.constant( [ [0.0_0_0_0, 0.0_0_0_0, 0.0_0_0_0, 0.0_0_0_0, 0.0_0_0_0], [0.8_4_1_5, 0.8_2_1_9, 0.8_0_2_0, 0.7_8_1_9, 0.7_6_1_7], [0.9_0_9_3, 0.9_3_6_4, 0.9_5_8_1, 0.9_7_4_9, 0.9_8_7_0], ] ) lowerCAmelCase__ : Optional[int] = TFRoFormerSinusoidalPositionalEmbedding(num_positions=512 , embedding_dim=512 ) emba([2, 16, 512] ) lowerCAmelCase__ : List[str] = emba.weight[:3, :5] tf.debugging.assert_near(a , a , atol=self.tolerance ) @require_tf class A__ ( unittest.TestCase ): lowercase = 1E-4 def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : int = tf.reshape(tf.range(2 * 12 * 16 * 64 , dtype=tf.floataa ) , shape=(2, 12, 16, 64) ) / 100 lowerCAmelCase__ : Union[str, Any] = -tf.reshape(tf.range(2 * 12 * 16 * 64 , dtype=tf.floataa ) , shape=(2, 12, 16, 64) ) / 100 lowerCAmelCase__ : str = TFRoFormerSinusoidalPositionalEmbedding(num_positions=32 , embedding_dim=64 ) lowerCAmelCase__ : Optional[int] = embed_positions([2, 16, 768] )[None, None, :, :] lowerCAmelCase__ , lowerCAmelCase__ : Dict = TFRoFormerSelfAttention.apply_rotary_position_embeddings( a , a , a ) lowerCAmelCase__ : Any = tf.constant( [ [0.0_0_0_0, 0.0_1_0_0, 0.0_2_0_0, 0.0_3_0_0, 0.0_4_0_0, 0.0_5_0_0, 0.0_6_0_0, 0.0_7_0_0], [-0.2_0_1_2, 0.8_8_9_7, 0.0_2_6_3, 0.9_4_0_1, 0.2_0_7_4, 0.9_4_6_3, 0.3_4_8_1, 0.9_3_4_3], [-1.7_0_5_7, 0.6_2_7_1, -1.2_1_4_5, 1.3_8_9_7, -0.6_3_0_3, 1.7_6_4_7, -0.1_1_7_3, 1.8_9_8_5], [-2.1_7_3_1, -1.6_3_9_7, -2.7_3_5_8, 0.2_8_5_4, -2.1_8_4_0, 1.7_1_8_3, -1.3_0_1_8, 2.4_8_7_1], [0.2_7_1_7, -3.6_1_7_3, -2.9_2_0_6, -2.1_9_8_8, -3.6_6_3_8, 0.3_8_5_8, -2.9_1_5_5, 2.2_9_8_0], [3.9_8_5_9, -2.1_5_8_0, -0.7_9_8_4, -4.4_9_0_4, -4.1_1_8_1, -2.0_2_5_2, -4.4_7_8_2, 1.1_2_5_3], ] ) lowerCAmelCase__ : Optional[int] = tf.constant( [ [0.0_0_0_0, -0.0_1_0_0, -0.0_2_0_0, -0.0_3_0_0, -0.0_4_0_0, -0.0_5_0_0, -0.0_6_0_0, -0.0_7_0_0], [0.2_0_1_2, -0.8_8_9_7, -0.0_2_6_3, -0.9_4_0_1, -0.2_0_7_4, -0.9_4_6_3, -0.3_4_8_1, -0.9_3_4_3], [1.7_0_5_7, -0.6_2_7_1, 1.2_1_4_5, -1.3_8_9_7, 0.6_3_0_3, -1.7_6_4_7, 0.1_1_7_3, -1.8_9_8_5], [2.1_7_3_1, 1.6_3_9_7, 2.7_3_5_8, -0.2_8_5_4, 2.1_8_4_0, -1.7_1_8_3, 1.3_0_1_8, -2.4_8_7_1], [-0.2_7_1_7, 3.6_1_7_3, 2.9_2_0_6, 2.1_9_8_8, 3.6_6_3_8, -0.3_8_5_8, 2.9_1_5_5, -2.2_9_8_0], [-3.9_8_5_9, 2.1_5_8_0, 0.7_9_8_4, 4.4_9_0_4, 4.1_1_8_1, 2.0_2_5_2, 4.4_7_8_2, -1.1_2_5_3], ] ) tf.debugging.assert_near(query_layer[0, 0, :6, :8] , a , atol=self.tolerance ) tf.debugging.assert_near(key_layer[0, 0, :6, :8] , a , atol=self.tolerance )
69
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
1
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } lowerCamelCase__ = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } lowerCamelCase__ = { """facebook/blenderbot_small-90M""": 512, } class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = BlenderbotSmallTokenizer def __init__( self : Union[str, Any] , a : Tuple=None , a : Optional[Any]=None , a : List[str]="<|endoftext|>" , a : Optional[int]="<|endoftext|>" , a : List[str]="<|endoftext|>" , a : List[str]=False , a : Union[str, Any]=True , **a : Any , ): '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=a , merges=a , add_prefix_space=a , trim_offsets=a , ) , bos_token=a , eos_token=a , unk_token=a , **a , ) lowerCAmelCase__ : Any = add_prefix_space def _lowerCamelCase ( self : Any , a : Dict , a : Union[str, Any]=None ): '''simple docstring''' lowerCAmelCase__ : List[Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowerCamelCase ( self : Any , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : List[Any] = [self.sep_token_id] lowerCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
69
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
1
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
69
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
1
import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets lowerCamelCase__ = """\ @inproceedings{kakwani2020indicnlpsuite, title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}}, author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar}, year={2020}, booktitle={Findings of EMNLP}, } """ lowerCamelCase__ = """\ IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te. """ lowerCamelCase__ = """ Compute IndicGLUE evaluation metric associated to each IndicGLUE dataset. Args: predictions: list of predictions to score (as int64), except for 'cvit-mkb-clsr' where each prediction is a vector (of float32). references: list of ground truth labels corresponding to the predictions (as int64), except for 'cvit-mkb-clsr' where each reference is a vector (of float32). Returns: depending on the IndicGLUE subset, one or several of: \"accuracy\": Accuracy \"f1\": F1 score \"precision\": Precision@10 Examples: >>> indic_glue_metric = datasets.load_metric('indic_glue', 'wnli') # 'wnli' or any of [\"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> indic_glue_metric = datasets.load_metric('indic_glue', 'wiki-ner') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> indic_glue_metric = datasets.load_metric('indic_glue', 'cvit-mkb-clsr') >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'precision@10': 1.0} """ def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: return float((preds == labels).mean() ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> List[Any]: lowerCAmelCase__ : Tuple = simple_accuracy(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : List[Any] = float(fa_score(y_true=SCREAMING_SNAKE_CASE_ , y_pred=SCREAMING_SNAKE_CASE_ ) ) return { "accuracy": acc, "f1": fa, } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : Dict = np.array(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = np.array(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = en_sentvecs.shape[0] # mean centering lowerCAmelCase__ : Any = en_sentvecs - np.mean(SCREAMING_SNAKE_CASE_ , axis=0 ) lowerCAmelCase__ : Union[str, Any] = in_sentvecs - np.mean(SCREAMING_SNAKE_CASE_ , axis=0 ) lowerCAmelCase__ : Tuple = cdist(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'cosine' ) lowerCAmelCase__ : Any = np.array(range(SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase__ : Optional[int] = sim.argsort(axis=1 )[:, :10] lowerCAmelCase__ : int = np.any(preds == actual[:, None] , axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( 'You should supply a configuration name selected in ' '["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ' '"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ' '"wiki-ner"]' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('int64' ) if self.config_name != 'cvit-mkb-clsr' else datasets.Sequence(datasets.Value('float32' ) ), 'references': datasets.Value('int64' ) if self.config_name != 'cvit-mkb-clsr' else datasets.Sequence(datasets.Value('float32' ) ), } ) , codebase_urls=[] , reference_urls=[] , format='numpy' if self.config_name != 'cvit-mkb-clsr' else None , ) def _lowerCamelCase ( self : List[Any] , a : Optional[int] , a : int ): '''simple docstring''' if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(a , a )} elif self.config_name in ["wiki-ner"]: return acc_and_fa(a , a ) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(a , a )} else: raise KeyError( 'You should supply a configuration name selected in ' '["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ' '"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ' '"wiki-ner"]' )
69
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
1
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml lowerCamelCase__ = NewType("""DataClass""", Any) lowerCamelCase__ = NewType("""DataClassType""", Any) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[Any]: if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F'''Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).''' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Callable[[str], Any]: lowerCAmelCase__ : Tuple = {str(SCREAMING_SNAKE_CASE_ ): choice for choice in choices} return lambda SCREAMING_SNAKE_CASE_ : str_to_choice.get(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( *, SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = dataclasses.MISSING , SCREAMING_SNAKE_CASE_ = dataclasses.MISSING , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls lowerCAmelCase__ : Optional[Any] = {} if aliases is not None: lowerCAmelCase__ : Dict = aliases if help is not None: lowerCAmelCase__ : int = help return dataclasses.field(metadata=SCREAMING_SNAKE_CASE_ , default=SCREAMING_SNAKE_CASE_ , default_factory=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) class A__ ( __magic_name__ ): lowercase = 42 def __init__( self : Optional[Any] , a : Union[DataClassType, Iterable[DataClassType]] , **a : Optional[Any] ): '''simple docstring''' if "formatter_class" not in kwargs: lowerCAmelCase__ : Any = ArgumentDefaultsHelpFormatter super().__init__(**a ) if dataclasses.is_dataclass(a ): lowerCAmelCase__ : Tuple = [dataclass_types] lowerCAmelCase__ : Optional[int] = list(a ) for dtype in self.dataclass_types: self._add_dataclass_arguments(a ) @staticmethod def _lowerCamelCase ( a : ArgumentParser , a : dataclasses.Field ): '''simple docstring''' lowerCAmelCase__ : Tuple = f'''--{field.name}''' lowerCAmelCase__ : Union[str, Any] = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , a ): raise RuntimeError( 'Unresolved type detected, which should have been done with the help of ' '`typing.get_type_hints` method by default' ) lowerCAmelCase__ : Dict = kwargs.pop('aliases' , [] ) if isinstance(a , a ): lowerCAmelCase__ : Union[str, Any] = [aliases] lowerCAmelCase__ : Union[str, Any] = getattr(field.type , '__origin__' , field.type ) if origin_type is Union or (hasattr(a , 'UnionType' ) and isinstance(a , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(a ) not in field.type.__args__ ): raise ValueError( 'Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because' ' the argument parser only supports one type per argument.' f''' Problem encountered in field \'{field.name}\'.''' ) if type(a ) not in field.type.__args__: # filter `str` in Union lowerCAmelCase__ : Dict = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] lowerCAmelCase__ : Optional[Any] = getattr(field.type , '__origin__' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) lowerCAmelCase__ : Optional[int] = ( field.type.__args__[0] if isinstance(a , field.type.__args__[1] ) else field.type.__args__[1] ) lowerCAmelCase__ : Optional[Any] = getattr(field.type , '__origin__' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) lowerCAmelCase__ : Any = {} if origin_type is Literal or (isinstance(field.type , a ) and issubclass(field.type , a )): if origin_type is Literal: lowerCAmelCase__ : Optional[Any] = field.type.__args__ else: lowerCAmelCase__ : List[str] = [x.value for x in field.type] lowerCAmelCase__ : Optional[int] = make_choice_type_function(kwargs['choices'] ) if field.default is not dataclasses.MISSING: lowerCAmelCase__ : Union[str, Any] = field.default else: lowerCAmelCase__ : Union[str, Any] = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument lowerCAmelCase__ : List[str] = copy(a ) # Hack because type=bool in argparse does not behave as we want. lowerCAmelCase__ : Any = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. lowerCAmelCase__ : Tuple = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way lowerCAmelCase__ : List[Any] = default # This tells argparse we accept 0 or 1 value after --field_name lowerCAmelCase__ : List[str] = '?' # This is the value that will get picked if we do --field_name (without value) lowerCAmelCase__ : List[Any] = True elif isclass(a ) and issubclass(a , a ): lowerCAmelCase__ : str = field.type.__args__[0] lowerCAmelCase__ : List[Any] = '+' if field.default_factory is not dataclasses.MISSING: lowerCAmelCase__ : Optional[int] = field.default_factory() elif field.default is dataclasses.MISSING: lowerCAmelCase__ : Any = True else: lowerCAmelCase__ : List[Any] = field.type if field.default is not dataclasses.MISSING: lowerCAmelCase__ : Tuple = field.default elif field.default_factory is not dataclasses.MISSING: lowerCAmelCase__ : Dict = field.default_factory() else: lowerCAmelCase__ : Dict = True parser.add_argument(a , *a , **a ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): lowerCAmelCase__ : str = False parser.add_argument(f'''--no_{field.name}''' , action='store_false' , dest=field.name , **a ) def _lowerCamelCase ( self : Union[str, Any] , a : DataClassType ): '''simple docstring''' if hasattr(a , '_argument_group_name' ): lowerCAmelCase__ : Union[str, Any] = self.add_argument_group(dtype._argument_group_name ) else: lowerCAmelCase__ : Optional[Any] = self try: lowerCAmelCase__ : Dict[str, type] = get_type_hints(a ) except NameError: raise RuntimeError( f'''Type resolution failed for {dtype}. Try declaring the class in global scope or ''' 'removing line of `from __future__ import annotations` which opts in Postponed ' 'Evaluation of Annotations (PEP 563)' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(a ): lowerCAmelCase__ : List[str] = '.'.join(map(a , sys.version_info[:3] ) ) raise RuntimeError( f'''Type resolution failed for {dtype} on Python {python_version}. Try removing ''' 'line of `from __future__ import annotations` which opts in union types as ' '`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ' 'support Python versions that lower than 3.10, you need to use ' '`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ' '`X | None`.' ) from ex raise for field in dataclasses.fields(a ): if not field.init: continue lowerCAmelCase__ : Dict = type_hints[field.name] self._parse_dataclass_field(a , a ) def _lowerCamelCase ( self : List[Any] , a : Tuple=None , a : Dict=False , a : Optional[Any]=True , a : List[Any]=None , a : int=None , ): '''simple docstring''' if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): lowerCAmelCase__ : List[str] = [] if args_filename: args_files.append(Path(a ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('.args' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values lowerCAmelCase__ : Tuple = ArgumentParser() args_file_parser.add_argument(a , type=a , action='append' ) # Use only remaining args for further parsing (remove the args_file_flag) lowerCAmelCase__ , lowerCAmelCase__ : Dict = args_file_parser.parse_known_args(args=a ) lowerCAmelCase__ : Optional[int] = vars(a ).get(args_file_flag.lstrip('-' ) , a ) if cmd_args_file_paths: args_files.extend([Path(a ) for p in cmd_args_file_paths] ) lowerCAmelCase__ : Tuple = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last lowerCAmelCase__ : int = file_args + args if args is not None else file_args + sys.argv[1:] lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.parse_known_args(args=a ) lowerCAmelCase__ : Union[str, Any] = [] for dtype in self.dataclass_types: lowerCAmelCase__ : Union[str, Any] = {f.name for f in dataclasses.fields(a ) if f.init} lowerCAmelCase__ : Optional[int] = {k: v for k, v in vars(a ).items() if k in keys} for k in keys: delattr(a , a ) lowerCAmelCase__ : Optional[int] = dtype(**a ) outputs.append(a ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(a ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f'''Some specified arguments are not used by the HfArgumentParser: {remaining_args}''' ) return (*outputs,) def _lowerCamelCase ( self : List[Any] , a : Dict[str, Any] , a : bool = False ): '''simple docstring''' lowerCAmelCase__ : int = set(args.keys() ) lowerCAmelCase__ : List[Any] = [] for dtype in self.dataclass_types: lowerCAmelCase__ : List[str] = {f.name for f in dataclasses.fields(a ) if f.init} lowerCAmelCase__ : str = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) lowerCAmelCase__ : str = dtype(**a ) outputs.append(a ) if not allow_extra_keys and unused_keys: raise ValueError(f'''Some keys are not used by the HfArgumentParser: {sorted(a )}''' ) return tuple(a ) def _lowerCamelCase ( self : Tuple , a : str , a : bool = False ): '''simple docstring''' with open(Path(a ) , encoding='utf-8' ) as open_json_file: lowerCAmelCase__ : int = json.loads(open_json_file.read() ) lowerCAmelCase__ : List[Any] = self.parse_dict(a , allow_extra_keys=a ) return tuple(a ) def _lowerCamelCase ( self : List[str] , a : str , a : bool = False ): '''simple docstring''' lowerCAmelCase__ : str = self.parse_dict(yaml.safe_load(Path(a ).read_text() ) , allow_extra_keys=a ) return tuple(a )
69
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
69
1
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class A__ : def __init__( self : List[Any] , a : Optional[int] , a : List[str]=12 , a : List[Any]=7 , a : Dict=True , a : str=True , a : Union[str, Any]=True , a : Optional[int]=99 , a : str=32 , a : List[Any]=32 , a : Optional[Any]=2 , a : Any=4 , a : List[str]=37 , a : List[Any]=0.1 , a : Any=0.1 , a : str=512 , a : Optional[int]=0.0_2 , a : Tuple=0 , a : Dict=None , ): '''simple docstring''' lowerCAmelCase__ : List[Any] = parent lowerCAmelCase__ : Tuple = batch_size lowerCAmelCase__ : List[str] = seq_length lowerCAmelCase__ : str = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_labels lowerCAmelCase__ : Union[str, Any] = vocab_size lowerCAmelCase__ : int = hidden_size lowerCAmelCase__ : Optional[Any] = projection_dim lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : Dict = num_attention_heads lowerCAmelCase__ : Union[str, Any] = intermediate_size lowerCAmelCase__ : Optional[Any] = dropout lowerCAmelCase__ : Tuple = attention_dropout lowerCAmelCase__ : List[str] = max_position_embeddings lowerCAmelCase__ : Union[str, Any] = initializer_range lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : List[Any] = bos_token_id def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Optional[int] = None if self.use_input_mask: lowerCAmelCase__ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowerCAmelCase__ : Any = input_mask.numpy() lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = input_mask.shape lowerCAmelCase__ : Union[str, Any] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a ): lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[Any] = 0 lowerCAmelCase__ : str = self.get_config() return config, input_ids, tf.convert_to_tensor(a ) def _lowerCamelCase ( self : str ): '''simple docstring''' return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _lowerCamelCase ( self : List[str] , a : str , a : Dict , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = TFBlipTextModel(config=a ) lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , training=a ) lowerCAmelCase__ : str = model(a , training=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : int = config_and_inputs lowerCAmelCase__ : str = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class A__ ( __magic_name__ , unittest.TestCase ): lowercase = (TFBlipTextModel,) if is_tf_available() else () lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Dict = BlipTextModelTester(self ) lowerCAmelCase__ : Tuple = ConfigTester(self , config_class=a , hidden_size=37 ) def _lowerCamelCase ( self : str ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' pass def _lowerCamelCase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Union[str, Any] = TFBlipTextModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Union[str, Any] , a : Any=True ): '''simple docstring''' super().test_pt_tf_model_equivalence(allow_missing_keys=a )
69
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __magic_name__ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'LayoutLMv3ImageProcessor' lowercase = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self : Optional[int] , a : Union[str, Any]=None , a : Optional[Any]=None , **a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , a , ) lowerCAmelCase__ : int = kwargs.pop('feature_extractor' ) lowerCAmelCase__ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(a , a ) def __call__( self : List[Any] , a : List[Any] , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a : Union[List[List[int]], List[List[List[int]]]] = None , a : Optional[Union[List[int], List[List[int]]]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : str , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor lowerCAmelCase__ : List[str] = self.image_processor(images=a , return_tensors=a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a , a ): lowerCAmelCase__ : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ : List[str] = features['words'] lowerCAmelCase__ : List[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_token_type_ids=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_length=a , verbose=a , return_tensors=a , **a , ) # add pixel values lowerCAmelCase__ : Tuple = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase__ : List[str] = self.get_overflowing_images(a , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase__ : List[str] = images return encoded_inputs def _lowerCamelCase ( self : Any , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a ) != len(a ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' f''' {len(a )} and {len(a )}''' ) return images_with_overflow def _lowerCamelCase ( self : Union[str, Any] , *a : Optional[Any] , **a : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Tuple , *a : List[str] , **a : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*a , **a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , a , ) return self.image_processor_class @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , a , ) return self.image_processor
69
1
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
69
1
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: assert ( isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) and number_of_steps > 0 ), F'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 lowerCAmelCase__ , lowerCAmelCase__ : Any = 1, 1 for _ in range(number_of_steps - 1 ): lowerCAmelCase__ , lowerCAmelCase__ : Dict = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
69
lowerCamelCase__ = """Alexander Joslin""" import operator as op from .stack import Stack def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} lowerCAmelCase__ : Stack[int] = Stack() lowerCAmelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE_ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE_ ) elif i == ")": # RULE 4 lowerCAmelCase__ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : List[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : Tuple = operators[opr](SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) operand_stack.push(SCREAMING_SNAKE_CASE_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
69
1
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class A__ : def __init__( self : int , a : str = "cpu" , a : str = "openai/clip-vit-large-patch14" ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = device lowerCAmelCase__ : Any = CLIPTokenizerFast.from_pretrained(a ) lowerCAmelCase__ : Tuple = [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] lowerCAmelCase__ : Optional[Any] = [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] lowerCAmelCase__ : List[str] = torchvision.transforms.Normalize(self.image_mean , self.image_std ) lowerCAmelCase__ : str = torchvision.transforms.Resize(224 ) lowerCAmelCase__ : Union[str, Any] = torchvision.transforms.CenterCrop(224 ) def _lowerCamelCase ( self : int , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.resize(a ) lowerCAmelCase__ : Tuple = self.center_crop(a ) lowerCAmelCase__ : Any = self.normalize(a ) return images def __call__( self : Union[str, Any] , a : Optional[int]=None , a : Any=None , **a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.tokenizer(text=a , **a ) lowerCAmelCase__ : Optional[int] = self.preprocess_img(a ) lowerCAmelCase__ : Any = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class A__ ( nn.Module ): def __init__( self : Union[str, Any] , a : List[Any]=10 , a : Optional[int]=0.0_1 , a : List[Any]=None , a : Optional[int]=None , a : str=None , a : Optional[Any]=None , a : Union[str, Any]=None , a : List[Any]=None , a : Dict=False , a : Optional[int]=True , a : Optional[int]="image" , a : Tuple=True , a : int=False , a : Optional[int]=False , a : int=False , ): '''simple docstring''' super().__init__() lowerCAmelCase__ : str = None lowerCAmelCase__ : Optional[int] = device if device else get_device() if vqgan: lowerCAmelCase__ : Any = vqgan else: lowerCAmelCase__ : Dict = load_vqgan(self.device , conf_path=a , ckpt_path=a ) self.vqgan.eval() if clip: lowerCAmelCase__ : Dict = clip else: lowerCAmelCase__ : Optional[Any] = CLIPModel.from_pretrained('openai/clip-vit-base-patch32' ) self.clip.to(self.device ) lowerCAmelCase__ : Any = ProcessorGradientFlow(device=self.device ) lowerCAmelCase__ : Any = iterations lowerCAmelCase__ : Union[str, Any] = lr lowerCAmelCase__ : Tuple = log lowerCAmelCase__ : int = make_grid lowerCAmelCase__ : Optional[Any] = return_val lowerCAmelCase__ : List[Any] = quantize lowerCAmelCase__ : Optional[Any] = self.vqgan.decoder.z_shape def _lowerCamelCase ( self : str , a : List[Any]=None , a : List[Any]=None , a : int=5 , a : Dict=True ): '''simple docstring''' lowerCAmelCase__ : Any = [] if output_path is None: lowerCAmelCase__ : Optional[int] = './animation.gif' if input_path is None: lowerCAmelCase__ : List[str] = self.save_path lowerCAmelCase__ : Tuple = sorted(glob(input_path + '/*' ) ) if not len(a ): raise ValueError( 'No images found in save path, aborting (did you pass save_intermediate=True to the generate' ' function?)' ) if len(a ) == 1: print('Only one image found in save path, (did you pass save_intermediate=True to the generate function?)' ) lowerCAmelCase__ : List[Any] = total_duration / len(a ) lowerCAmelCase__ : Union[str, Any] = [frame_duration] * len(a ) if extend_frames: lowerCAmelCase__ : Union[str, Any] = 1.5 lowerCAmelCase__ : Union[str, Any] = 3 for file_name in paths: if file_name.endswith('.png' ): images.append(imageio.imread(a ) ) imageio.mimsave(a , a , duration=a ) print(f'''gif saved to {output_path}''' ) def _lowerCamelCase ( self : str , a : Dict=None , a : List[str]=None ): '''simple docstring''' if not (path or img): raise ValueError('Input either path or tensor' ) if img is not None: raise NotImplementedError lowerCAmelCase__ : Any = preprocess(Image.open(a ) , target_image_size=256 ).to(self.device ) lowerCAmelCase__ : str = preprocess_vqgan(a ) lowerCAmelCase__ , *lowerCAmelCase__ : List[Any] = self.vqgan.encode(a ) return z def _lowerCamelCase ( self : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.latent.detach().requires_grad_() lowerCAmelCase__ : Optional[int] = base_latent + transform_vector if self.quantize: lowerCAmelCase__ , *lowerCAmelCase__ : Dict = self.vqgan.quantize(a ) else: lowerCAmelCase__ : int = trans_latent return self.vqgan.decode(a ) def _lowerCamelCase ( self : Dict , a : Any , a : Any , a : List[Any]=None ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.clip_preprocessor(text=a , images=a , return_tensors='pt' , padding=a ) lowerCAmelCase__ : int = self.clip(**a ) lowerCAmelCase__ : Optional[int] = clip_outputs.logits_per_image if weights is not None: lowerCAmelCase__ : List[str] = similarity_logits * weights return similarity_logits.sum() def _lowerCamelCase ( self : List[Any] , a : Optional[Any] , a : Any , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self._get_clip_similarity(pos_prompts['prompts'] , a , weights=(1 / pos_prompts['weights']) ) if neg_prompts: lowerCAmelCase__ : Optional[Any] = self._get_clip_similarity(neg_prompts['prompts'] , a , weights=neg_prompts['weights'] ) else: lowerCAmelCase__ : str = torch.tensor([1] , device=self.device ) lowerCAmelCase__ : Dict = -torch.log(a ) + torch.log(a ) return loss def _lowerCamelCase ( self : Union[str, Any] , a : str , a : Dict , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = torch.randn_like(self.latent , requires_grad=a , device=self.device ) lowerCAmelCase__ : Tuple = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() lowerCAmelCase__ : Optional[Any] = self._add_vector(a ) lowerCAmelCase__ : List[Any] = loop_post_process(a ) lowerCAmelCase__ : Union[str, Any] = self._get_CLIP_loss(a , a , a ) print('CLIP loss' , a ) if self.log: wandb.log({'CLIP Loss': clip_loss} ) clip_loss.backward(retain_graph=a ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def _lowerCamelCase ( self : Optional[int] , a : Dict , a : List[Any] , a : Union[str, Any] ): '''simple docstring''' wandb.init(reinit=a , project='face-editor' ) wandb.config.update({'Positive Prompts': positive_prompts} ) wandb.config.update({'Negative Prompts': negative_prompts} ) wandb.config.update({'lr': self.lr, 'iterations': self.iterations} ) if image_path: lowerCAmelCase__ : Optional[Any] = Image.open(a ) lowerCAmelCase__ : Dict = image.resize((256, 256) ) wandb.log('Original Image' , wandb.Image(a ) ) def _lowerCamelCase ( self : Union[str, Any] , a : List[Any] ): '''simple docstring''' if not prompts: return [] lowerCAmelCase__ : Tuple = [] lowerCAmelCase__ : int = [] if isinstance(a , a ): lowerCAmelCase__ : Tuple = [prompt.strip() for prompt in prompts.split('|' )] for prompt in prompts: if isinstance(a , (tuple, list) ): lowerCAmelCase__ : Any = prompt[0] lowerCAmelCase__ : List[str] = float(prompt[1] ) elif ":" in prompt: lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = prompt.split(':' ) lowerCAmelCase__ : Optional[Any] = float(a ) else: lowerCAmelCase__ : Optional[int] = prompt lowerCAmelCase__ : Optional[Any] = 1.0 processed_prompts.append(a ) weights.append(a ) return { "prompts": processed_prompts, "weights": torch.tensor(a , device=self.device ), } def _lowerCamelCase ( self : Tuple , a : Tuple , a : Any=None , a : List[Any]=None , a : Union[str, Any]=True , a : int=False , a : Optional[Any]=True , a : Optional[Any]=True , a : Any=None , ): '''simple docstring''' if image_path: lowerCAmelCase__ : Dict = self._get_latent(a ) else: lowerCAmelCase__ : Tuple = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(a , a , a ) assert pos_prompts, "You must provide at least one positive prompt." lowerCAmelCase__ : Union[str, Any] = self.process_prompts(a ) lowerCAmelCase__ : Optional[Any] = self.process_prompts(a ) if save_final and save_path is None: lowerCAmelCase__ : Optional[int] = os.path.join('./outputs/' , '_'.join(pos_prompts['prompts'] ) ) if not os.path.exists(a ): os.makedirs(a ) else: lowerCAmelCase__ : Optional[int] = save_path + '_' + get_timestamp() os.makedirs(a ) lowerCAmelCase__ : Tuple = save_path lowerCAmelCase__ : str = self.vqgan.decode(self.latent )[0] if show_intermediate: print('Original Image' ) show_pil(custom_to_pil(a ) ) lowerCAmelCase__ : Tuple = loop_post_process(a ) for iter, transformed_img in enumerate(self._optimize_CLIP(a , a , a ) ): if show_intermediate: show_pil(a ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , f'''iter_{iter:03d}.png''' ) ) if self.log: wandb.log({'Image': wandb.Image(a )} ) if show_final: show_pil(a ) if save_final: transformed_img.save(os.path.join(self.save_path , f'''iter_{iter:03d}_final.png''' ) )
69
import numpy class A__ : def __init__( self : Tuple , a : numpy.ndarray , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : int = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Dict = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : List[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowerCAmelCase__ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : List[Any] = numpy.zeros(output_array.shape ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowerCAmelCase__ : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowerCAmelCase__ : int = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowerCamelCase ( self : Optional[int] , a : numpy.ndarray , a : int , a : bool ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): lowerCAmelCase__ : Any = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Tuple = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def _lowerCamelCase ( self : Optional[Any] , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : Dict = input_arr lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return (value) * (1 - (value)) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : List[str] = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ , output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ , iterations=10 , give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
69
1
import numpy as np def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> np.array: return 1 / (1 + np.exp(-vector )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> np.array: return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
69
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A__ : def __init__( self : int , a : str , a : Union[str, Any]=13 , a : int=32 , a : Optional[Any]=2 , a : Tuple=3 , a : List[Any]=16 , a : List[str]=[1, 2, 1] , a : int=[2, 2, 4] , a : int=2 , a : Optional[Any]=2.0 , a : Optional[int]=True , a : Dict=0.0 , a : Any=0.0 , a : int=0.1 , a : List[str]="gelu" , a : Optional[Any]=False , a : str=True , a : Dict=0.0_2 , a : Any=1E-5 , a : Optional[int]=True , a : str=None , a : str=True , a : int=10 , a : str=8 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : Optional[Any] = patch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = embed_dim lowerCAmelCase__ : Tuple = depths lowerCAmelCase__ : List[str] = num_heads lowerCAmelCase__ : List[Any] = window_size lowerCAmelCase__ : Any = mlp_ratio lowerCAmelCase__ : Optional[Any] = qkv_bias lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : int = drop_path_rate lowerCAmelCase__ : Optional[Any] = hidden_act lowerCAmelCase__ : int = use_absolute_embeddings lowerCAmelCase__ : List[str] = patch_norm lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Dict = use_labels lowerCAmelCase__ : List[Any] = type_sequence_label_size lowerCAmelCase__ : Optional[Any] = encoder_stride def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = None if self.use_labels: lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : int = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : List[str] , a : Any , a : str , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = SwinvaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model(a ) lowerCAmelCase__ : str = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase__ : Tuple = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SwinvaForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : Dict = SwinvaForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Union[str, Any] , a : int , a : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.type_sequence_label_size lowerCAmelCase__ : List[Any] = SwinvaForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = config_and_inputs lowerCAmelCase__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = SwinvaModelTester(self ) lowerCAmelCase__ : int = ConfigTester(self , config_class=a , embed_dim=37 ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[int] = model_class(a ) lowerCAmelCase__ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Tuple = [*signature.parameters.keys()] lowerCAmelCase__ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = True for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = True lowerCAmelCase__ : str = False lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Dict = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : int = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Dict = outputs.attentions lowerCAmelCase__ : Dict = len(self.model_tester.depths ) self.assertEqual(len(a ) , a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Optional[int] = config.window_size**2 lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowerCAmelCase__ : Tuple = len(a ) # Check attention is always last and order is fine lowerCAmelCase__ : str = True lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(**self._prepare_for_class(a , a ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): lowerCAmelCase__ : Optional[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase__ : Any = 2 self.assertEqual(out_len + added_hidden_states , len(a ) ) lowerCAmelCase__ : Dict = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _lowerCamelCase ( self : int , a : Optional[int] , a : int , a : Optional[Any] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.hidden_states lowerCAmelCase__ : str = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a ) , a ) # Swinv2 has a different seq_length lowerCAmelCase__ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowerCAmelCase__ : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(a ) , a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = reshaped_hidden_states[0].shape lowerCAmelCase__ : List[str] = ( reshaped_hidden_states[0].view(a , a , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : str = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : List[str] = True self.check_hidden_states_output(a , a , a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Any = 3 lowerCAmelCase__ : int = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase__ : str = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : str = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase__ : Any = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase__ : str = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[str] = SwinvaModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = _config_zero_init(a ) for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(config=a ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( a ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Any = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**a ) # verify the logits lowerCAmelCase__ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Optional[Any] = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) )
69
1
from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
69
from itertools import permutations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase__ : str = [7, 11, 13, 17] for i, test in enumerate(SCREAMING_SNAKE_CASE_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 10 ) -> int: return sum( int(''.join(map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) ) for num in permutations(range(SCREAMING_SNAKE_CASE_ ) ) if is_substring_divisible(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
69
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """microsoft/deberta-v2-xlarge""": """https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json""", """microsoft/deberta-v2-xxlarge""": """https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json""", """microsoft/deberta-v2-xlarge-mnli""": ( """https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json""" ), """microsoft/deberta-v2-xxlarge-mnli""": ( """https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json""" ), } class A__ ( __magic_name__ ): lowercase = 'deberta-v2' def __init__( self : Dict , a : Any=128_100 , a : List[Any]=1_536 , a : List[str]=24 , a : List[Any]=24 , a : int=6_144 , a : List[Any]="gelu" , a : int=0.1 , a : Union[str, Any]=0.1 , a : Dict=512 , a : Optional[int]=0 , a : Tuple=0.0_2 , a : Optional[int]=1E-7 , a : Optional[int]=False , a : Tuple=-1 , a : Dict=0 , a : Dict=True , a : Optional[int]=None , a : Optional[Any]=0 , a : Optional[int]="gelu" , **a : Tuple , ): '''simple docstring''' super().__init__(**a ) lowerCAmelCase__ : int = hidden_size lowerCAmelCase__ : List[str] = num_hidden_layers lowerCAmelCase__ : List[Any] = num_attention_heads lowerCAmelCase__ : Any = intermediate_size lowerCAmelCase__ : str = hidden_act lowerCAmelCase__ : List[Any] = hidden_dropout_prob lowerCAmelCase__ : List[str] = attention_probs_dropout_prob lowerCAmelCase__ : List[str] = max_position_embeddings lowerCAmelCase__ : Union[str, Any] = type_vocab_size lowerCAmelCase__ : Optional[Any] = initializer_range lowerCAmelCase__ : List[Any] = relative_attention lowerCAmelCase__ : Optional[Any] = max_relative_positions lowerCAmelCase__ : Union[str, Any] = pad_token_id lowerCAmelCase__ : str = position_biased_input # Backwards compatibility if type(a ) == str: lowerCAmelCase__ : Optional[Any] = [x.strip() for x in pos_att_type.lower().split('|' )] lowerCAmelCase__ : Dict = pos_att_type lowerCAmelCase__ : Tuple = vocab_size lowerCAmelCase__ : Dict = layer_norm_eps lowerCAmelCase__ : str = kwargs.get('pooler_hidden_size' , a ) lowerCAmelCase__ : Any = pooler_dropout lowerCAmelCase__ : Dict = pooler_hidden_act class A__ ( __magic_name__ ): @property def _lowerCamelCase ( self : Any ): '''simple docstring''' if self.task == "multiple-choice": lowerCAmelCase__ : Optional[Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: lowerCAmelCase__ : int = {0: 'batch', 1: 'sequence'} if self._config.type_vocab_size > 0: return OrderedDict( [('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis)] ) else: return OrderedDict([('input_ids', dynamic_axis), ('attention_mask', dynamic_axis)] ) @property def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return 12 def _lowerCamelCase ( self : Optional[Any] , a : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , a : int = -1 , a : int = -1 , a : int = -1 , a : bool = False , a : Optional["TensorType"] = None , a : int = 3 , a : int = 40 , a : int = 40 , a : "PreTrainedTokenizerBase" = None , ): '''simple docstring''' lowerCAmelCase__ : List[str] = super().generate_dummy_inputs(preprocessor=a , framework=a ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
69
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
69
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """microsoft/focalnet-tiny""": """https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json""", } class A__ ( __magic_name__ , __magic_name__ ): lowercase = 'focalnet' def __init__( self : List[str] , a : str=224 , a : List[Any]=4 , a : str=3 , a : List[str]=96 , a : str=False , a : List[str]=[192, 384, 768, 768] , a : Optional[int]=[2, 2, 6, 2] , a : List[Any]=[2, 2, 2, 2] , a : List[str]=[3, 3, 3, 3] , a : Dict="gelu" , a : Any=4.0 , a : Dict=0.0 , a : List[str]=0.1 , a : str=False , a : str=1E-4 , a : str=False , a : Optional[int]=False , a : Any=False , a : Dict=0.0_2 , a : str=1E-5 , a : int=32 , a : Union[str, Any]=None , a : Optional[int]=None , **a : Optional[Any] , ): '''simple docstring''' super().__init__(**a ) lowerCAmelCase__ : int = image_size lowerCAmelCase__ : List[str] = patch_size lowerCAmelCase__ : int = num_channels lowerCAmelCase__ : Any = embed_dim lowerCAmelCase__ : str = use_conv_embed lowerCAmelCase__ : int = hidden_sizes lowerCAmelCase__ : Any = depths lowerCAmelCase__ : str = focal_levels lowerCAmelCase__ : Tuple = focal_windows lowerCAmelCase__ : Union[str, Any] = hidden_act lowerCAmelCase__ : Any = mlp_ratio lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : str = drop_path_rate lowerCAmelCase__ : Any = use_layerscale lowerCAmelCase__ : List[str] = layerscale_value lowerCAmelCase__ : str = use_post_layernorm lowerCAmelCase__ : Optional[Any] = use_post_layernorm_in_modulation lowerCAmelCase__ : List[Any] = normalize_modulator lowerCAmelCase__ : str = initializer_range lowerCAmelCase__ : Tuple = layer_norm_eps lowerCAmelCase__ : Any = encoder_stride lowerCAmelCase__ : Any = ['stem'] + [f'''stage{idx}''' for idx in range(1 , len(self.depths ) + 1 )] lowerCAmelCase__ , lowerCAmelCase__ : Tuple = get_aligned_output_features_output_indices( out_features=a , out_indices=a , stage_names=self.stage_names )
69
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ , lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
69
1
from __future__ import annotations from random import random from typing import Generic, TypeVar lowerCamelCase__ = TypeVar("""KT""") lowerCamelCase__ = TypeVar("""VT""") class A__ ( Generic[KT, VT] ): def __init__( self : List[Any] , a : KT | str = "root" , a : VT | None = None ): '''simple docstring''' lowerCAmelCase__ : List[str] = key lowerCAmelCase__ : Tuple = value lowerCAmelCase__ : list[Node[KT, VT]] = [] def __repr__( self : Dict ): '''simple docstring''' return f'''Node({self.key}: {self.value})''' @property def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' return len(self.forward ) class A__ ( Generic[KT, VT] ): def __init__( self : List[str] , a : float = 0.5 , a : int = 16 ): '''simple docstring''' lowerCAmelCase__ : Node[KT, VT] = Node[KT, VT]() lowerCAmelCase__ : Tuple = 0 lowerCAmelCase__ : Tuple = p lowerCAmelCase__ : List[Any] = max_level def __str__( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = list(self ) if len(a ) == 0: return f'''SkipList(level={self.level})''' lowerCAmelCase__ : List[str] = max((len(str(a ) ) for item in items) , default=4 ) lowerCAmelCase__ : List[Any] = max(a , 4 ) + 4 lowerCAmelCase__ : List[str] = self.head lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : Optional[Any] = node.forward.copy() lines.append(f'''[{node.key}]'''.ljust(a , '-' ) + '* ' * len(a ) ) lines.append(' ' * label_size + '| ' * len(a ) ) while len(node.forward ) != 0: lowerCAmelCase__ : Tuple = node.forward[0] lines.append( f'''[{node.key}]'''.ljust(a , '-' ) + ' '.join(str(n.key ) if n.key == node.key else '|' for n in forwards ) ) lines.append(' ' * label_size + '| ' * len(a ) ) lowerCAmelCase__ : Tuple = node.forward lines.append('None'.ljust(a ) + '* ' * len(a ) ) return f'''SkipList(level={self.level})\n''' + "\n".join(a ) def __iter__( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = self.head while len(node.forward ) != 0: yield node.forward[0].key lowerCAmelCase__ : List[Any] = node.forward[0] def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Any = 1 while random() < self.p and level < self.max_level: level += 1 return level def _lowerCamelCase ( self : Dict , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : int = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: lowerCAmelCase__ : List[str] = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(a ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def _lowerCamelCase ( self : List[str] , a : KT ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = self._locate_node(a ) if node is not None: for i, update_node in enumerate(a ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: lowerCAmelCase__ : str = node.forward[i] else: lowerCAmelCase__ : List[str] = update_node.forward[:i] def _lowerCamelCase ( self : List[str] , a : KT , a : VT ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[str] = self._locate_node(a ) if node is not None: lowerCAmelCase__ : Union[str, Any] = value else: lowerCAmelCase__ : int = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , a ): update_vector.append(self.head ) lowerCAmelCase__ : List[Any] = level lowerCAmelCase__ : List[Any] = Node(a , a ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(a ) else: lowerCAmelCase__ : Any = new_node def _lowerCamelCase ( self : List[str] , a : VT ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[str] = self._locate_node(a ) if node is not None: return node.value return None def lowerCAmelCase__ ( ) -> Dict: lowerCAmelCase__ : Tuple = SkipList() skip_list.insert('Key1' , 3 ) skip_list.insert('Key2' , 12 ) skip_list.insert('Key3' , 41 ) skip_list.insert('Key4' , -19 ) lowerCAmelCase__ : Optional[int] = skip_list.head lowerCAmelCase__ : Optional[Any] = {} while node.level != 0: lowerCAmelCase__ : int = node.forward[0] lowerCAmelCase__ : Any = node.value assert len(SCREAMING_SNAKE_CASE_ ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def lowerCAmelCase__ ( ) -> Optional[int]: lowerCAmelCase__ : int = SkipList() skip_list.insert('Key1' , 10 ) skip_list.insert('Key1' , 12 ) skip_list.insert('Key5' , 7 ) skip_list.insert('Key7' , 10 ) skip_list.insert('Key10' , 5 ) skip_list.insert('Key7' , 7 ) skip_list.insert('Key5' , 5 ) skip_list.insert('Key10' , 10 ) lowerCAmelCase__ : Union[str, Any] = skip_list.head lowerCAmelCase__ : Tuple = {} while node.level != 0: lowerCAmelCase__ : List[Any] = node.forward[0] lowerCAmelCase__ : int = node.value if len(SCREAMING_SNAKE_CASE_ ) != 4: print() assert len(SCREAMING_SNAKE_CASE_ ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def lowerCAmelCase__ ( ) -> Optional[int]: lowerCAmelCase__ : int = SkipList() assert skip_list.find('Some key' ) is None def lowerCAmelCase__ ( ) -> Tuple: lowerCAmelCase__ : Optional[int] = SkipList() skip_list.insert('Key2' , 20 ) assert skip_list.find('Key2' ) == 20 skip_list.insert('Some Key' , 10 ) skip_list.insert('Key2' , 8 ) skip_list.insert('V' , 13 ) assert skip_list.find('Y' ) is None assert skip_list.find('Key2' ) == 8 assert skip_list.find('Some Key' ) == 10 assert skip_list.find('V' ) == 13 def lowerCAmelCase__ ( ) -> Tuple: lowerCAmelCase__ : Any = SkipList() skip_list.delete('Some key' ) assert len(skip_list.head.forward ) == 0 def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Optional[Any] = SkipList() skip_list.insert('Key1' , 12 ) skip_list.insert('V' , 13 ) skip_list.insert('X' , 14 ) skip_list.insert('Key2' , 15 ) skip_list.delete('V' ) skip_list.delete('Key2' ) assert skip_list.find('V' ) is None assert skip_list.find('Key2' ) is None def lowerCAmelCase__ ( ) -> Optional[Any]: lowerCAmelCase__ : Union[str, Any] = SkipList() skip_list.insert('Key1' , 12 ) skip_list.insert('V' , 13 ) skip_list.insert('X' , 14 ) skip_list.insert('Key2' , 15 ) skip_list.delete('V' ) assert skip_list.find('V' ) is None assert skip_list.find('X' ) == 14 assert skip_list.find('Key1' ) == 12 assert skip_list.find('Key2' ) == 15 skip_list.delete('X' ) assert skip_list.find('V' ) is None assert skip_list.find('X' ) is None assert skip_list.find('Key1' ) == 12 assert skip_list.find('Key2' ) == 15 skip_list.delete('Key1' ) assert skip_list.find('V' ) is None assert skip_list.find('X' ) is None assert skip_list.find('Key1' ) is None assert skip_list.find('Key2' ) == 15 skip_list.delete('Key2' ) assert skip_list.find('V' ) is None assert skip_list.find('X' ) is None assert skip_list.find('Key1' ) is None assert skip_list.find('Key2' ) is None def lowerCAmelCase__ ( ) -> Union[str, Any]: lowerCAmelCase__ : List[Any] = SkipList() skip_list.insert('Key1' , 12 ) skip_list.insert('V' , 13 ) skip_list.insert('X' , 142 ) skip_list.insert('Key2' , 15 ) skip_list.delete('X' ) def traverse_keys(SCREAMING_SNAKE_CASE_ ): yield node.key for forward_node in node.forward: yield from traverse_keys(SCREAMING_SNAKE_CASE_ ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def lowerCAmelCase__ ( ) -> Union[str, Any]: def is_sorted(SCREAMING_SNAKE_CASE_ ): return all(next_item >= item for item, next_item in zip(SCREAMING_SNAKE_CASE_ , lst[1:] ) ) lowerCAmelCase__ : int = SkipList() for i in range(10 ): skip_list.insert(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) assert is_sorted(list(SCREAMING_SNAKE_CASE_ ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(SCREAMING_SNAKE_CASE_ ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(SCREAMING_SNAKE_CASE_ ) ) def lowerCAmelCase__ ( ) -> int: for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def lowerCAmelCase__ ( ) -> List[str]: lowerCAmelCase__ : int = SkipList() skip_list.insert(2 , '2' ) skip_list.insert(4 , '4' ) skip_list.insert(6 , '4' ) skip_list.insert(4 , '5' ) skip_list.insert(8 , '4' ) skip_list.insert(9 , '4' ) skip_list.delete(4 ) print(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
69
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
1
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = r""" Args: input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax or scores for each vocabulary token after SoftMax. kwargs (`Dict[str, Any]`, *optional*): Additional stopping criteria specific kwargs. Return: `bool`. `False` indicates we should continue, `True` indicates we should stop. """ class A__ ( __magic_name__ ): @add_start_docstrings(a ) def __call__( self : int , a : torch.LongTensor , a : torch.FloatTensor , **a : Tuple ): '''simple docstring''' raise NotImplementedError('StoppingCriteria needs to be subclassed' ) class A__ ( __magic_name__ ): def __init__( self : int , a : int , a : Optional[int] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = max_length lowerCAmelCase__ : Optional[int] = max_position_embeddings @add_start_docstrings(a ) def __call__( self : Dict , a : torch.LongTensor , a : torch.FloatTensor , **a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = input_ids.shape[-1] lowerCAmelCase__ : Any = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( 'This is a friendly reminder - the current text generation call will exceed the model\'s predefined ' f'''maximum length ({self.max_position_embeddings}). Depending on the model, you may observe ''' 'exceptions, performance degradation, or nothing at all.' ) return is_done class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : int , a : int ): '''simple docstring''' warnings.warn( 'The class `MaxNewTokensCriteria` is deprecated. ' f'''Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` ''' 'with `max_length = start_length + max_new_tokens` instead.' , a , ) lowerCAmelCase__ : int = start_length lowerCAmelCase__ : List[str] = max_new_tokens lowerCAmelCase__ : Dict = start_length + max_new_tokens @add_start_docstrings(a ) def __call__( self : Any , a : torch.LongTensor , a : torch.FloatTensor , **a : str ): '''simple docstring''' return input_ids.shape[-1] >= self.max_length class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : float , a : Optional[float] = None ): '''simple docstring''' lowerCAmelCase__ : Any = max_time lowerCAmelCase__ : int = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(a ) def __call__( self : Optional[Any] , a : torch.LongTensor , a : torch.FloatTensor , **a : str ): '''simple docstring''' return time.time() - self.initial_timestamp > self.max_time class A__ ( __magic_name__ ): @add_start_docstrings(a ) def __call__( self : str , a : torch.LongTensor , a : torch.FloatTensor , **a : List[Any] ): '''simple docstring''' return any(criteria(a , a ) for criteria in self ) @property def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for stopping_criterium in self: if isinstance(a , a ): return stopping_criterium.max_length elif isinstance(a , a ): return stopping_criterium.max_length return None def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> StoppingCriteriaList: lowerCAmelCase__ : Tuple = stopping_criteria.max_length lowerCAmelCase__ : Optional[int] = deepcopy(SCREAMING_SNAKE_CASE_ ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn('You set different `max_length` for stopping criteria and `max_length` parameter' , SCREAMING_SNAKE_CASE_ ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=SCREAMING_SNAKE_CASE_ ) ) return new_stopping_criteria
69
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """google/rembert""": 256, } lowerCamelCase__ = """▁""" class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = RemBertTokenizer def __init__( self : Optional[Any] , a : str=None , a : Any=None , a : List[Any]=True , a : str=True , a : Dict=False , a : Dict="[CLS]" , a : int="[SEP]" , a : Tuple="<unk>" , a : Optional[Any]="[SEP]" , a : Tuple="<pad>" , a : Dict="[CLS]" , a : Optional[Any]="[MASK]" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( a , tokenizer_file=a , do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : int = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : List[Any] = keep_accents lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
69
1
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : int = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : List[Any] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : int = None lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Dict = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Union[str, Any] = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
69
1
import cmath import math def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> complex: lowerCAmelCase__ : Union[str, Any] = math.radians(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = math.radians(SCREAMING_SNAKE_CASE_ ) # Convert voltage and current to rectangular form lowerCAmelCase__ : Optional[int] = cmath.rect(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Tuple = cmath.rect(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
69
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
1
import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesseract, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class A__ : @staticmethod def _lowerCamelCase ( *a : List[Any] , **a : Any ): '''simple docstring''' pass @is_pipeline_test @require_vision @require_timm @require_torch class A__ ( unittest.TestCase ): lowercase = MODEL_FOR_OBJECT_DETECTION_MAPPING def _lowerCamelCase ( self : int , a : Any , a : Any , a : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = ObjectDetectionPipeline(model=a , image_processor=a ) return object_detector, ["./tests/fixtures/tests_samples/COCO/000000039769.png"] def _lowerCamelCase ( self : List[Any] , a : str , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = object_detector('./tests/fixtures/tests_samples/COCO/000000039769.png' , threshold=0.0 ) self.assertGreater(len(a ) , 0 ) for detected_object in outputs: self.assertEqual( a , { 'score': ANY(a ), 'label': ANY(a ), 'box': {'xmin': ANY(a ), 'ymin': ANY(a ), 'xmax': ANY(a ), 'ymax': ANY(a )}, } , ) import datasets lowerCAmelCase__ : Any = datasets.load_dataset('hf-internal-testing/fixtures_image_utils' , 'image' , split='test' ) lowerCAmelCase__ : Union[str, Any] = [ Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ), 'http://images.cocodataset.org/val2017/000000039769.jpg', # RGBA dataset[0]['file'], # LA dataset[1]['file'], # L dataset[2]['file'], ] lowerCAmelCase__ : Union[str, Any] = object_detector(a , threshold=0.0 ) self.assertEqual(len(a ) , len(a ) ) for outputs in batch_outputs: self.assertGreater(len(a ) , 0 ) for detected_object in outputs: self.assertEqual( a , { 'score': ANY(a ), 'label': ANY(a ), 'box': {'xmin': ANY(a ), 'ymin': ANY(a ), 'xmax': ANY(a ), 'ymax': ANY(a )}, } , ) @require_tf @unittest.skip('Object detection not implemented in TF' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' pass @require_torch def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'hf-internal-testing/tiny-detr-mobilenetsv3' lowerCAmelCase__ : Optional[Any] = AutoModelForObjectDetection.from_pretrained(a ) lowerCAmelCase__ : List[Any] = AutoFeatureExtractor.from_pretrained(a ) lowerCAmelCase__ : Optional[Any] = ObjectDetectionPipeline(model=a , feature_extractor=a ) lowerCAmelCase__ : Optional[int] = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' , threshold=0.0 ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, ] , ) lowerCAmelCase__ : Optional[int] = object_detector( [ 'http://images.cocodataset.org/val2017/000000039769.jpg', 'http://images.cocodataset.org/val2017/000000039769.jpg', ] , threshold=0.0 , ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ [ {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, ], [ {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, ], ] , ) @require_torch @slow def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'facebook/detr-resnet-50' lowerCAmelCase__ : Optional[Any] = AutoModelForObjectDetection.from_pretrained(a ) lowerCAmelCase__ : List[str] = AutoFeatureExtractor.from_pretrained(a ) lowerCAmelCase__ : Any = ObjectDetectionPipeline(model=a , feature_extractor=a ) lowerCAmelCase__ : List[str] = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ] , ) lowerCAmelCase__ : int = object_detector( [ 'http://images.cocodataset.org/val2017/000000039769.jpg', 'http://images.cocodataset.org/val2017/000000039769.jpg', ] ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], ] , ) @require_torch @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Dict = 'facebook/detr-resnet-50' lowerCAmelCase__ : Tuple = pipeline('object-detection' , model=a ) lowerCAmelCase__ : List[str] = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ] , ) lowerCAmelCase__ : Tuple = object_detector( [ 'http://images.cocodataset.org/val2017/000000039769.jpg', 'http://images.cocodataset.org/val2017/000000039769.jpg', ] ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], ] , ) @require_torch @slow def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = 0.9_9_8_5 lowerCAmelCase__ : str = 'facebook/detr-resnet-50' lowerCAmelCase__ : Optional[Any] = pipeline('object-detection' , model=a ) lowerCAmelCase__ : Dict = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' , threshold=a ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ] , ) @require_torch @require_pytesseract @slow def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Dict = 'Narsil/layoutlmv3-finetuned-funsd' lowerCAmelCase__ : str = 0.9_9_9_3 lowerCAmelCase__ : int = pipeline('object-detection' , model=a , threshold=a ) lowerCAmelCase__ : Union[str, Any] = object_detector( 'https://huggingface.co/spaces/impira/docquery/resolve/2359223c1837a7587402bda0f2643382a6eefeab/invoice.png' ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.9_9_9_3, 'label': 'I-ANSWER', 'box': {'xmin': 294, 'ymin': 254, 'xmax': 343, 'ymax': 264}}, {'score': 0.9_9_9_3, 'label': 'I-ANSWER', 'box': {'xmin': 294, 'ymin': 254, 'xmax': 343, 'ymax': 264}}, ] , )
69
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
1
from typing import List, Optional, Union import numpy as np from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['input_values', 'padding_mask'] def __init__( self : Union[str, Any] , a : int = 1 , a : int = 24_000 , a : float = 0.0 , a : float = None , a : float = None , **a : List[str] , ): '''simple docstring''' super().__init__(feature_size=a , sampling_rate=a , padding_value=a , **a ) lowerCAmelCase__ : Any = chunk_length_s lowerCAmelCase__ : Optional[int] = overlap @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) def __call__( self : Dict , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[bool, str, PaddingStrategy]] = None , a : Optional[bool] = False , a : Optional[int] = None , a : Optional[Union[str, TensorType]] = None , a : Optional[int] = None , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided audio input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) if padding and truncation: raise ValueError('Both padding and truncation were set. Make sure you only set one.' ) elif padding is None: # by default let's pad the inputs lowerCAmelCase__ : Any = True lowerCAmelCase__ : List[Any] = bool( isinstance(a , (list, tuple) ) and (isinstance(raw_audio[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : Tuple = [np.asarray(a , dtype=np.floataa ).T for audio in raw_audio] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : str = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_audio.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : List[str] = raw_audio.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : str = [np.asarray(a ).T] # verify inputs are valid for idx, example in enumerate(a ): if example.ndim > 2: raise ValueError(f'''Expected input shape (channels, length) but got shape {example.shape}''' ) if self.feature_size == 1 and example.ndim != 1: raise ValueError(f'''Expected mono audio but example has {example.shape[-1]} channels''' ) if self.feature_size == 2 and example.shape[-1] != 2: raise ValueError(f'''Expected stereo audio but example has {example.shape[-1]} channels''' ) lowerCAmelCase__ : Union[str, Any] = None lowerCAmelCase__ : Any = BatchFeature({'input_values': raw_audio} ) if self.chunk_stride is not None and self.chunk_length is not None and max_length is None: if truncation: lowerCAmelCase__ : Any = min(array.shape[0] for array in raw_audio ) lowerCAmelCase__ : int = int(np.floor(max_length / self.chunk_stride ) ) lowerCAmelCase__ : Dict = (nb_step - 1) * self.chunk_stride + self.chunk_length elif padding: lowerCAmelCase__ : str = max(array.shape[0] for array in raw_audio ) lowerCAmelCase__ : Optional[int] = int(np.ceil(max_length / self.chunk_stride ) ) lowerCAmelCase__ : Dict = (nb_step - 1) * self.chunk_stride + self.chunk_length lowerCAmelCase__ : Any = 'max_length' else: lowerCAmelCase__ : List[str] = input_values # normal padding on batch if padded_inputs is None: lowerCAmelCase__ : Any = self.pad( a , max_length=a , truncation=a , padding=a , return_attention_mask=a , ) if padding: lowerCAmelCase__ : str = padded_inputs.pop('attention_mask' ) lowerCAmelCase__ : Optional[int] = [] for example in padded_inputs.pop('input_values' ): if self.feature_size == 1: lowerCAmelCase__ : Any = example[..., None] input_values.append(example.T ) lowerCAmelCase__ : List[Any] = input_values if return_tensors is not None: lowerCAmelCase__ : List[str] = padded_inputs.convert_to_tensors(a ) return padded_inputs
69
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
1
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise TypeError('Input value must be an \'int\' type' ) lowerCAmelCase__ : List[Any] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
69
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
1
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list[list[int]]: lowerCAmelCase__ : Union[str, Any] = [] if len(SCREAMING_SNAKE_CASE_ ) == 1: return [nums.copy()] for _ in range(len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ : Optional[int] = nums.pop(0 ) lowerCAmelCase__ : List[str] = permute(SCREAMING_SNAKE_CASE_ ) for perm in permutations: perm.append(SCREAMING_SNAKE_CASE_ ) result.extend(SCREAMING_SNAKE_CASE_ ) nums.append(SCREAMING_SNAKE_CASE_ ) return result def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: def backtrack(SCREAMING_SNAKE_CASE_ ): if start == len(SCREAMING_SNAKE_CASE_ ) - 1: output.append(nums[:] ) else: for i in range(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ , lowerCAmelCase__ : str = nums[i], nums[start] backtrack(start + 1 ) lowerCAmelCase__ , lowerCAmelCase__ : Tuple = nums[i], nums[start] # backtrack lowerCAmelCase__ : Optional[int] = [] backtrack(0 ) return output if __name__ == "__main__": import doctest # use res to print the data in permute2 function lowerCamelCase__ = permutea([1, 2, 3]) print(res) doctest.testmod()
69
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
1
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ , lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
69
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
1
import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class A__ : def __init__( self : List[str] , a : Optional[int]=2 , a : Optional[Any]=3 , a : List[Any]=64 , a : Any=None ): '''simple docstring''' lowerCAmelCase__ : Dict = np.random.default_rng(a ) lowerCAmelCase__ : Tuple = length lowerCAmelCase__ : Tuple = rng.normal(size=(length,) ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self : List[str] ): '''simple docstring''' return self.length def __getitem__( self : Union[str, Any] , a : Dict ): '''simple docstring''' return {"x": self.x[i], "y": self.y[i]} class A__ ( torch.nn.Module ): def __init__( self : Dict , a : Tuple=0 , a : Optional[Any]=0 , a : Optional[int]=False ): '''simple docstring''' super().__init__() lowerCAmelCase__ : Union[str, Any] = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) lowerCAmelCase__ : List[Any] = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) lowerCAmelCase__ : Any = True def _lowerCamelCase ( self : Union[str, Any] , a : Optional[int]=None ): '''simple docstring''' if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) lowerCAmelCase__ : Optional[Any] = False return x * self.a[0] + self.b[0] class A__ ( torch.nn.Module ): def __init__( self : Any , a : List[Any]=0 , a : Optional[Any]=0 , a : Tuple=False ): '''simple docstring''' super().__init__() lowerCAmelCase__ : int = torch.nn.Parameter(torch.tensor(a ).float() ) lowerCAmelCase__ : str = torch.nn.Parameter(torch.tensor(a ).float() ) lowerCAmelCase__ : Tuple = True def _lowerCamelCase ( self : str , a : int=None ): '''simple docstring''' if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) lowerCAmelCase__ : Optional[int] = False return x * self.a + self.b def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 16 ) -> Dict: from datasets import load_dataset from transformers import AutoTokenizer lowerCAmelCase__ : List[str] = AutoTokenizer.from_pretrained('bert-base-cased' ) lowerCAmelCase__ : int = {'train': 'tests/test_samples/MRPC/train.csv', 'validation': 'tests/test_samples/MRPC/dev.csv'} lowerCAmelCase__ : str = load_dataset('csv' , data_files=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Tuple = datasets['train'].unique('label' ) lowerCAmelCase__ : Union[str, Any] = {v: i for i, v in enumerate(SCREAMING_SNAKE_CASE_ )} def tokenize_function(SCREAMING_SNAKE_CASE_ ): # max_length=None => use the model max length (it's actually the default) lowerCAmelCase__ : Union[str, Any] = tokenizer( examples['sentence1'] , examples['sentence2'] , truncation=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , padding='max_length' ) if "label" in examples: lowerCAmelCase__ : Optional[Any] = [label_to_id[l] for l in examples['label']] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset lowerCAmelCase__ : Union[str, Any] = datasets.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , remove_columns=['sentence1', 'sentence2', 'label'] , ) def collate_fn(SCREAMING_SNAKE_CASE_ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(SCREAMING_SNAKE_CASE_ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. lowerCAmelCase__ : int = DataLoader(tokenized_datasets['train'] , shuffle=SCREAMING_SNAKE_CASE_ , collate_fn=SCREAMING_SNAKE_CASE_ , batch_size=2 ) lowerCAmelCase__ : List[Any] = DataLoader(tokenized_datasets['validation'] , shuffle=SCREAMING_SNAKE_CASE_ , collate_fn=SCREAMING_SNAKE_CASE_ , batch_size=1 ) return train_dataloader, eval_dataloader
69
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
1
from typing import Dict from transformers import EvalPrediction, HfArgumentParser, TrainingArguments, is_torch_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, get_torch_dist_unique_port, require_torch_multi_gpu, require_torch_neuroncore, ) from transformers.training_args import ParallelMode from transformers.utils import logging lowerCamelCase__ = logging.get_logger(__name__) if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset from transformers import Trainer class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : int = 101 ): '''simple docstring''' lowerCAmelCase__ : int = length def __len__( self : int ): '''simple docstring''' return self.length def __getitem__( self : List[Any] , a : int ): '''simple docstring''' return i class A__ : def __call__( self : Any , a : List[str] ): '''simple docstring''' return {"input_ids": torch.tensor(a ), "labels": torch.tensor(a )} class A__ ( nn.Module ): def __init__( self : Tuple ): '''simple docstring''' super().__init__() # Add some (unused) params otherwise DDP will complain. lowerCAmelCase__ : Union[str, Any] = nn.Linear(120 , 80 ) def _lowerCamelCase ( self : str , a : Tuple , a : str=None ): '''simple docstring''' if labels is not None: return torch.tensor(0.0 , device=input_ids.device ), input_ids else: return input_ids class A__ ( __magic_name__ ): @require_torch_neuroncore def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = f'''--nproc_per_node=2 --master_port={get_torch_dist_unique_port()} {self.test_file_dir}/test_trainer_distributed.py '''.split() lowerCAmelCase__ : str = self.get_auto_remove_tmp_dir() lowerCAmelCase__ : str = f'''--output_dir {output_dir}'''.split() lowerCAmelCase__ : Optional[Any] = ['torchrun'] + distributed_args + args execute_subprocess_async(a , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call class A__ ( __magic_name__ ): @require_torch_multi_gpu def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : int = f'''--nproc_per_node={torch.cuda.device_count()} --master_port={get_torch_dist_unique_port()} {self.test_file_dir}/test_trainer_distributed.py '''.split() lowerCAmelCase__ : Union[str, Any] = self.get_auto_remove_tmp_dir() lowerCAmelCase__ : List[Any] = f'''--output_dir {output_dir}'''.split() lowerCAmelCase__ : Optional[int] = ['torchrun'] + distributed_args + args execute_subprocess_async(a , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call if __name__ == "__main__": # The script below is meant to be run under torch.distributed, on a machine with multiple GPUs: # # PYTHONPATH="src" python -m torch.distributed.run --nproc_per_node 2 --output_dir output_dir ./tests/test_trainer_distributed.py lowerCamelCase__ = HfArgumentParser((TrainingArguments,)) lowerCamelCase__ = parser.parse_args_into_dataclasses()[0] logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}, """ F"""distributed training: {training_args.parallel_mode != ParallelMode.NOT_DISTRIBUTED}""" ) # Essentially, what we want to verify in the distributed case is that we get all samples back, # in the right order. (this is crucial for prediction for instance) for dataset_length in [101, 40, 7]: lowerCamelCase__ = DummyDataset(dataset_length) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Dict: lowerCAmelCase__ : Union[str, Any] = list(range(len(SCREAMING_SNAKE_CASE_ ) ) ) lowerCAmelCase__ : Dict = p.predictions.tolist() == sequential and p.label_ids.tolist() == sequential if not success and training_args.local_rank == 0: logger.warning( 'Predictions and/or labels do not match expected results:\n - predictions: ' F'''{p.predictions.tolist()}\n - labels: {p.label_ids.tolist()}\n - expected: {sequential}''' ) return {"success": success} lowerCamelCase__ = Trainer( model=DummyModel(), args=training_args, data_collator=DummyDataCollator(), eval_dataset=dataset, compute_metrics=compute_metrics, ) lowerCamelCase__ = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) lowerCamelCase__ = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) lowerCamelCase__ = 2 lowerCamelCase__ = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) lowerCamelCase__ = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) lowerCamelCase__ = None
69
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
1
import inspect import unittest from transformers import ViTMSNConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMSNForImageClassification, ViTMSNModel from transformers.models.vit_msn.modeling_vit_msn import VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A__ : def __init__( self : Any , a : Dict , a : int=13 , a : Optional[Any]=30 , a : List[Any]=2 , a : str=3 , a : str=True , a : Optional[Any]=True , a : List[Any]=32 , a : Union[str, Any]=5 , a : List[Any]=4 , a : Any=37 , a : List[Any]="gelu" , a : List[str]=0.1 , a : str=0.1 , a : List[str]=10 , a : Optional[int]=0.0_2 , a : Dict=None , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = parent lowerCAmelCase__ : Optional[Any] = batch_size lowerCAmelCase__ : int = image_size lowerCAmelCase__ : List[Any] = patch_size lowerCAmelCase__ : Optional[Any] = num_channels lowerCAmelCase__ : int = is_training lowerCAmelCase__ : Optional[Any] = use_labels lowerCAmelCase__ : int = hidden_size lowerCAmelCase__ : Union[str, Any] = num_hidden_layers lowerCAmelCase__ : Tuple = num_attention_heads lowerCAmelCase__ : Any = intermediate_size lowerCAmelCase__ : Union[str, Any] = hidden_act lowerCAmelCase__ : List[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : str = type_sequence_label_size lowerCAmelCase__ : Optional[Any] = initializer_range lowerCAmelCase__ : List[Any] = scope # in ViT MSN, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCAmelCase__ : str = (image_size // patch_size) ** 2 lowerCAmelCase__ : Tuple = num_patches + 1 def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Union[str, Any] = None if self.use_labels: lowerCAmelCase__ : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Dict = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : Dict ): '''simple docstring''' return ViTMSNConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def _lowerCamelCase ( self : List[str] , a : Any , a : Any , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = ViTMSNModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : int , a : Optional[Any] , a : Tuple , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.type_sequence_label_size lowerCAmelCase__ : Optional[Any] = ViTMSNForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , labels=a ) print('Pixel and labels shape: {pixel_values.shape}, {labels.shape}' ) print('Labels: {labels}' ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : List[Any] = ViTMSNForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : int = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = config_and_inputs lowerCAmelCase__ : Tuple = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = (ViTMSNModel, ViTMSNForImageClassification) if is_torch_available() else () lowercase = ( {'feature-extraction': ViTMSNModel, 'image-classification': ViTMSNForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = ViTMSNModelTester(self ) lowerCAmelCase__ : int = ConfigTester(self , config_class=a , has_text_modality=a , hidden_size=37 ) def _lowerCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='ViTMSN does not use inputs_embeds' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : List[Any] = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[Any] = model_class(a ) lowerCAmelCase__ : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Optional[int] = [*signature.parameters.keys()] lowerCAmelCase__ : Tuple = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : str ): '''simple docstring''' for model_name in VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[Any] = ViTMSNModel.from_pretrained(a ) self.assertIsNotNone(a ) def lowerCAmelCase__ ( ) -> Any: lowerCAmelCase__ : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Dict ): '''simple docstring''' return ViTImageProcessor.from_pretrained('facebook/vit-msn-small' ) if is_vision_available() else None @slow def _lowerCamelCase ( self : str ): '''simple docstring''' torch.manual_seed(2 ) lowerCAmelCase__ : List[Any] = ViTMSNForImageClassification.from_pretrained('facebook/vit-msn-small' ).to(a ) lowerCAmelCase__ : Optional[int] = self.default_image_processor lowerCAmelCase__ : Dict = prepare_img() lowerCAmelCase__ : Union[str, Any] = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : int = model(**a ) # verify the logits lowerCAmelCase__ : Any = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : int = torch.tensor([-0.0_8_0_3, -0.4_4_5_4, -0.2_3_7_5] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) )
69
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
1
import tensorflow as tf from ...tf_utils import shape_list class A__ ( tf.keras.layers.Layer ): def __init__( self : Any , a : List[str] , a : Optional[int] , a : Tuple , a : int , a : Optional[int]=1 , a : Dict=False , **a : Optional[Any] ): '''simple docstring''' super().__init__(**a ) lowerCAmelCase__ : int = vocab_size lowerCAmelCase__ : List[Any] = d_embed lowerCAmelCase__ : str = d_proj lowerCAmelCase__ : Optional[Any] = cutoffs + [vocab_size] lowerCAmelCase__ : Dict = [0] + self.cutoffs lowerCAmelCase__ : Optional[int] = div_val lowerCAmelCase__ : int = self.cutoffs[0] lowerCAmelCase__ : List[Any] = len(self.cutoffs ) - 1 lowerCAmelCase__ : str = self.shortlist_size + self.n_clusters lowerCAmelCase__ : Optional[Any] = keep_order lowerCAmelCase__ : int = [] lowerCAmelCase__ : Union[str, Any] = [] def _lowerCamelCase ( self : Union[str, Any] , a : List[str] ): '''simple docstring''' if self.n_clusters > 0: lowerCAmelCase__ : Optional[Any] = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer='zeros' , trainable=a , name='cluster_weight' ) lowerCAmelCase__ : Tuple = self.add_weight( shape=(self.n_clusters,) , initializer='zeros' , trainable=a , name='cluster_bias' ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: lowerCAmelCase__ : Any = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer='zeros' , trainable=a , name=f'''out_projs_._{i}''' , ) self.out_projs.append(a ) else: self.out_projs.append(a ) lowerCAmelCase__ : Optional[int] = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer='zeros' , trainable=a , name=f'''out_layers_._{i}_._weight''' , ) lowerCAmelCase__ : List[str] = self.add_weight( shape=(self.vocab_size,) , initializer='zeros' , trainable=a , name=f'''out_layers_._{i}_._bias''' , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): lowerCAmelCase__ , lowerCAmelCase__ : str = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowerCAmelCase__ : List[Any] = self.d_embed // (self.div_val**i) lowerCAmelCase__ : List[Any] = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer='zeros' , trainable=a , name=f'''out_projs_._{i}''' ) self.out_projs.append(a ) lowerCAmelCase__ : List[Any] = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer='zeros' , trainable=a , name=f'''out_layers_._{i}_._weight''' , ) lowerCAmelCase__ : List[Any] = self.add_weight( shape=(r_idx - l_idx,) , initializer='zeros' , trainable=a , name=f'''out_layers_._{i}_._bias''' , ) self.out_layers.append((weight, bias) ) super().build(a ) @staticmethod def _lowerCamelCase ( a : str , a : Tuple , a : Union[str, Any] , a : List[str]=None ): '''simple docstring''' lowerCAmelCase__ : str = x if proj is not None: lowerCAmelCase__ : Any = tf.einsum('ibd,ed->ibe' , a , a ) return tf.einsum('ibd,nd->ibn' , a , a ) + b @staticmethod def _lowerCamelCase ( a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = shape_list(a ) lowerCAmelCase__ : List[str] = tf.range(lp_size[0] , dtype=target.dtype ) lowerCAmelCase__ : Optional[Any] = tf.stack([r, target] , 1 ) return tf.gather_nd(a , a ) def _lowerCamelCase ( self : Any , a : Union[str, Any] , a : int , a : Optional[Any]=True , a : Dict=False ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 0 if self.n_clusters == 0: lowerCAmelCase__ : str = self._logit(a , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: lowerCAmelCase__ : Tuple = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=a , logits=a ) lowerCAmelCase__ : List[str] = tf.nn.log_softmax(a , axis=-1 ) else: lowerCAmelCase__ : Tuple = shape_list(a ) lowerCAmelCase__ : str = [] lowerCAmelCase__ : Optional[Any] = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: lowerCAmelCase__ : List[Any] = (target >= l_idx) & (target < r_idx) lowerCAmelCase__ : Dict = tf.where(a ) lowerCAmelCase__ : int = tf.boolean_mask(a , a ) - l_idx if self.div_val == 1: lowerCAmelCase__ : Any = self.out_layers[0][0][l_idx:r_idx] lowerCAmelCase__ : Optional[int] = self.out_layers[0][1][l_idx:r_idx] else: lowerCAmelCase__ : Union[str, Any] = self.out_layers[i][0] lowerCAmelCase__ : List[str] = self.out_layers[i][1] if i == 0: lowerCAmelCase__ : Optional[Any] = tf.concat([cur_W, self.cluster_weight] , 0 ) lowerCAmelCase__ : Optional[int] = tf.concat([cur_b, self.cluster_bias] , 0 ) lowerCAmelCase__ : Optional[int] = self._logit(a , a , a , self.out_projs[0] ) lowerCAmelCase__ : Tuple = tf.nn.log_softmax(a ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: lowerCAmelCase__ : List[str] = tf.boolean_mask(a , a ) lowerCAmelCase__ : int = self._gather_logprob(a , a ) else: lowerCAmelCase__ : List[str] = self._logit(a , a , a , self.out_projs[i] ) lowerCAmelCase__ : Any = tf.nn.log_softmax(a ) lowerCAmelCase__ : Optional[int] = self.cutoffs[0] + i - 1 # No probability for the head cluster lowerCAmelCase__ : Optional[int] = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(a ) if target is not None: lowerCAmelCase__ : Optional[Any] = tf.boolean_mask(a , a ) lowerCAmelCase__ : Union[str, Any] = tf.boolean_mask(a , a ) lowerCAmelCase__ : str = self._gather_logprob(a , a ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(a , -cur_logprob , shape_list(a ) ) lowerCAmelCase__ : Dict = tf.concat(a , axis=-1 ) if target is not None: if return_mean: lowerCAmelCase__ : str = tf.reduce_mean(a ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(a ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(a , name=self.name , aggregation='mean' if return_mean else '' ) return out
69
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
1
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowerCamelCase__ = get_tests_dir("""fixtures/spiece.model""") @require_sentencepiece @require_tokenizers class A__ ( __magic_name__ , unittest.TestCase ): lowercase = AlbertTokenizer lowercase = AlbertTokenizerFast lowercase = True lowercase = True lowercase = True def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase__ : Dict = AlbertTokenizer(a ) tokenizer.save_pretrained(self.tmpdirname ) def _lowerCamelCase ( self : Any , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = 'this is a test' lowerCAmelCase__ : List[str] = 'this is a test' return input_text, output_text def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = '<pad>' lowerCAmelCase__ : Tuple = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a ) , a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a ) , a ) def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<pad>' ) self.assertEqual(vocab_keys[1] , '<unk>' ) self.assertEqual(vocab_keys[-1] , '▁eloquent' ) self.assertEqual(len(a ) , 30_000 ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 30_000 ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' if not self.test_rust_tokenizer: return lowerCAmelCase__ : List[str] = self.get_tokenizer() lowerCAmelCase__ : List[str] = self.get_rust_tokenizer() lowerCAmelCase__ : Optional[int] = 'I was born in 92000, and this is falsé.' lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) lowerCAmelCase__ : Optional[Any] = rust_tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Any = tokenizer.encode(a , add_special_tokens=a ) lowerCAmelCase__ : List[str] = rust_tokenizer.encode(a , add_special_tokens=a ) self.assertListEqual(a , a ) lowerCAmelCase__ : List[str] = self.get_rust_tokenizer() lowerCAmelCase__ : Dict = tokenizer.encode(a ) lowerCAmelCase__ : List[Any] = rust_tokenizer.encode(a ) self.assertListEqual(a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[str] = AlbertTokenizer(a , keep_accents=a ) lowerCAmelCase__ : Optional[int] = tokenizer.tokenize('This is a test' ) self.assertListEqual(a , ['▁this', '▁is', '▁a', '▁test'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , [48, 25, 21, 1_289] ) lowerCAmelCase__ : Dict = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( a , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.'] ) lowerCAmelCase__ : Any = tokenizer.convert_tokens_to_ids(a ) self.assertListEqual(a , [31, 23, 386, 19, 561, 3_050, 15, 17, 48, 25, 8_256, 18, 1, 9] ) lowerCAmelCase__ : Union[str, Any] = tokenizer.convert_ids_to_tokens(a ) self.assertListEqual( a , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.'] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = AlbertTokenizer(a ) lowerCAmelCase__ : Tuple = tokenizer.encode('sequence builders' ) lowerCAmelCase__ : str = tokenizer.encode('multi-sequence build' ) lowerCAmelCase__ : Any = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : Dict = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = {'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'input_ids': [[2, 21_970, 13, 5, 6_092, 167, 28, 7_103, 2_153, 673, 8, 7_028, 12_051, 18, 17, 7_103, 2_153, 673, 8, 3_515, 18_684, 8, 4_461, 6, 1_927, 297, 8, 12_060, 2_607, 18, 13, 5, 4_461, 15, 10_538, 38, 8, 135, 15, 822, 58, 15, 993, 10_363, 15, 1_460, 8_005, 4_461, 15, 993, 255, 2_328, 9, 9, 9, 6, 26, 1_112, 816, 3_260, 13, 5, 103, 2_377, 6, 17, 1_112, 816, 2_782, 13, 5, 103, 10_641, 6, 29, 84, 2_512, 2_430, 782, 18_684, 2_761, 19, 808, 2_430, 2_556, 17, 855, 1_480, 9_477, 4_091, 128, 11_712, 15, 7_103, 2_153, 673, 17, 24_883, 9_990, 9, 3], [2, 11_502, 25, 1_006, 20, 782, 8, 11_809, 855, 1_732, 19_393, 18_667, 37, 367, 21_018, 69, 1_854, 34, 11_860, 19_124, 27, 156, 225, 17, 193, 4_141, 19, 65, 9_124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2_231, 886, 2_385, 17_659, 84, 14, 16_792, 1_952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a , model_name='albert-base-v2' , revision='6b6560eaf5ff2e250b00c50f380c5389a9c2d82e' , )
69
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
69
1
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) lowerCamelCase__ = logging.getLogger(__name__) @dataclass(frozen=__magic_name__ ) class A__ : lowercase = 42 lowercase = 42 lowercase = None lowercase = None lowercase = None @dataclass(frozen=__magic_name__ ) class A__ : lowercase = 42 lowercase = None lowercase = None lowercase = None lowercase = None if is_torch_available(): import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): lowercase = 42 def __init__( self : List[str] , a : str , a : PreTrainedTokenizer , a : str , a : Optional[int] = None , a : str=False , a : bool = False , ): '''simple docstring''' lowerCAmelCase__ : int = hans_processors[task]() lowerCAmelCase__ : Dict = os.path.join( a , 'cached_{}_{}_{}_{}'.format( 'dev' if evaluate else 'train' , tokenizer.__class__.__name__ , str(a ) , a , ) , ) lowerCAmelCase__ : Any = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase__ , lowerCAmelCase__ : int = label_list[2], label_list[1] lowerCAmelCase__ : int = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCAmelCase__ : List[str] = cached_features_file + '.lock' with FileLock(a ): if os.path.exists(a ) and not overwrite_cache: logger.info(f'''Loading features from cached file {cached_features_file}''' ) lowerCAmelCase__ : Optional[Any] = torch.load(a ) else: logger.info(f'''Creating features from dataset file at {data_dir}''' ) lowerCAmelCase__ : Tuple = ( processor.get_dev_examples(a ) if evaluate else processor.get_train_examples(a ) ) logger.info('Training examples: %s' , len(a ) ) lowerCAmelCase__ : Optional[Any] = hans_convert_examples_to_features(a , a , a , a ) logger.info('Saving features into cached file %s' , a ) torch.save(self.features , a ) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.features ) def __getitem__( self : Optional[Any] , a : int ): '''simple docstring''' return self.features[i] def _lowerCamelCase ( self : str ): '''simple docstring''' return self.label_list if is_tf_available(): import tensorflow as tf class A__ : lowercase = 42 def __init__( self : Any , a : str , a : PreTrainedTokenizer , a : str , a : Optional[int] = 128 , a : Union[str, Any]=False , a : bool = False , ): '''simple docstring''' lowerCAmelCase__ : List[str] = hans_processors[task]() lowerCAmelCase__ : List[str] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase__ , lowerCAmelCase__ : Dict = label_list[2], label_list[1] lowerCAmelCase__ : Any = label_list lowerCAmelCase__ : Optional[Any] = processor.get_dev_examples(a ) if evaluate else processor.get_train_examples(a ) lowerCAmelCase__ : Optional[Any] = hans_convert_examples_to_features(a , a , a , a ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='convert examples to features' ): if ex_index % 10_000 == 0: logger.info('Writing example %d of %d' % (ex_index, len(a )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowerCAmelCase__ : Optional[Any] = tf.data.Dataset.from_generator( a , ( { 'example_id': tf.intaa, 'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa, }, tf.intaa, ) , ( { 'example_id': tf.TensorShape([] ), 'input_ids': tf.TensorShape([None, None] ), 'attention_mask': tf.TensorShape([None, None] ), 'token_type_ids': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' return self.dataset def __len__( self : Tuple ): '''simple docstring''' return len(self.features ) def __getitem__( self : Dict , a : Union[str, Any] ): '''simple docstring''' return self.features[i] def _lowerCamelCase ( self : str ): '''simple docstring''' return self.label_list class A__ ( __magic_name__ ): def _lowerCamelCase ( self : str , a : Dict ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(a , 'heuristics_train_set.txt' ) ) , 'train' ) def _lowerCamelCase ( self : Dict , a : Optional[Any] ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(a , 'heuristics_evaluation_set.txt' ) ) , 'dev' ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return ["contradiction", "entailment", "neutral"] def _lowerCamelCase ( self : Dict , a : Dict , a : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = [] for i, line in enumerate(a ): if i == 0: continue lowerCAmelCase__ : Optional[int] = '%s-%s' % (set_type, line[0]) lowerCAmelCase__ : Tuple = line[5] lowerCAmelCase__ : Any = line[6] lowerCAmelCase__ : Union[str, Any] = line[7][2:] if line[7].startswith('ex' ) else line[7] lowerCAmelCase__ : List[str] = line[0] examples.append(InputExample(guid=a , text_a=a , text_b=a , label=a , pairID=a ) ) return examples def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[int]: lowerCAmelCase__ : Dict = {label: i for i, label in enumerate(SCREAMING_SNAKE_CASE_ )} lowerCAmelCase__ : Dict = [] for ex_index, example in tqdm.tqdm(enumerate(SCREAMING_SNAKE_CASE_ ) , desc='convert examples to features' ): if ex_index % 10_000 == 0: logger.info('Writing example %d' % (ex_index) ) lowerCAmelCase__ : Union[str, Any] = tokenizer( example.text_a , example.text_b , add_special_tokens=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , padding='max_length' , truncation=SCREAMING_SNAKE_CASE_ , return_overflowing_tokens=SCREAMING_SNAKE_CASE_ , ) lowerCAmelCase__ : List[str] = label_map[example.label] if example.label in label_map else 0 lowerCAmelCase__ : Any = int(example.pairID ) features.append(InputFeatures(**SCREAMING_SNAKE_CASE_ , label=SCREAMING_SNAKE_CASE_ , pairID=SCREAMING_SNAKE_CASE_ ) ) for i, example in enumerate(examples[:5] ): logger.info('*** Example ***' ) logger.info(F'''guid: {example}''' ) logger.info(F'''features: {features[i]}''' ) return features lowerCamelCase__ = { """hans""": 3, } lowerCamelCase__ = { """hans""": HansProcessor, }
69
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __magic_name__ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'LayoutLMv3ImageProcessor' lowercase = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self : Optional[int] , a : Union[str, Any]=None , a : Optional[Any]=None , **a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , a , ) lowerCAmelCase__ : int = kwargs.pop('feature_extractor' ) lowerCAmelCase__ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(a , a ) def __call__( self : List[Any] , a : List[Any] , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a : Union[List[List[int]], List[List[List[int]]]] = None , a : Optional[Union[List[int], List[List[int]]]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : str , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor lowerCAmelCase__ : List[str] = self.image_processor(images=a , return_tensors=a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a , a ): lowerCAmelCase__ : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ : List[str] = features['words'] lowerCAmelCase__ : List[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_token_type_ids=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_length=a , verbose=a , return_tensors=a , **a , ) # add pixel values lowerCAmelCase__ : Tuple = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase__ : List[str] = self.get_overflowing_images(a , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase__ : List[str] = images return encoded_inputs def _lowerCamelCase ( self : Any , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a ) != len(a ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' f''' {len(a )} and {len(a )}''' ) return images_with_overflow def _lowerCamelCase ( self : Union[str, Any] , *a : Optional[Any] , **a : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Tuple , *a : List[str] , **a : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*a , **a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , a , ) return self.image_processor_class @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , a , ) return self.image_processor
69
1
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> float: if principal <= 0: raise Exception('Principal borrowed must be > 0' ) if rate_per_annum < 0: raise Exception('Rate of interest must be >= 0' ) if years_to_repay <= 0 or not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise Exception('Years to repay must be an integer > 0' ) # Yearly rate is divided by 12 to get monthly rate lowerCAmelCase__ : Optional[int] = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly lowerCAmelCase__ : str = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
69
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
69
1
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['input_features', 'attention_mask'] def __init__( self : Optional[Any] , a : Union[str, Any]=80 , a : List[Any]=16_000 , a : Union[str, Any]=0.0 , a : int=10 , a : Any=25 , a : str="hamming_window" , a : str=3_2_7_6_8.0 , a : Any=0.9_7 , a : Union[str, Any]=1.0 , a : Optional[int]=True , a : Optional[int]=True , a : int=False , **a : Dict , ): '''simple docstring''' super().__init__(feature_size=a , sampling_rate=a , padding_value=a , **a ) lowerCAmelCase__ : Optional[Any] = feature_size lowerCAmelCase__ : Any = sampling_rate lowerCAmelCase__ : int = padding_value lowerCAmelCase__ : Optional[Any] = hop_length lowerCAmelCase__ : Any = win_length lowerCAmelCase__ : List[str] = frame_signal_scale lowerCAmelCase__ : Union[str, Any] = preemphasis_coeff lowerCAmelCase__ : Tuple = mel_floor lowerCAmelCase__ : List[str] = normalize_means lowerCAmelCase__ : int = normalize_vars lowerCAmelCase__ : List[Any] = win_function lowerCAmelCase__ : int = return_attention_mask lowerCAmelCase__ : int = win_length * sampling_rate // 1_000 lowerCAmelCase__ : Tuple = hop_length * sampling_rate // 1_000 lowerCAmelCase__ : Tuple = optimal_fft_length(self.sample_size ) lowerCAmelCase__ : Any = (self.n_fft // 2) + 1 def _lowerCamelCase ( self : Optional[Any] , a : np.array ): '''simple docstring''' if self.win_function == "hamming_window": lowerCAmelCase__ : Optional[int] = window_function(window_length=self.sample_size , name=self.win_function , periodic=a ) else: lowerCAmelCase__ : Tuple = window_function(window_length=self.sample_size , name=self.win_function ) lowerCAmelCase__ : Tuple = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) lowerCAmelCase__ : Any = spectrogram( one_waveform * self.frame_signal_scale , window=a , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=a , preemphasis=self.preemphasis_coeff , mel_filters=a , mel_floor=self.mel_floor , log_mel='log' , ) return msfc_features.T def _lowerCamelCase ( self : Optional[Any] , a : int , a : Union[str, Any] , a : List[str] ): '''simple docstring''' if self.normalize_means: lowerCAmelCase__ : Dict = x[:input_length].mean(axis=0 ) lowerCAmelCase__ : List[Any] = np.subtract(a , a ) if self.normalize_vars: lowerCAmelCase__ : Union[str, Any] = x[:input_length].std(axis=0 ) lowerCAmelCase__ : Any = np.divide(a , a ) if input_length < x.shape[0]: lowerCAmelCase__ : List[str] = padding_value # make sure array is in float32 lowerCAmelCase__ : Optional[int] = x.astype(np.floataa ) return x def _lowerCamelCase ( self : Union[str, Any] , a : List[np.ndarray] , a : Optional[np.ndarray] = None ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(a , a , self.padding_value ) for x, n in zip(a , a )] def __call__( self : Union[str, Any] , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Union[bool, str, PaddingStrategy] = False , a : Optional[int] = None , a : bool = False , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[Union[str, TensorType]] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the ``sampling_rate`` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : int = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : int = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : str = [np.asarray(a , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Dict = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : Optional[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : str = [raw_speech] # extract fbank features lowerCAmelCase__ : Dict = [self._extract_mfsc_features(a ) for one_waveform in raw_speech] # convert into correct format for padding lowerCAmelCase__ : List[str] = BatchFeature({'input_features': features} ) lowerCAmelCase__ : Optional[Any] = self.pad( a , padding=a , max_length=a , truncation=a , pad_to_multiple_of=a , return_attention_mask=a , **a , ) # make sure list is in array format lowerCAmelCase__ : List[str] = padded_inputs.get('input_features' ) if isinstance(input_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in input_features] lowerCAmelCase__ : Optional[Any] = padded_inputs.get('attention_mask' ) if attention_mask is not None: lowerCAmelCase__ : List[Any] = [np.asarray(a , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCAmelCase__ : str = ( np.array(a , dtype=np.intaa ) if self._get_padding_strategies(a , max_length=a ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCAmelCase__ : Tuple = self.normalize( padded_inputs['input_features'] , attention_mask=a ) if return_tensors is not None: lowerCAmelCase__ : List[str] = padded_inputs.convert_to_tensors(a ) return padded_inputs
69
lowerCamelCase__ = """Alexander Joslin""" import operator as op from .stack import Stack def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} lowerCAmelCase__ : Stack[int] = Stack() lowerCAmelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE_ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE_ ) elif i == ")": # RULE 4 lowerCAmelCase__ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : List[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : Tuple = operators[opr](SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) operand_stack.push(SCREAMING_SNAKE_CASE_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
69
1
from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class A__ : lowercase = XGLMConfig lowercase = {} lowercase = 'gelu' def __init__( self : Tuple , a : int , a : str=14 , a : Tuple=7 , a : Optional[int]=True , a : Optional[int]=True , a : List[str]=True , a : int=99 , a : Optional[int]=32 , a : Any=2 , a : Optional[Any]=4 , a : Optional[int]=37 , a : List[Any]="gelu" , a : Union[str, Any]=0.1 , a : Any=0.1 , a : Optional[int]=512 , a : Tuple=0.0_2 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Tuple = batch_size lowerCAmelCase__ : Optional[int] = seq_length lowerCAmelCase__ : List[Any] = is_training lowerCAmelCase__ : str = use_input_mask lowerCAmelCase__ : Optional[Any] = use_labels lowerCAmelCase__ : List[str] = vocab_size lowerCAmelCase__ : Optional[int] = d_model lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : Any = num_attention_heads lowerCAmelCase__ : Dict = ffn_dim lowerCAmelCase__ : Any = activation_function lowerCAmelCase__ : Optional[Any] = activation_dropout lowerCAmelCase__ : Dict = attention_dropout lowerCAmelCase__ : int = max_position_embeddings lowerCAmelCase__ : Union[str, Any] = initializer_range lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Optional[int] = 0 lowerCAmelCase__ : str = 2 lowerCAmelCase__ : List[str] = 1 def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return XGLMConfig.from_pretrained('facebook/xglm-564M' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[str] = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) lowerCAmelCase__ : Union[str, Any] = None if self.use_input_mask: lowerCAmelCase__ : Any = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : Optional[Any] = self.get_config() lowerCAmelCase__ : Tuple = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=a , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=a , ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : Tuple = config_and_inputs lowerCAmelCase__ : Tuple = { 'input_ids': input_ids, 'head_mask': head_mask, } return config, inputs_dict @require_tf class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () lowercase = (TFXGLMForCausalLM,) if is_tf_available() else () lowercase = ( {'feature-extraction': TFXGLMModel, 'text-generation': TFXGLMForCausalLM} if is_tf_available() else {} ) lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = TFXGLMModelTester(self ) lowerCAmelCase__ : Any = ConfigTester(self , config_class=a , n_embd=37 ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' self.config_tester.run_common_tests() @slow def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Tuple = TFXGLMModel.from_pretrained(a ) self.assertIsNotNone(a ) @unittest.skip(reason='Currently, model embeddings are going to undergo a major refactor.' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' super().test_resize_token_embeddings() @require_tf class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Tuple , a : Optional[Any]=True ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) lowerCAmelCase__ : Dict = tf.convert_to_tensor([[2, 268, 9_865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off lowerCAmelCase__ : int = [2, 268, 9_865, 67, 11, 1_988, 57_252, 9_865, 5, 984, 67, 1_988, 213_838, 1_658, 53, 70_446, 33, 6_657, 278, 1_581] # fmt: on lowerCAmelCase__ : Dict = model.generate(a , do_sample=a , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) lowerCAmelCase__ : str = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) tf.random.set_seed(0 ) lowerCAmelCase__ : int = tokenizer('Today is a nice day and' , return_tensors='tf' ) lowerCAmelCase__ : List[Any] = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(':/CPU:0' ): lowerCAmelCase__ : Tuple = model.generate(a , do_sample=a , seed=[7, 0] ) lowerCAmelCase__ : Union[str, Any] = tokenizer.decode(output_ids[0] , skip_special_tokens=a ) lowerCAmelCase__ : int = ( 'Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due' ) self.assertEqual(a , a ) @slow def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) lowerCAmelCase__ : Union[str, Any] = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) lowerCAmelCase__ : Any = 'left' # use different length sentences to test batching lowerCAmelCase__ : List[Any] = [ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When', 'Hello, my dog is a little', ] lowerCAmelCase__ : Dict = tokenizer(a , return_tensors='tf' , padding=a ) lowerCAmelCase__ : int = inputs['input_ids'] lowerCAmelCase__ : Dict = model.generate(input_ids=a , attention_mask=inputs['attention_mask'] , max_new_tokens=12 ) lowerCAmelCase__ : Union[str, Any] = tokenizer(sentences[0] , return_tensors='tf' ).input_ids lowerCAmelCase__ : int = model.generate(input_ids=a , max_new_tokens=12 ) lowerCAmelCase__ : Tuple = tokenizer(sentences[1] , return_tensors='tf' ).input_ids lowerCAmelCase__ : Any = model.generate(input_ids=a , max_new_tokens=12 ) lowerCAmelCase__ : List[str] = tokenizer.batch_decode(a , skip_special_tokens=a ) lowerCAmelCase__ : Optional[int] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=a ) lowerCAmelCase__ : Optional[int] = tokenizer.decode(output_padded[0] , skip_special_tokens=a ) lowerCAmelCase__ : Any = [ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ' 'a single', 'Hello, my dog is a little bit of a shy one, but he is very friendly', ] self.assertListEqual(a , a ) self.assertListEqual(a , [non_padded_sentence, padded_sentence] )
69
import numpy class A__ : def __init__( self : Tuple , a : numpy.ndarray , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : int = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Dict = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : List[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowerCAmelCase__ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : List[Any] = numpy.zeros(output_array.shape ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowerCAmelCase__ : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowerCAmelCase__ : int = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowerCamelCase ( self : Optional[int] , a : numpy.ndarray , a : int , a : bool ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): lowerCAmelCase__ : Any = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Tuple = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def _lowerCamelCase ( self : Optional[Any] , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : Dict = input_arr lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return (value) * (1 - (value)) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : List[str] = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ , output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ , iterations=10 , give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
69
1
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) lowerCamelCase__ = logging.getLogger(__name__) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : Tuple = np.argmax(SCREAMING_SNAKE_CASE_ , axis=1 ) return np.sum(outputs == labels ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: with open(SCREAMING_SNAKE_CASE_ , encoding='utf_8' ) as f: lowerCAmelCase__ : List[Any] = csv.reader(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Union[str, Any] = [] next(SCREAMING_SNAKE_CASE_ ) # skip the first line for line in tqdm(SCREAMING_SNAKE_CASE_ ): output.append((' '.join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : Optional[Any] = [] for dataset in encoded_datasets: lowerCAmelCase__ : Any = len(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : List[str] = np.zeros((n_batch, 2, input_len) , dtype=np.intaa ) lowerCAmelCase__ : List[Any] = np.zeros((n_batch, 2) , dtype=np.intaa ) lowerCAmelCase__ : Union[str, Any] = np.full((n_batch, 2, input_len) , fill_value=-100 , dtype=np.intaa ) lowerCAmelCase__ : str = np.zeros((n_batch,) , dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Dict = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] lowerCAmelCase__ : Optional[Any] = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] lowerCAmelCase__ : Dict = with_conta lowerCAmelCase__ : List[Any] = with_conta lowerCAmelCase__ : Tuple = len(SCREAMING_SNAKE_CASE_ ) - 1 lowerCAmelCase__ : List[str] = len(SCREAMING_SNAKE_CASE_ ) - 1 lowerCAmelCase__ : str = with_conta lowerCAmelCase__ : List[str] = with_conta lowerCAmelCase__ : Dict = mc_label lowerCAmelCase__ : str = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(SCREAMING_SNAKE_CASE_ ) for t in all_inputs ) ) return tensor_datasets def lowerCAmelCase__ ( ) -> Any: lowerCAmelCase__ : Optional[Any] = argparse.ArgumentParser() parser.add_argument('--model_name' , type=SCREAMING_SNAKE_CASE_ , default='openai-gpt' , help='pretrained model name' ) parser.add_argument('--do_train' , action='store_true' , help='Whether to run training.' ) parser.add_argument('--do_eval' , action='store_true' , help='Whether to run eval on the dev set.' ) parser.add_argument( '--output_dir' , default=SCREAMING_SNAKE_CASE_ , type=SCREAMING_SNAKE_CASE_ , required=SCREAMING_SNAKE_CASE_ , help='The output directory where the model predictions and checkpoints will be written.' , ) parser.add_argument('--train_dataset' , type=SCREAMING_SNAKE_CASE_ , default='' ) parser.add_argument('--eval_dataset' , type=SCREAMING_SNAKE_CASE_ , default='' ) parser.add_argument('--seed' , type=SCREAMING_SNAKE_CASE_ , default=42 ) parser.add_argument('--num_train_epochs' , type=SCREAMING_SNAKE_CASE_ , default=3 ) parser.add_argument('--train_batch_size' , type=SCREAMING_SNAKE_CASE_ , default=8 ) parser.add_argument('--eval_batch_size' , type=SCREAMING_SNAKE_CASE_ , default=16 ) parser.add_argument('--adam_epsilon' , default=1e-8 , type=SCREAMING_SNAKE_CASE_ , help='Epsilon for Adam optimizer.' ) parser.add_argument('--max_grad_norm' , type=SCREAMING_SNAKE_CASE_ , default=1 ) parser.add_argument( '--max_steps' , default=-1 , type=SCREAMING_SNAKE_CASE_ , help=( 'If > 0: set total number of training steps to perform. Override num_train_epochs.' ) , ) parser.add_argument( '--gradient_accumulation_steps' , type=SCREAMING_SNAKE_CASE_ , default=1 , help='Number of updates steps to accumulate before performing a backward/update pass.' , ) parser.add_argument('--learning_rate' , type=SCREAMING_SNAKE_CASE_ , default=6.25e-5 ) parser.add_argument('--warmup_steps' , default=0 , type=SCREAMING_SNAKE_CASE_ , help='Linear warmup over warmup_steps.' ) parser.add_argument('--lr_schedule' , type=SCREAMING_SNAKE_CASE_ , default='warmup_linear' ) parser.add_argument('--weight_decay' , type=SCREAMING_SNAKE_CASE_ , default=0.01 ) parser.add_argument('--lm_coef' , type=SCREAMING_SNAKE_CASE_ , default=0.9 ) parser.add_argument('--n_valid' , type=SCREAMING_SNAKE_CASE_ , default=374 ) parser.add_argument('--server_ip' , type=SCREAMING_SNAKE_CASE_ , default='' , help='Can be used for distant debugging.' ) parser.add_argument('--server_port' , type=SCREAMING_SNAKE_CASE_ , default='' , help='Can be used for distant debugging.' ) lowerCAmelCase__ : int = parser.parse_args() print(SCREAMING_SNAKE_CASE_ ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('Waiting for debugger attach' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=SCREAMING_SNAKE_CASE_ ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) lowerCAmelCase__ : Optional[int] = torch.device('cuda' if torch.cuda.is_available() else 'cpu' ) lowerCAmelCase__ : Dict = torch.cuda.device_count() logger.info('device: {}, n_gpu {}'.format(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) if not args.do_train and not args.do_eval: raise ValueError('At least one of `do_train` or `do_eval` must be True.' ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset lowerCAmelCase__ : Dict = ['_start_', '_delimiter_', '_classify_'] lowerCAmelCase__ : List[Any] = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : List[Any] = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(SCREAMING_SNAKE_CASE_ ) ) model.to(SCREAMING_SNAKE_CASE_ ) # Load and encode the datasets def tokenize_and_encode(SCREAMING_SNAKE_CASE_ ): if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(SCREAMING_SNAKE_CASE_ ) ) elif isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return obj return [tokenize_and_encode(SCREAMING_SNAKE_CASE_ ) for o in obj] logger.info('Encoding dataset...' ) lowerCAmelCase__ : Tuple = load_rocstories_dataset(args.train_dataset ) lowerCAmelCase__ : List[str] = load_rocstories_dataset(args.eval_dataset ) lowerCAmelCase__ : Tuple = (train_dataset, eval_dataset) lowerCAmelCase__ : Any = tokenize_and_encode(SCREAMING_SNAKE_CASE_ ) # Compute the max input length for the Transformer lowerCAmelCase__ : int = model.config.n_positions // 2 - 2 lowerCAmelCase__ : Any = max( len(story[:max_length] ) + max(len(conta[:max_length] ) , len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) lowerCAmelCase__ : int = min(SCREAMING_SNAKE_CASE_ , model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders lowerCAmelCase__ : Dict = pre_process_datasets(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , *SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ , lowerCAmelCase__ : Tuple = tensor_datasets[0], tensor_datasets[1] lowerCAmelCase__ : str = TensorDataset(*SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Tuple = RandomSampler(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Union[str, Any] = DataLoader(SCREAMING_SNAKE_CASE_ , sampler=SCREAMING_SNAKE_CASE_ , batch_size=args.train_batch_size ) lowerCAmelCase__ : int = TensorDataset(*SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : List[str] = SequentialSampler(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[int] = DataLoader(SCREAMING_SNAKE_CASE_ , sampler=SCREAMING_SNAKE_CASE_ , batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: lowerCAmelCase__ : int = args.max_steps lowerCAmelCase__ : Dict = args.max_steps // (len(SCREAMING_SNAKE_CASE_ ) // args.gradient_accumulation_steps) + 1 else: lowerCAmelCase__ : List[str] = len(SCREAMING_SNAKE_CASE_ ) // args.gradient_accumulation_steps * args.num_train_epochs lowerCAmelCase__ : Optional[Any] = list(model.named_parameters() ) lowerCAmelCase__ : int = ['bias', 'LayerNorm.bias', 'LayerNorm.weight'] lowerCAmelCase__ : int = [ { 'params': [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], 'weight_decay': args.weight_decay, }, {'params': [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], 'weight_decay': 0.0}, ] lowerCAmelCase__ : Union[str, Any] = AdamW(SCREAMING_SNAKE_CASE_ , lr=args.learning_rate , eps=args.adam_epsilon ) lowerCAmelCase__ : Optional[int] = get_linear_schedule_with_warmup( SCREAMING_SNAKE_CASE_ , num_warmup_steps=args.warmup_steps , num_training_steps=SCREAMING_SNAKE_CASE_ ) if args.do_train: lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) , desc='Epoch' ): lowerCAmelCase__ : List[str] = 0 lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : Dict = tqdm(SCREAMING_SNAKE_CASE_ , desc='Training' ) for step, batch in enumerate(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Union[str, Any] = tuple(t.to(SCREAMING_SNAKE_CASE_ ) for t in batch ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = batch lowerCAmelCase__ : List[Any] = model(SCREAMING_SNAKE_CASE_ , mc_token_ids=SCREAMING_SNAKE_CASE_ , lm_labels=SCREAMING_SNAKE_CASE_ , mc_labels=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Tuple = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() lowerCAmelCase__ : Optional[Any] = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 lowerCAmelCase__ : Any = 'Training loss: {:.2e} lr: {:.2e}'.format(SCREAMING_SNAKE_CASE_ , scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer lowerCAmelCase__ : Any = model.module if hasattr(SCREAMING_SNAKE_CASE_ , 'module' ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` lowerCAmelCase__ : str = os.path.join(args.output_dir , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = os.path.join(args.output_dir , SCREAMING_SNAKE_CASE_ ) torch.save(model_to_save.state_dict() , SCREAMING_SNAKE_CASE_ ) model_to_save.config.to_json_file(SCREAMING_SNAKE_CASE_ ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned lowerCAmelCase__ : Optional[Any] = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) lowerCAmelCase__ : int = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(SCREAMING_SNAKE_CASE_ ) if args.do_eval: model.eval() lowerCAmelCase__ , lowerCAmelCase__ : List[str] = 0, 0 lowerCAmelCase__ , lowerCAmelCase__ : Tuple = 0, 0 for batch in tqdm(SCREAMING_SNAKE_CASE_ , desc='Evaluating' ): lowerCAmelCase__ : List[str] = tuple(t.to(SCREAMING_SNAKE_CASE_ ) for t in batch ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = batch with torch.no_grad(): lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = model( SCREAMING_SNAKE_CASE_ , mc_token_ids=SCREAMING_SNAKE_CASE_ , lm_labels=SCREAMING_SNAKE_CASE_ , mc_labels=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = mc_logits.detach().cpu().numpy() lowerCAmelCase__ : Optional[Any] = mc_labels.to('cpu' ).numpy() lowerCAmelCase__ : Dict = accuracy(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 lowerCAmelCase__ : List[str] = eval_loss / nb_eval_steps lowerCAmelCase__ : Tuple = eval_accuracy / nb_eval_examples lowerCAmelCase__ : List[str] = tr_loss / nb_tr_steps if args.do_train else None lowerCAmelCase__ : Tuple = {'eval_loss': eval_loss, 'eval_accuracy': eval_accuracy, 'train_loss': train_loss} lowerCAmelCase__ : int = os.path.join(args.output_dir , 'eval_results.txt' ) with open(SCREAMING_SNAKE_CASE_ , 'w' ) as writer: logger.info('***** Eval results *****' ) for key in sorted(result.keys() ): logger.info(' %s = %s' , SCREAMING_SNAKE_CASE_ , str(result[key] ) ) writer.write('%s = %s\n' % (key, str(result[key] )) ) if __name__ == "__main__": main()
69
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A__ : def __init__( self : int , a : str , a : Union[str, Any]=13 , a : int=32 , a : Optional[Any]=2 , a : Tuple=3 , a : List[Any]=16 , a : List[str]=[1, 2, 1] , a : int=[2, 2, 4] , a : int=2 , a : Optional[Any]=2.0 , a : Optional[int]=True , a : Dict=0.0 , a : Any=0.0 , a : int=0.1 , a : List[str]="gelu" , a : Optional[Any]=False , a : str=True , a : Dict=0.0_2 , a : Any=1E-5 , a : Optional[int]=True , a : str=None , a : str=True , a : int=10 , a : str=8 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : Optional[Any] = patch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = embed_dim lowerCAmelCase__ : Tuple = depths lowerCAmelCase__ : List[str] = num_heads lowerCAmelCase__ : List[Any] = window_size lowerCAmelCase__ : Any = mlp_ratio lowerCAmelCase__ : Optional[Any] = qkv_bias lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : int = drop_path_rate lowerCAmelCase__ : Optional[Any] = hidden_act lowerCAmelCase__ : int = use_absolute_embeddings lowerCAmelCase__ : List[str] = patch_norm lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Dict = use_labels lowerCAmelCase__ : List[Any] = type_sequence_label_size lowerCAmelCase__ : Optional[Any] = encoder_stride def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = None if self.use_labels: lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : int = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : List[str] , a : Any , a : str , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = SwinvaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model(a ) lowerCAmelCase__ : str = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase__ : Tuple = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SwinvaForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : Dict = SwinvaForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Union[str, Any] , a : int , a : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.type_sequence_label_size lowerCAmelCase__ : List[Any] = SwinvaForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = config_and_inputs lowerCAmelCase__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = SwinvaModelTester(self ) lowerCAmelCase__ : int = ConfigTester(self , config_class=a , embed_dim=37 ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[int] = model_class(a ) lowerCAmelCase__ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Tuple = [*signature.parameters.keys()] lowerCAmelCase__ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = True for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = True lowerCAmelCase__ : str = False lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Dict = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : int = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Dict = outputs.attentions lowerCAmelCase__ : Dict = len(self.model_tester.depths ) self.assertEqual(len(a ) , a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Optional[int] = config.window_size**2 lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowerCAmelCase__ : Tuple = len(a ) # Check attention is always last and order is fine lowerCAmelCase__ : str = True lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(**self._prepare_for_class(a , a ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): lowerCAmelCase__ : Optional[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase__ : Any = 2 self.assertEqual(out_len + added_hidden_states , len(a ) ) lowerCAmelCase__ : Dict = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _lowerCamelCase ( self : int , a : Optional[int] , a : int , a : Optional[Any] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.hidden_states lowerCAmelCase__ : str = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a ) , a ) # Swinv2 has a different seq_length lowerCAmelCase__ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowerCAmelCase__ : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(a ) , a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = reshaped_hidden_states[0].shape lowerCAmelCase__ : List[str] = ( reshaped_hidden_states[0].view(a , a , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : str = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : List[str] = True self.check_hidden_states_output(a , a , a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Any = 3 lowerCAmelCase__ : int = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase__ : str = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : str = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase__ : Any = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase__ : str = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[str] = SwinvaModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = _config_zero_init(a ) for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(config=a ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( a ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Any = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**a ) # verify the logits lowerCAmelCase__ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Optional[Any] = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) )
69
1
import json import os import tempfile from unittest.mock import patch import torch from torch.utils.data import DataLoader, TensorDataset from accelerate import DistributedType, infer_auto_device_map, init_empty_weights from accelerate.accelerator import Accelerator from accelerate.state import GradientState, PartialState from accelerate.test_utils import require_bnb, require_multi_gpu, slow from accelerate.test_utils.testing import AccelerateTestCase, require_cuda from accelerate.utils import patch_environment def lowerCAmelCase__ ( ) -> str: lowerCAmelCase__ : int = torch.nn.Linear(2 , 4 ) lowerCAmelCase__ : int = torch.optim.AdamW(model.parameters() , lr=1.0 ) lowerCAmelCase__ : str = torch.optim.lr_scheduler.OneCycleLR(SCREAMING_SNAKE_CASE_ , max_lr=0.01 , steps_per_epoch=2 , epochs=1 ) lowerCAmelCase__ : Optional[int] = DataLoader(TensorDataset(torch.tensor([1, 2, 3] ) ) ) lowerCAmelCase__ : Tuple = DataLoader(TensorDataset(torch.tensor([4, 5, 6] ) ) ) return model, optimizer, scheduler, train_dl, valid_dl def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[Any]: return (model.weight.abs().sum() + model.bias.abs().sum()).item() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[Any]: lowerCAmelCase__ : Optional[int] = torch.nn.Linear(*tuple(model.weight.T.shape ) ).state_dict() model.load_state_dict(SCREAMING_SNAKE_CASE_ ) class A__ ( __magic_name__ ): @require_cuda def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : str = Accelerator() assert PartialState._shared_state["_cpu"] is False assert PartialState._shared_state["device"].type == "cuda" with self.assertRaises(a ): lowerCAmelCase__ : Optional[Any] = Accelerator(cpu=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : int = Accelerator() lowerCAmelCase__ : List[Any] = GradientState() assert state.num_steps == 1 lowerCAmelCase__ : Union[str, Any] = 4 assert state.num_steps == 4 assert state.sync_gradients is True lowerCAmelCase__ : Optional[Any] = False assert state.sync_gradients is False GradientState._reset_state() def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = Accelerator() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = create_components() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : Any = accelerator.prepare(a , a , a , a , a ) self.assertTrue(prepared_model in accelerator._models ) self.assertTrue(prepared_optimizer in accelerator._optimizers ) self.assertTrue(prepared_scheduler in accelerator._schedulers ) self.assertTrue(prepared_train_dl in accelerator._dataloaders ) self.assertTrue(prepared_valid_dl in accelerator._dataloaders ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = Accelerator() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Dict = create_components() accelerator.prepare(a , a , a , a , a ) accelerator.free_memory() self.assertTrue(len(accelerator._models ) == 0 ) self.assertTrue(len(accelerator._optimizers ) == 0 ) self.assertTrue(len(accelerator._schedulers ) == 0 ) self.assertTrue(len(accelerator._dataloaders ) == 0 ) def _lowerCamelCase ( self : str ): '''simple docstring''' PartialState._reset_state() # Mock torch.cuda.set_device to avoid an exception as the device doesn't exist def noop(*a : List[str] , **a : Tuple ): pass with patch('torch.cuda.set_device' , a ), patch_environment(ACCELERATE_TORCH_DEVICE='cuda:64' ): lowerCAmelCase__ : Tuple = Accelerator() self.assertEqual(str(accelerator.state.device ) , 'cuda:64' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = Accelerator() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : List[str] = create_components() accelerator.prepare(a , a , a , a , a ) lowerCAmelCase__ : List[Any] = get_signature(a ) with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(a ) # make sure random weights don't match load_random_weights(a ) self.assertTrue(abs(model_signature - get_signature(a ) ) > 1E-3 ) # make sure loaded weights match accelerator.load_state(a ) self.assertTrue(abs(model_signature - get_signature(a ) ) < 1E-3 ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : int = Accelerator() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : List[str] = create_components() accelerator.prepare(a , a , a , a , a ) lowerCAmelCase__ : Tuple = get_signature(a ) # saving hook def save_config(a : Any , a : Optional[Any] , a : List[Any] ): lowerCAmelCase__ : List[Any] = {'class_name': models[0].__class__.__name__} with open(os.path.join(a , 'data.json' ) , 'w' ) as f: json.dump(a , a ) # loading hook def load_config(a : Optional[Any] , a : List[str] ): with open(os.path.join(a , 'data.json' ) , 'r' ) as f: lowerCAmelCase__ : Optional[int] = json.load(a ) lowerCAmelCase__ : List[Any] = config['class_name'] lowerCAmelCase__ : Union[str, Any] = accelerator.register_save_state_pre_hook(a ) lowerCAmelCase__ : str = accelerator.register_load_state_pre_hook(a ) with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(a ) # make sure random weights don't match with hooks load_random_weights(a ) self.assertTrue(abs(model_signature - get_signature(a ) ) > 1E-3 ) # random class name to verify correct one is loaded lowerCAmelCase__ : Union[str, Any] = 'random' # make sure loaded weights match with hooks accelerator.load_state(a ) self.assertTrue(abs(model_signature - get_signature(a ) ) < 1E-3 ) # mode.class_name is loaded from config self.assertTrue(model.class_name == model.__class__.__name__ ) # remove hooks save_hook.remove() load_hook.remove() with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(a ) # make sure random weights don't match with hooks removed load_random_weights(a ) self.assertTrue(abs(model_signature - get_signature(a ) ) > 1E-3 ) # random class name to verify correct one is loaded lowerCAmelCase__ : Optional[int] = 'random' # make sure loaded weights match with hooks removed accelerator.load_state(a ) self.assertTrue(abs(model_signature - get_signature(a ) ) < 1E-3 ) # mode.class_name is NOT loaded from config self.assertTrue(model.class_name != model.__class__.__name__ ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = Accelerator() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = create_components() lowerCAmelCase__ : Dict = None # This should work lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Dict = accelerator.prepare( a , a , a , a , a , a ) self.assertTrue(dummy_obj is None ) def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = Accelerator() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = create_components() lowerCAmelCase__ : Optional[Any] = [1, 2, 3] # This should work lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = accelerator.prepare( a , a , a , a , a , a ) self.assertEqual( getattr(a , '_is_accelerate_prepared' , a ) , a , 'Dummy object should have `_is_accelerate_prepared` set to `True`' , ) self.assertEqual( getattr(a , '_is_accelerate_prepared' , a ) , a , 'Model is missing `_is_accelerator_prepared` or is set to `False`' , ) self.assertEqual( getattr(a , '_is_accelerate_prepared' , a ) , a , 'Optimizer is missing `_is_accelerator_prepared` or is set to `False`' , ) self.assertEqual( getattr(a , '_is_accelerate_prepared' , a ) , a , 'Scheduler is missing `_is_accelerator_prepared` or is set to `False`' , ) self.assertEqual( getattr(a , '_is_accelerate_prepared' , a ) , a , 'Train Dataloader is missing `_is_accelerator_prepared` or is set to `False`' , ) self.assertEqual( getattr(a , '_is_accelerate_prepared' , a ) , a , 'Valid Dataloader is missing `_is_accelerator_prepared` or is set to `False`' , ) @slow @require_bnb def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' from transformers import AutoModelForCausalLM lowerCAmelCase__ : Tuple = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , load_in_abit=a , device_map={'': 0} , ) lowerCAmelCase__ : Union[str, Any] = Accelerator() # This should work lowerCAmelCase__ : Union[str, Any] = accelerator.prepare(a ) @slow @require_bnb def _lowerCamelCase ( self : Dict ): '''simple docstring''' from transformers import AutoModelForCausalLM lowerCAmelCase__ : Any = Accelerator() with init_empty_weights(): lowerCAmelCase__ : List[Any] = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , ) model.tie_weights() lowerCAmelCase__ : Any = infer_auto_device_map(a ) lowerCAmelCase__ : Optional[int] = 'cpu' lowerCAmelCase__ : int = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , device_map=a , load_in_abit=a , llm_inta_enable_fpaa_cpu_offload=a ) # This should not work and get value error with self.assertRaises(a ): lowerCAmelCase__ : Any = accelerator.prepare(a ) @slow @require_bnb @require_multi_gpu def _lowerCamelCase ( self : str ): '''simple docstring''' from transformers import AutoModelForCausalLM lowerCAmelCase__ : Tuple = {'distributed_type': DistributedType.MULTI_GPU} with init_empty_weights(): lowerCAmelCase__ : Tuple = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , ) model.tie_weights() lowerCAmelCase__ : Union[str, Any] = infer_auto_device_map(a ) lowerCAmelCase__ : Optional[int] = 1 lowerCAmelCase__ : Union[str, Any] = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , load_in_abit=a , device_map=a , ) lowerCAmelCase__ : Any = Accelerator() # This should not work and get value error with self.assertRaises(a ): lowerCAmelCase__ : Dict = accelerator.prepare(a ) PartialState._reset_state() @slow @require_bnb @require_multi_gpu def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' from transformers import AutoModelForCausalLM with init_empty_weights(): lowerCAmelCase__ : Tuple = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , ) lowerCAmelCase__ : Optional[Any] = infer_auto_device_map(a ) lowerCAmelCase__ : Optional[int] = 1 lowerCAmelCase__ : List[Any] = AutoModelForCausalLM.from_pretrained( 'EleutherAI/gpt-neo-125m' , load_in_abit=a , device_map=a , ) lowerCAmelCase__ : List[str] = Accelerator() # This should work lowerCAmelCase__ : Any = accelerator.prepare(a ) @require_cuda def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Dict = torch.nn.Linear(10 , 10 ) lowerCAmelCase__ : Tuple = torch.optim.SGD(model.parameters() , lr=0.0_1 ) lowerCAmelCase__ : List[str] = Accelerator(cpu=a ) lowerCAmelCase__ : List[Any] = accelerator.prepare(a )
69
from itertools import permutations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase__ : str = [7, 11, 13, 17] for i, test in enumerate(SCREAMING_SNAKE_CASE_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 10 ) -> int: return sum( int(''.join(map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) ) for num in permutations(range(SCREAMING_SNAKE_CASE_ ) ) if is_substring_divisible(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
69
1
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask lowerCamelCase__ = logging.getLogger(__name__) class A__ ( __magic_name__ ): lowercase = 'token-classification' def __init__( self : List[str] , a : int ): '''simple docstring''' if type(a ) == dict: lowerCAmelCase__ : Any = Namespace(**a ) lowerCAmelCase__ : Tuple = import_module('tasks' ) try: lowerCAmelCase__ : List[Any] = getattr(a , hparams.task_type ) lowerCAmelCase__ : TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( f'''Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ''' f'''Available tasks classes are: {TokenClassificationTask.__subclasses__()}''' ) lowerCAmelCase__ : Optional[Any] = self.token_classification_task.get_labels(hparams.labels ) lowerCAmelCase__ : List[Any] = CrossEntropyLoss().ignore_index super().__init__(a , len(self.labels ) , self.mode ) def _lowerCamelCase ( self : int , **a : Union[str, Any] ): '''simple docstring''' return self.model(**a ) def _lowerCamelCase ( self : Optional[int] , a : str , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : List[str] = {'input_ids': batch[0], 'attention_mask': batch[1], 'labels': batch[3]} if self.config.model_type != "distilbert": lowerCAmelCase__ : Tuple = ( batch[2] if self.config.model_type in ['bert', 'xlnet'] else None ) # XLM and RoBERTa don"t use token_type_ids lowerCAmelCase__ : int = self(**a ) lowerCAmelCase__ : Union[str, Any] = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.hparams for mode in ["train", "dev", "test"]: lowerCAmelCase__ : str = self._feature_file(a ) if os.path.exists(a ) and not args.overwrite_cache: logger.info('Loading features from cached file %s' , a ) lowerCAmelCase__ : int = torch.load(a ) else: logger.info('Creating features from dataset file at %s' , args.data_dir ) lowerCAmelCase__ : Optional[int] = self.token_classification_task.read_examples_from_file(args.data_dir , a ) lowerCAmelCase__ : Dict = self.token_classification_task.convert_examples_to_features( a , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ['xlnet'] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ['xlnet'] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=a , pad_on_left=bool(self.config.model_type in ['xlnet'] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info('Saving features into cached file %s' , a ) torch.save(a , a ) def _lowerCamelCase ( self : Any , a : int , a : int , a : bool = False ): '''simple docstring''' lowerCAmelCase__ : Dict = self._feature_file(a ) logger.info('Loading features from cached file %s' , a ) lowerCAmelCase__ : Optional[Any] = torch.load(a ) lowerCAmelCase__ : Optional[int] = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) lowerCAmelCase__ : Optional[Any] = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: lowerCAmelCase__ : Any = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: lowerCAmelCase__ : Optional[int] = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) lowerCAmelCase__ : Optional[int] = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(a , a , a , a ) , batch_size=a ) def _lowerCamelCase ( self : Optional[Any] , a : Optional[int] , a : List[str] ): '''simple docstring''' """Compute validation""" "" lowerCAmelCase__ : List[Any] = {'input_ids': batch[0], 'attention_mask': batch[1], 'labels': batch[3]} if self.config.model_type != "distilbert": lowerCAmelCase__ : List[Any] = ( batch[2] if self.config.model_type in ['bert', 'xlnet'] else None ) # XLM and RoBERTa don"t use token_type_ids lowerCAmelCase__ : Optional[Any] = self(**a ) lowerCAmelCase__ , lowerCAmelCase__ : List[str] = outputs[:2] lowerCAmelCase__ : List[str] = logits.detach().cpu().numpy() lowerCAmelCase__ : Optional[int] = inputs['labels'].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def _lowerCamelCase ( self : str , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.stack([x['val_loss'] for x in outputs] ).mean() lowerCAmelCase__ : Optional[int] = np.concatenate([x['pred'] for x in outputs] , axis=0 ) lowerCAmelCase__ : Union[str, Any] = np.argmax(a , axis=2 ) lowerCAmelCase__ : int = np.concatenate([x['target'] for x in outputs] , axis=0 ) lowerCAmelCase__ : List[str] = dict(enumerate(self.labels ) ) lowerCAmelCase__ : Any = [[] for _ in range(out_label_ids.shape[0] )] lowerCAmelCase__ : List[str] = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) lowerCAmelCase__ : Tuple = { 'val_loss': val_loss_mean, 'accuracy_score': accuracy_score(a , a ), 'precision': precision_score(a , a ), 'recall': recall_score(a , a ), 'f1': fa_score(a , a ), } lowerCAmelCase__ : Optional[int] = dict(results.items() ) lowerCAmelCase__ : Tuple = results return ret, preds_list, out_label_list def _lowerCamelCase ( self : Union[str, Any] , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self._eval_end(a ) lowerCAmelCase__ : int = ret['log'] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def _lowerCamelCase ( self : str , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self._eval_end(a ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 lowerCAmelCase__ : int = ret['log'] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def _lowerCamelCase ( a : int , a : Union[str, Any] ): '''simple docstring''' BaseTransformer.add_model_specific_args(a , a ) parser.add_argument( '--task_type' , default='NER' , type=a , help='Task type to fine tune in training (e.g. NER, POS, etc)' ) parser.add_argument( '--max_seq_length' , default=128 , type=a , help=( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) , ) parser.add_argument( '--labels' , default='' , type=a , help='Path to a file containing all labels. If not specified, CoNLL-2003 labels are used.' , ) parser.add_argument( '--gpus' , default=0 , type=a , help='The number of GPUs allocated for this, it is by default 0 meaning none' , ) parser.add_argument( '--overwrite_cache' , action='store_true' , help='Overwrite the cached training and evaluation sets' ) return parser if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) lowerCamelCase__ = NERTransformer.add_model_specific_args(parser, os.getcwd()) lowerCamelCase__ = parser.parse_args() lowerCamelCase__ = NERTransformer(args) lowerCamelCase__ = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 lowerCamelCase__ = sorted(glob.glob(os.path.join(args.output_dir, """checkpoint-epoch=*.ckpt"""), recursive=True)) lowerCamelCase__ = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
69
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
69
1
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 50 ) -> int: lowerCAmelCase__ : str = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(F"""{solution() = }""")
69
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ , lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
69
1
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} # See all MVP models at https://huggingface.co/models?filter=mvp lowerCamelCase__ = { """vocab_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json""", }, """added_tokens.json""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json""", }, """merges_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt""", }, """tokenizer_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """RUCAIBox/mvp""": 1024, } class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = ['input_ids', 'attention_mask'] lowercase = MvpTokenizer def __init__( self : Optional[Any] , a : Tuple=None , a : Any=None , a : Any=None , a : Union[str, Any]="replace" , a : str="<s>" , a : Any="</s>" , a : List[str]="</s>" , a : Optional[Any]="<s>" , a : Optional[int]="<unk>" , a : str="<pad>" , a : Any="<mask>" , a : List[Any]=False , a : Any=True , **a : Optional[Any] , ): '''simple docstring''' super().__init__( a , a , tokenizer_file=a , errors=a , bos_token=a , eos_token=a , sep_token=a , cls_token=a , unk_token=a , pad_token=a , mask_token=a , add_prefix_space=a , trim_offsets=a , **a , ) lowerCAmelCase__ : List[Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , a ) != add_prefix_space: lowerCAmelCase__ : Union[str, Any] = getattr(a , pre_tok_state.pop('type' ) ) lowerCAmelCase__ : int = add_prefix_space lowerCAmelCase__ : List[Any] = pre_tok_class(**a ) lowerCAmelCase__ : List[Any] = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowerCAmelCase__ : List[str] = 'post_processor' lowerCAmelCase__ : Union[str, Any] = getattr(self.backend_tokenizer , a , a ) if tokenizer_component_instance: lowerCAmelCase__ : Any = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCAmelCase__ : Union[str, Any] = tuple(state['sep'] ) if "cls" in state: lowerCAmelCase__ : Tuple = tuple(state['cls'] ) lowerCAmelCase__ : Tuple = False if state.get('add_prefix_space' , a ) != add_prefix_space: lowerCAmelCase__ : Optional[Any] = add_prefix_space lowerCAmelCase__ : int = True if state.get('trim_offsets' , a ) != trim_offsets: lowerCAmelCase__ : List[str] = trim_offsets lowerCAmelCase__ : List[str] = True if changes_to_apply: lowerCAmelCase__ : Optional[int] = getattr(a , state.pop('type' ) ) lowerCAmelCase__ : Optional[Any] = component_class(**a ) setattr(self.backend_tokenizer , a , a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def _lowerCamelCase ( self : str , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else value lowerCAmelCase__ : List[str] = value def _lowerCamelCase ( self : List[str] , *a : int , **a : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = kwargs.get('is_split_into_words' , a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' 'to use it with pretokenized inputs.' ) return super()._batch_encode_plus(*a , **a ) def _lowerCamelCase ( self : Optional[Any] , *a : Optional[Any] , **a : str ): '''simple docstring''' lowerCAmelCase__ : Tuple = kwargs.get('is_split_into_words' , a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' 'to use it with pretokenized inputs.' ) return super()._encode_plus(*a , **a ) def _lowerCamelCase ( self : str , a : str , a : Optional[str] = None ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self._tokenizer.model.save(a , name=a ) return tuple(a ) def _lowerCamelCase ( self : str , a : List[Any] , a : Any=None ): '''simple docstring''' lowerCAmelCase__ : str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowerCamelCase ( self : Optional[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
69
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
1
import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder lowerCamelCase__ = """__DUMMY_TRANSFORMERS_USER__""" lowerCamelCase__ = """Dummy User""" lowerCamelCase__ = """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""" lowerCamelCase__ = """https://hub-ci.huggingface.co""" lowerCamelCase__ = CI_HUB_ENDPOINT + """/datasets/{repo_id}/resolve/{revision}/{path}""" lowerCamelCase__ = CI_HUB_ENDPOINT + """/{repo_id}/resolve/{revision}/{filename}""" lowerCamelCase__ = Path("""~/.huggingface/hub_ci_token""").expanduser() @pytest.fixture def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Dict: monkeypatch.setattr( 'huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE' , SCREAMING_SNAKE_CASE_ ) @pytest.fixture def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: monkeypatch.setattr('datasets.config.HF_ENDPOINT' , SCREAMING_SNAKE_CASE_ ) monkeypatch.setattr('datasets.config.HUB_DATASETS_URL' , SCREAMING_SNAKE_CASE_ ) @pytest.fixture def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[Any]: monkeypatch.setattr('huggingface_hub.hf_api.HfFolder.path_token' , SCREAMING_SNAKE_CASE_ ) @pytest.fixture def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: HfFolder.save_token(SCREAMING_SNAKE_CASE_ ) yield HfFolder.delete_token() @pytest.fixture(scope='session' ) def lowerCAmelCase__ ( ) -> Optional[Any]: return HfApi(endpoint=SCREAMING_SNAKE_CASE_ ) @pytest.fixture(scope='session' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = HfFolder.get_token() HfFolder.save_token(SCREAMING_SNAKE_CASE_ ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(SCREAMING_SNAKE_CASE_ ) @pytest.fixture def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Optional[int]: def _cleanup_repo(SCREAMING_SNAKE_CASE_ ): hf_api.delete_repo(SCREAMING_SNAKE_CASE_ , token=SCREAMING_SNAKE_CASE_ , repo_type='dataset' ) return _cleanup_repo @pytest.fixture def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: @contextmanager def _temporary_repo(SCREAMING_SNAKE_CASE_ ): try: yield repo_id finally: cleanup_repo(SCREAMING_SNAKE_CASE_ ) return _temporary_repo @pytest.fixture(scope='session' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : List[Any] = F'''repo_txt_data-{int(time.time() * 10e3 )}''' lowerCAmelCase__ : Tuple = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(SCREAMING_SNAKE_CASE_ , token=SCREAMING_SNAKE_CASE_ , repo_type='dataset' , private=SCREAMING_SNAKE_CASE_ ) hf_api.upload_file( token=SCREAMING_SNAKE_CASE_ , path_or_fileobj=str(SCREAMING_SNAKE_CASE_ ) , path_in_repo='data/text_data.txt' , repo_id=SCREAMING_SNAKE_CASE_ , repo_type='dataset' , ) yield repo_id try: hf_api.delete_repo(SCREAMING_SNAKE_CASE_ , token=SCREAMING_SNAKE_CASE_ , repo_type='dataset' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope='session' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : int = F'''repo_zipped_txt_data-{int(time.time() * 10e3 )}''' lowerCAmelCase__ : int = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(SCREAMING_SNAKE_CASE_ , token=SCREAMING_SNAKE_CASE_ , repo_type='dataset' , private=SCREAMING_SNAKE_CASE_ ) hf_api.upload_file( token=SCREAMING_SNAKE_CASE_ , path_or_fileobj=str(SCREAMING_SNAKE_CASE_ ) , path_in_repo='data.zip' , repo_id=SCREAMING_SNAKE_CASE_ , repo_type='dataset' , ) yield repo_id try: hf_api.delete_repo(SCREAMING_SNAKE_CASE_ , token=SCREAMING_SNAKE_CASE_ , repo_type='dataset' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope='session' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: lowerCAmelCase__ : int = F'''repo_zipped_img_data-{int(time.time() * 10e3 )}''' lowerCAmelCase__ : Optional[int] = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(SCREAMING_SNAKE_CASE_ , token=SCREAMING_SNAKE_CASE_ , repo_type='dataset' , private=SCREAMING_SNAKE_CASE_ ) hf_api.upload_file( token=SCREAMING_SNAKE_CASE_ , path_or_fileobj=str(SCREAMING_SNAKE_CASE_ ) , path_in_repo='data.zip' , repo_id=SCREAMING_SNAKE_CASE_ , repo_type='dataset' , ) yield repo_id try: hf_api.delete_repo(SCREAMING_SNAKE_CASE_ , token=SCREAMING_SNAKE_CASE_ , repo_type='dataset' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> List[str]: return hf_private_dataset_repo_zipped_img_data_
69
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """google/rembert""": 256, } lowerCamelCase__ = """▁""" class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = RemBertTokenizer def __init__( self : Optional[Any] , a : str=None , a : Any=None , a : List[Any]=True , a : str=True , a : Dict=False , a : Dict="[CLS]" , a : int="[SEP]" , a : Tuple="<unk>" , a : Optional[Any]="[SEP]" , a : Tuple="<pad>" , a : Dict="[CLS]" , a : Optional[Any]="[MASK]" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( a , tokenizer_file=a , do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : int = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : List[Any] = keep_accents lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
69
1
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class A__ : @staticmethod def _lowerCamelCase ( *a : int , **a : List[str] ): '''simple docstring''' pass @is_pipeline_test @require_vision class A__ ( unittest.TestCase ): @require_torch def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = pipeline( model='hf-internal-testing/tiny-random-clip-zero-shot-image-classification' , ) lowerCAmelCase__ : Union[str, Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Optional[int] = image_classifier(a , candidate_labels=['a', 'b', 'c'] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(a ) , [ [{'score': 0.3_3_3, 'label': 'a'}, {'score': 0.3_3_3, 'label': 'b'}, {'score': 0.3_3_3, 'label': 'c'}], [{'score': 0.3_3_3, 'label': 'a'}, {'score': 0.3_3_3, 'label': 'c'}, {'score': 0.3_3_3, 'label': 'b'}], ] , ) lowerCAmelCase__ : str = image_classifier([image] * 5 , candidate_labels=['A', 'B', 'C'] , batch_size=2 ) self.assertEqual( nested_simplify(a ) , [ [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], ] , ) @require_tf def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = pipeline( model='hf-internal-testing/tiny-random-clip-zero-shot-image-classification' , framework='tf' ) lowerCAmelCase__ : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : List[Any] = image_classifier(a , candidate_labels=['a', 'b', 'c'] ) self.assertEqual( nested_simplify(a ) , [{'score': 0.3_3_3, 'label': 'a'}, {'score': 0.3_3_3, 'label': 'b'}, {'score': 0.3_3_3, 'label': 'c'}] , ) lowerCAmelCase__ : Optional[int] = image_classifier([image] * 5 , candidate_labels=['A', 'B', 'C'] , batch_size=2 ) self.assertEqual( nested_simplify(a ) , [ [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], ] , ) @slow @require_torch def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = pipeline( task='zero-shot-image-classification' , model='openai/clip-vit-base-patch32' , ) # This is an image of 2 cats with remotes and no planes lowerCAmelCase__ : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Optional[Any] = image_classifier(a , candidate_labels=['cat', 'plane', 'remote'] ) self.assertEqual( nested_simplify(a ) , [ {'score': 0.5_1_1, 'label': 'remote'}, {'score': 0.4_8_5, 'label': 'cat'}, {'score': 0.0_0_4, 'label': 'plane'}, ] , ) lowerCAmelCase__ : Dict = image_classifier([image] * 5 , candidate_labels=['cat', 'plane', 'remote'] , batch_size=2 ) self.assertEqual( nested_simplify(a ) , [ [ {'score': 0.5_1_1, 'label': 'remote'}, {'score': 0.4_8_5, 'label': 'cat'}, {'score': 0.0_0_4, 'label': 'plane'}, ], ] * 5 , ) @slow @require_tf def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = pipeline( task='zero-shot-image-classification' , model='openai/clip-vit-base-patch32' , framework='tf' ) # This is an image of 2 cats with remotes and no planes lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Optional[Any] = image_classifier(a , candidate_labels=['cat', 'plane', 'remote'] ) self.assertEqual( nested_simplify(a ) , [ {'score': 0.5_1_1, 'label': 'remote'}, {'score': 0.4_8_5, 'label': 'cat'}, {'score': 0.0_0_4, 'label': 'plane'}, ] , ) lowerCAmelCase__ : Any = image_classifier([image] * 5 , candidate_labels=['cat', 'plane', 'remote'] , batch_size=2 ) self.assertEqual( nested_simplify(a ) , [ [ {'score': 0.5_1_1, 'label': 'remote'}, {'score': 0.4_8_5, 'label': 'cat'}, {'score': 0.0_0_4, 'label': 'plane'}, ], ] * 5 , )
69
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : int = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : List[Any] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : int = None lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Dict = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Union[str, Any] = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
69
1
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
69
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
1
import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class A__ ( unittest.TestCase , __magic_name__ ): def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[str] = load_tool('text-to-speech' ) self.tool.setup() def _lowerCamelCase ( self : Tuple ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : str = self.tool('hey' ) lowerCAmelCase__ : Dict = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_0_0_5_9_6_6_6_6_8_8_3_2_1_1_5_8_2_9, -0.0_0_0_3_6_5_7_6_4_0_1_9_0_7_9_5_0_6_4, -0.0_0_0_1_3_4_3_9_5_0_2_7_9_9_8_8_3_4_8_5] ) , ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : str = self.tool('hey' ) lowerCAmelCase__ : List[str] = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_0_0_5_9_6_6_6_6_8_8_3_2_1_1_5_8_2_9, -0.0_0_0_3_6_5_7_6_4_0_1_9_0_7_9_5_0_6_4, -0.0_0_0_1_3_4_3_9_5_0_2_7_9_9_8_8_3_4_8_5] ) , ) )
69
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
1
import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") lowerCamelCase__ = logging.getLogger(__name__) @dataclass class A__ : lowercase = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) lowercase = field( default=__magic_name__ , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) lowercase = field( default=__magic_name__ , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) lowercase = field( default=__magic_name__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Whether to use one of the fast tokenizer (backed by the tokenizers library) or not.'} , ) lowercase = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) lowercase = field( default=__magic_name__ , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) @dataclass class A__ : lowercase = field(default=__magic_name__ , metadata={'help': 'The input training data file (a text file).'} ) lowercase = field( default=__magic_name__ , metadata={'help': 'An optional input evaluation data file to evaluate the perplexity on (a text file).'} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) lowercase = field( default=__magic_name__ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) lowercase = field( default=__magic_name__ , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. If passed, sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) lowercase = field( default=__magic_name__ , metadata={ 'help': ( 'Whether to pad all samples to the maximum sentence length. ' 'If False, will pad the samples dynamically when batching to the maximum length in the batch. More ' 'efficient on GPU but very bad for TPU.' ) } , ) lowercase = field( default=__magic_name__ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) lowercase = field( default=__magic_name__ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.train_file is not None: lowerCAmelCase__ : Tuple = self.train_file.split('.' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: lowerCAmelCase__ : List[str] = self.validation_file.split('.' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class A__ : lowercase = 42 lowercase = True lowercase = None lowercase = None def __call__( self : List[str] , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = 'label' if 'label' in features[0].keys() else 'labels' lowerCAmelCase__ : str = [feature.pop(a ) for feature in features] lowerCAmelCase__ : Union[str, Any] = len(a ) lowerCAmelCase__ : Optional[Any] = len(features[0]['input_ids'] ) lowerCAmelCase__ : str = [ [{k: v[i] for k, v in feature.items()} for i in range(a )] for feature in features ] lowerCAmelCase__ : Optional[int] = list(chain(*a ) ) lowerCAmelCase__ : str = self.tokenizer.pad( a , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) # Un-flatten lowerCAmelCase__ : Optional[int] = {k: v.view(a , a , -1 ) for k, v in batch.items()} # Add back labels lowerCAmelCase__ : int = torch.tensor(a , dtype=torch.intaa ) return batch def lowerCAmelCase__ ( ) -> int: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase__ : Dict = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_swag' , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowerCAmelCase__ : Dict = training_args.get_process_log_level() logger.setLevel(SCREAMING_SNAKE_CASE_ ) datasets.utils.logging.set_verbosity(SCREAMING_SNAKE_CASE_ ) transformers.utils.logging.set_verbosity(SCREAMING_SNAKE_CASE_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. lowerCAmelCase__ : Tuple = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCAmelCase__ : int = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: lowerCAmelCase__ : Union[str, Any] = {} if data_args.train_file is not None: lowerCAmelCase__ : Union[str, Any] = data_args.train_file if data_args.validation_file is not None: lowerCAmelCase__ : str = data_args.validation_file lowerCAmelCase__ : Union[str, Any] = data_args.train_file.split('.' )[-1] lowerCAmelCase__ : List[Any] = load_dataset( SCREAMING_SNAKE_CASE_ , data_files=SCREAMING_SNAKE_CASE_ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. lowerCAmelCase__ : Union[str, Any] = load_dataset( 'swag' , 'regular' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCAmelCase__ : List[str] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCAmelCase__ : Optional[Any] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCAmelCase__ : Tuple = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. lowerCAmelCase__ : Dict = [F'''ending{i}''' for i in range(4 )] lowerCAmelCase__ : Any = 'sent1' lowerCAmelCase__ : Dict = 'sent2' if data_args.max_seq_length is None: lowerCAmelCase__ : List[str] = tokenizer.model_max_length if max_seq_length > 1_024: logger.warning( 'The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value' ' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can' ' override this default with `--block_size xxx`.' ) lowerCAmelCase__ : Tuple = 1_024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' F'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) lowerCAmelCase__ : List[Any] = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : List[str] = [[context] * 4 for context in examples[context_name]] lowerCAmelCase__ : int = examples[question_header_name] lowerCAmelCase__ : List[str] = [ [F'''{header} {examples[end][i]}''' for end in ending_names] for i, header in enumerate(SCREAMING_SNAKE_CASE_ ) ] # Flatten out lowerCAmelCase__ : Union[str, Any] = list(chain(*SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase__ : Optional[Any] = list(chain(*SCREAMING_SNAKE_CASE_ ) ) # Tokenize lowerCAmelCase__ : Union[str, Any] = tokenizer( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , padding='max_length' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) lowerCAmelCase__ : Dict = raw_datasets['train'] if data_args.max_train_samples is not None: lowerCAmelCase__ : List[str] = min(len(SCREAMING_SNAKE_CASE_ ) , data_args.max_train_samples ) lowerCAmelCase__ : str = train_dataset.select(range(SCREAMING_SNAKE_CASE_ ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): lowerCAmelCase__ : int = train_dataset.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) lowerCAmelCase__ : Any = raw_datasets['validation'] if data_args.max_eval_samples is not None: lowerCAmelCase__ : Optional[int] = min(len(SCREAMING_SNAKE_CASE_ ) , data_args.max_eval_samples ) lowerCAmelCase__ : Dict = eval_dataset.select(range(SCREAMING_SNAKE_CASE_ ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): lowerCAmelCase__ : str = eval_dataset.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator lowerCAmelCase__ : List[str] = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=SCREAMING_SNAKE_CASE_ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = eval_predictions lowerCAmelCase__ : Tuple = np.argmax(SCREAMING_SNAKE_CASE_ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer lowerCAmelCase__ : Optional[int] = Trainer( model=SCREAMING_SNAKE_CASE_ , args=SCREAMING_SNAKE_CASE_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=SCREAMING_SNAKE_CASE_ , data_collator=SCREAMING_SNAKE_CASE_ , compute_metrics=SCREAMING_SNAKE_CASE_ , ) # Training if training_args.do_train: lowerCAmelCase__ : str = None if training_args.resume_from_checkpoint is not None: lowerCAmelCase__ : str = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowerCAmelCase__ : Union[str, Any] = last_checkpoint lowerCAmelCase__ : Optional[Any] = trainer.train(resume_from_checkpoint=SCREAMING_SNAKE_CASE_ ) trainer.save_model() # Saves the tokenizer too for easy upload lowerCAmelCase__ : Tuple = train_result.metrics lowerCAmelCase__ : List[str] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase__ : Optional[Any] = min(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ) trainer.log_metrics('train' , SCREAMING_SNAKE_CASE_ ) trainer.save_metrics('train' , SCREAMING_SNAKE_CASE_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) lowerCAmelCase__ : List[str] = trainer.evaluate() lowerCAmelCase__ : Optional[int] = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[int] = min(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ) trainer.log_metrics('eval' , SCREAMING_SNAKE_CASE_ ) trainer.save_metrics('eval' , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'multiple-choice', 'dataset_tags': 'swag', 'dataset_args': 'regular', 'dataset': 'SWAG', 'language': 'en', } if training_args.push_to_hub: trainer.push_to_hub(**SCREAMING_SNAKE_CASE_ ) else: trainer.create_model_card(**SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
69
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
1
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class A__ ( __magic_name__ ): def _lowerCamelCase ( self : Tuple , a : str ): '''simple docstring''' with open(a , encoding='utf-8' ) as input_file: lowerCAmelCase__ : Optional[Any] = re.compile(R'(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)' ) lowerCAmelCase__ : str = input_file.read() lowerCAmelCase__ : List[str] = regexp.search(a ) return match def _lowerCamelCase ( self : Dict , a : str ): '''simple docstring''' with open(a , encoding='utf-8' ) as input_file: lowerCAmelCase__ : Optional[int] = re.compile(R'#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()' , re.DOTALL ) lowerCAmelCase__ : Any = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` lowerCAmelCase__ : int = regexp.finditer(a ) lowerCAmelCase__ : List[str] = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = Path('./datasets' ) lowerCAmelCase__ : Optional[int] = list(dataset_paths.absolute().glob('**/*.py' ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(a ) ): raise AssertionError(f'''open(...) must use utf-8 encoding in {dataset}''' ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = Path('./datasets' ) lowerCAmelCase__ : List[str] = list(dataset_paths.absolute().glob('**/*.py' ) ) for dataset in dataset_files: if self._no_print_statements(str(a ) ): raise AssertionError(f'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
69
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
1
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class A__ : lowercase = 42 # [batch_size x 3] lowercase = 42 # [batch_size x 3] lowercase = 42 # [batch_size x 3] lowercase = 42 # [batch_size x 3] lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 def _lowerCamelCase ( self : str ): '''simple docstring''' assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = torch.arange(self.height * self.width ) lowerCAmelCase__ : List[Any] = torch.stack( [ pixel_indices % self.width, torch.div(a , self.width , rounding_mode='trunc' ), ] , axis=1 , ) return coords @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ , *lowerCAmelCase__ : Tuple = self.shape lowerCAmelCase__ : Dict = int(np.prod(a ) ) lowerCAmelCase__ : Union[str, Any] = self.get_image_coords() lowerCAmelCase__ : Union[str, Any] = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) lowerCAmelCase__ : Optional[Any] = self.get_camera_rays(a ) lowerCAmelCase__ : Any = rays.view(a , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def _lowerCamelCase ( self : Optional[int] , a : torch.Tensor ): '''simple docstring''' lowerCAmelCase__ , *lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] lowerCAmelCase__ : Union[str, Any] = coords.view(a , -1 , 2 ) lowerCAmelCase__ : Optional[int] = self.resolution() lowerCAmelCase__ : List[str] = self.fov() lowerCAmelCase__ : int = (flat.float() / (res - 1)) * 2 - 1 lowerCAmelCase__ : str = fracs * torch.tan(fov / 2 ) lowerCAmelCase__ : Tuple = fracs.view(a , -1 , 2 ) lowerCAmelCase__ : Optional[int] = ( self.z.view(a , 1 , 3 ) + self.x.view(a , 1 , 3 ) * fracs[:, :, :1] + self.y.view(a , 1 , 3 ) * fracs[:, :, 1:] ) lowerCAmelCase__ : List[Any] = directions / directions.norm(dim=-1 , keepdim=a ) lowerCAmelCase__ : Any = torch.stack( [ torch.broadcast_to(self.origin.view(a , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(a , *a , 2 , 3 ) def _lowerCamelCase ( self : Dict , a : int , a : int ): '''simple docstring''' assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=a , height=a , x_fov=self.x_fov , y_fov=self.y_fov , ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> DifferentiableProjectiveCamera: lowerCAmelCase__ : str = [] lowerCAmelCase__ : List[Any] = [] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Optional[int] = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): lowerCAmelCase__ : Optional[Any] = np.array([np.sin(SCREAMING_SNAKE_CASE_ ), np.cos(SCREAMING_SNAKE_CASE_ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) lowerCAmelCase__ : List[Any] = -z * 4 lowerCAmelCase__ : Tuple = np.array([np.cos(SCREAMING_SNAKE_CASE_ ), -np.sin(SCREAMING_SNAKE_CASE_ ), 0.0] ) lowerCAmelCase__ : Optional[Any] = np.cross(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) origins.append(SCREAMING_SNAKE_CASE_ ) xs.append(SCREAMING_SNAKE_CASE_ ) ys.append(SCREAMING_SNAKE_CASE_ ) zs.append(SCREAMING_SNAKE_CASE_ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(SCREAMING_SNAKE_CASE_ , axis=0 ) ).float() , x=torch.from_numpy(np.stack(SCREAMING_SNAKE_CASE_ , axis=0 ) ).float() , y=torch.from_numpy(np.stack(SCREAMING_SNAKE_CASE_ , axis=0 ) ).float() , z=torch.from_numpy(np.stack(SCREAMING_SNAKE_CASE_ , axis=0 ) ).float() , width=SCREAMING_SNAKE_CASE_ , height=SCREAMING_SNAKE_CASE_ , x_fov=0.7 , y_fov=0.7 , shape=(1, len(SCREAMING_SNAKE_CASE_ )) , )
69
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {} class A__ ( __magic_name__ ): lowercase = 'llama' lowercase = ['past_key_values'] def __init__( self : int , a : Dict=32_000 , a : Union[str, Any]=4_096 , a : str=11_008 , a : List[str]=32 , a : Tuple=32 , a : int=None , a : List[Any]="silu" , a : Optional[Any]=2_048 , a : Any=0.0_2 , a : List[str]=1E-6 , a : Tuple=True , a : Union[str, Any]=0 , a : int=1 , a : List[Any]=2 , a : Optional[Any]=1 , a : Dict=False , a : Optional[Any]=None , **a : Optional[Any] , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = vocab_size lowerCAmelCase__ : str = max_position_embeddings lowerCAmelCase__ : Optional[int] = hidden_size lowerCAmelCase__ : str = intermediate_size lowerCAmelCase__ : Any = num_hidden_layers lowerCAmelCase__ : List[Any] = num_attention_heads # for backward compatibility if num_key_value_heads is None: lowerCAmelCase__ : List[Any] = num_attention_heads lowerCAmelCase__ : str = num_key_value_heads lowerCAmelCase__ : str = hidden_act lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[str] = rms_norm_eps lowerCAmelCase__ : str = pretraining_tp lowerCAmelCase__ : Tuple = use_cache lowerCAmelCase__ : Dict = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=a , bos_token_id=a , eos_token_id=a , tie_word_embeddings=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling , a ) or len(self.rope_scaling ) != 2: raise ValueError( '`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ' f'''got {self.rope_scaling}''' ) lowerCAmelCase__ : str = self.rope_scaling.get('type' , a ) lowerCAmelCase__ : Tuple = self.rope_scaling.get('factor' , a ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f'''`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}''' ) if rope_scaling_factor is None or not isinstance(a , a ) or rope_scaling_factor <= 1.0: raise ValueError(f'''`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}''' )
69
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) lowerCAmelCase__ : List[Any] = get_activation('gelu' ) self.assertTrue(torch.allclose(gelu_python(a ) , torch_builtin(a ) ) ) self.assertFalse(torch.allclose(gelu_python(a ) , gelu_new(a ) ) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) lowerCAmelCase__ : Dict = get_activation('gelu' ) lowerCAmelCase__ : Tuple = get_activation('gelu_10' ) lowerCAmelCase__ : Dict = torch_builtin(a ) lowerCAmelCase__ : List[str] = geluaa(a ) lowerCAmelCase__ : List[Any] = torch.where(y_gelu_aa < 1_0.0 , 1 , 0 ) self.assertTrue(torch.max(a ).item() == 1_0.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' get_activation('gelu' ) get_activation('gelu_10' ) get_activation('gelu_fast' ) get_activation('gelu_new' ) get_activation('gelu_python' ) get_activation('gelu_pytorch_tanh' ) get_activation('linear' ) get_activation('mish' ) get_activation('quick_gelu' ) get_activation('relu' ) get_activation('sigmoid' ) get_activation('silu' ) get_activation('swish' ) get_activation('tanh' ) with self.assertRaises(a ): get_activation('bogus' ) with self.assertRaises(a ): get_activation(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = get_activation('gelu' ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Any = get_activation('gelu' ) self.assertEqual(acta.a , 1 ) with self.assertRaises(a ): lowerCAmelCase__ : Optional[int] = acta.a
69
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
1
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __magic_name__ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'LayoutLMv3ImageProcessor' lowercase = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self : Optional[int] , a : Union[str, Any]=None , a : Optional[Any]=None , **a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , a , ) lowerCAmelCase__ : int = kwargs.pop('feature_extractor' ) lowerCAmelCase__ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(a , a ) def __call__( self : List[Any] , a : List[Any] , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a : Union[List[List[int]], List[List[List[int]]]] = None , a : Optional[Union[List[int], List[List[int]]]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : str , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor lowerCAmelCase__ : List[str] = self.image_processor(images=a , return_tensors=a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a , a ): lowerCAmelCase__ : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ : List[str] = features['words'] lowerCAmelCase__ : List[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_token_type_ids=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_length=a , verbose=a , return_tensors=a , **a , ) # add pixel values lowerCAmelCase__ : Tuple = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase__ : List[str] = self.get_overflowing_images(a , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase__ : List[str] = images return encoded_inputs def _lowerCamelCase ( self : Any , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a ) != len(a ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' f''' {len(a )} and {len(a )}''' ) return images_with_overflow def _lowerCamelCase ( self : Union[str, Any] , *a : Optional[Any] , **a : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Tuple , *a : List[str] , **a : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*a , **a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , a , ) return self.image_processor_class @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , a , ) return self.image_processor
69
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
1
from ...processing_utils import ProcessorMixin class A__ ( __magic_name__ ): lowercase = 'SpeechT5FeatureExtractor' lowercase = 'SpeechT5Tokenizer' def __init__( self : Dict , a : int , a : str ): '''simple docstring''' super().__init__(a , a ) def __call__( self : List[str] , *a : Tuple , **a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = kwargs.pop('audio' , a ) lowerCAmelCase__ : str = kwargs.pop('text' , a ) lowerCAmelCase__ : Union[str, Any] = kwargs.pop('text_target' , a ) lowerCAmelCase__ : List[Any] = kwargs.pop('audio_target' , a ) lowerCAmelCase__ : Optional[Any] = kwargs.pop('sampling_rate' , a ) if audio is not None and text is not None: raise ValueError( 'Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?' ) if audio_target is not None and text_target is not None: raise ValueError( 'Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?' ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( 'You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.' ) if audio is not None: lowerCAmelCase__ : Optional[int] = self.feature_extractor(a , *a , sampling_rate=a , **a ) elif text is not None: lowerCAmelCase__ : List[str] = self.tokenizer(a , **a ) else: lowerCAmelCase__ : str = None if audio_target is not None: lowerCAmelCase__ : List[str] = self.feature_extractor(audio_target=a , *a , sampling_rate=a , **a ) lowerCAmelCase__ : Optional[Any] = targets['input_values'] elif text_target is not None: lowerCAmelCase__ : Tuple = self.tokenizer(a , **a ) lowerCAmelCase__ : Dict = targets['input_ids'] else: lowerCAmelCase__ : str = None if inputs is None: return targets if targets is not None: lowerCAmelCase__ : List[Any] = labels lowerCAmelCase__ : Any = targets.get('attention_mask' ) if decoder_attention_mask is not None: lowerCAmelCase__ : List[Any] = decoder_attention_mask return inputs def _lowerCamelCase ( self : List[str] , *a : Dict , **a : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = kwargs.pop('input_values' , a ) lowerCAmelCase__ : Any = kwargs.pop('input_ids' , a ) lowerCAmelCase__ : Any = kwargs.pop('labels' , a ) if input_values is not None and input_ids is not None: raise ValueError('Cannot process both `input_values` and `input_ids` inputs.' ) if input_values is None and input_ids is None and labels is None: raise ValueError( 'You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.' ) if input_values is not None: lowerCAmelCase__ : Dict = self.feature_extractor.pad(a , *a , **a ) elif input_ids is not None: lowerCAmelCase__ : List[str] = self.tokenizer.pad(a , **a ) else: lowerCAmelCase__ : Any = None if labels is not None: if "input_ids" in labels or (isinstance(a , a ) and "input_ids" in labels[0]): lowerCAmelCase__ : Union[str, Any] = self.tokenizer.pad(a , **a ) lowerCAmelCase__ : Dict = targets['input_ids'] else: lowerCAmelCase__ : List[Any] = self.feature_extractor.feature_size lowerCAmelCase__ : Union[str, Any] = self.feature_extractor.num_mel_bins lowerCAmelCase__ : Optional[Any] = self.feature_extractor.pad(a , *a , **a ) lowerCAmelCase__ : Dict = feature_size_hack lowerCAmelCase__ : List[str] = targets['input_values'] else: lowerCAmelCase__ : List[str] = None if inputs is None: return targets if targets is not None: lowerCAmelCase__ : str = labels lowerCAmelCase__ : str = targets.get('attention_mask' ) if decoder_attention_mask is not None: lowerCAmelCase__ : Tuple = decoder_attention_mask return inputs def _lowerCamelCase ( self : Any , *a : List[str] , **a : Dict ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Union[str, Any] , *a : List[Any] , **a : Dict ): '''simple docstring''' return self.tokenizer.decode(*a , **a )
69
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
1
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = KandinskyVaaImgaImgPipeline lowercase = ['image_embeds', 'negative_image_embeds', 'image'] lowercase = [ 'image_embeds', 'negative_image_embeds', 'image', ] lowercase = [ 'generator', 'height', 'width', 'strength', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] lowercase = False @property def _lowerCamelCase ( self : int ): '''simple docstring''' return 32 @property def _lowerCamelCase ( self : int ): '''simple docstring''' return 32 @property def _lowerCamelCase ( self : Any ): '''simple docstring''' return self.time_input_dim @property def _lowerCamelCase ( self : Any ): '''simple docstring''' return self.time_input_dim * 4 @property def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' return 100 @property def _lowerCamelCase ( self : Any ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : List[str] = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } lowerCAmelCase__ : Union[str, Any] = UNetaDConditionModel(**a ) return model @property def _lowerCamelCase ( self : str ): '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : Any = VQModel(**self.dummy_movq_kwargs ) return model def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.dummy_unet lowerCAmelCase__ : str = self.dummy_movq lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_schedule': 'linear', 'beta_start': 0.0_0_0_8_5, 'beta_end': 0.0_1_2, 'clip_sample': False, 'set_alpha_to_one': False, 'steps_offset': 0, 'prediction_type': 'epsilon', 'thresholding': False, } lowerCAmelCase__ : Tuple = DDIMScheduler(**a ) lowerCAmelCase__ : Dict = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def _lowerCamelCase ( self : List[str] , a : List[str] , a : List[Any]=0 ): '''simple docstring''' lowerCAmelCase__ : List[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(a ) ).to(a ) lowerCAmelCase__ : Optional[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( a ) # create init_image lowerCAmelCase__ : Optional[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(a ) ).to(a ) lowerCAmelCase__ : Dict = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase__ : Any = Image.fromarray(np.uinta(a ) ).convert('RGB' ).resize((256, 256) ) if str(a ).startswith('mps' ): lowerCAmelCase__ : int = torch.manual_seed(a ) else: lowerCAmelCase__ : Any = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : List[str] = { 'image': init_image, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 10, 'guidance_scale': 7.0, 'strength': 0.2, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = 'cpu' lowerCAmelCase__ : Any = self.get_dummy_components() lowerCAmelCase__ : str = self.pipeline_class(**a ) lowerCAmelCase__ : List[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = pipe(**self.get_dummy_inputs(a ) ) lowerCAmelCase__ : Optional[int] = output.images lowerCAmelCase__ : int = pipe( **self.get_dummy_inputs(a ) , return_dict=a , )[0] lowerCAmelCase__ : List[Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = np.array( [0.6_1_9_9_7_7_8, 0.6_3_9_8_4_4_0_6, 0.4_6_1_4_5_7_8_5, 0.6_2_9_4_4_9_8_4, 0.5_6_2_2_2_1_5, 0.4_7_3_0_6_1_3_2, 0.4_7_4_4_1_4_5_6, 0.4_6_0_7_6_0_6, 0.4_8_7_1_9_2_6_3] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Tuple ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_img2img_frog.npy' ) lowerCAmelCase__ : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) lowerCAmelCase__ : Optional[int] = 'A red cartoon frog, 4k' lowerCAmelCase__ : Dict = KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(a ) lowerCAmelCase__ : str = KandinskyVaaImgaImgPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-decoder' , torch_dtype=torch.floataa ) lowerCAmelCase__ : Optional[int] = pipeline.to(a ) pipeline.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[int] = torch.Generator(device='cpu' ).manual_seed(0 ) lowerCAmelCase__ , lowerCAmelCase__ : str = pipe_prior( a , generator=a , num_inference_steps=5 , negative_prompt='' , ).to_tuple() lowerCAmelCase__ : Optional[int] = pipeline( image=a , image_embeds=a , negative_image_embeds=a , generator=a , num_inference_steps=100 , height=768 , width=768 , strength=0.2 , output_type='np' , ) lowerCAmelCase__ : Tuple = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(a , a )
69
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class A__ ( __magic_name__ ): lowercase = 'openai-gpt' lowercase = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : str , a : List[str]=40_478 , a : Union[str, Any]=512 , a : List[str]=768 , a : str=12 , a : Dict=12 , a : str="gelu" , a : Dict=0.1 , a : Any=0.1 , a : Optional[int]=0.1 , a : str=1E-5 , a : Dict=0.0_2 , a : Any="cls_index" , a : Tuple=True , a : Union[str, Any]=None , a : Union[str, Any]=True , a : Any=0.1 , **a : int , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = vocab_size lowerCAmelCase__ : str = n_positions lowerCAmelCase__ : str = n_embd lowerCAmelCase__ : Dict = n_layer lowerCAmelCase__ : Any = n_head lowerCAmelCase__ : int = afn lowerCAmelCase__ : Dict = resid_pdrop lowerCAmelCase__ : Union[str, Any] = embd_pdrop lowerCAmelCase__ : int = attn_pdrop lowerCAmelCase__ : List[Any] = layer_norm_epsilon lowerCAmelCase__ : Union[str, Any] = initializer_range lowerCAmelCase__ : Optional[Any] = summary_type lowerCAmelCase__ : int = summary_use_proj lowerCAmelCase__ : Tuple = summary_activation lowerCAmelCase__ : Union[str, Any] = summary_first_dropout lowerCAmelCase__ : str = summary_proj_to_labels super().__init__(**a )
69
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
69
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available lowerCamelCase__ = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
700
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __magic_name__ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'LayoutLMv3ImageProcessor' lowercase = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self : Optional[int] , a : Union[str, Any]=None , a : Optional[Any]=None , **a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , a , ) lowerCAmelCase__ : int = kwargs.pop('feature_extractor' ) lowerCAmelCase__ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(a , a ) def __call__( self : List[Any] , a : List[Any] , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a : Union[List[List[int]], List[List[List[int]]]] = None , a : Optional[Union[List[int], List[List[int]]]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : str , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor lowerCAmelCase__ : List[str] = self.image_processor(images=a , return_tensors=a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a , a ): lowerCAmelCase__ : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ : List[str] = features['words'] lowerCAmelCase__ : List[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_token_type_ids=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_length=a , verbose=a , return_tensors=a , **a , ) # add pixel values lowerCAmelCase__ : Tuple = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase__ : List[str] = self.get_overflowing_images(a , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase__ : List[str] = images return encoded_inputs def _lowerCamelCase ( self : Any , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a ) != len(a ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' f''' {len(a )} and {len(a )}''' ) return images_with_overflow def _lowerCamelCase ( self : Union[str, Any] , *a : Optional[Any] , **a : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Tuple , *a : List[str] , **a : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*a , **a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , a , ) return self.image_processor_class @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , a , ) return self.image_processor
69
0
import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """vocab.json"""} lowerCamelCase__ = { """vocab_file""": { """mgp-str""": """https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json""", } } lowerCamelCase__ = {"""mgp-str""": 27} class A__ ( _snake_case ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : List[str] , a : Any , a : Dict="[GO]" , a : List[str]="[GO]" , a : Tuple="[s]" , a : Union[str, Any]="[GO]" , **a : List[str] ): '''simple docstring''' super().__init__( unk_token=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) with open(lowerCAmelCase__ , encoding='utf-8' ) as vocab_handle: lowerCAmelCase__ : Dict = json.load(lowerCAmelCase__ ) lowerCAmelCase__ : Tuple = {v: k for k, v in self.vocab.items()} @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return len(self.vocab ) def _lowerCamelCase ( self : Any ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def _lowerCamelCase ( self : Dict , a : str ): '''simple docstring''' lowerCAmelCase__ : Dict = [] for s in text: char_tokens.extend(lowerCAmelCase__ ) return char_tokens def _lowerCamelCase ( self : Union[str, Any] , a : Union[str, Any] ): '''simple docstring''' return self.vocab.get(lowerCAmelCase__ , self.vocab.get(self.unk_token ) ) def _lowerCamelCase ( self : Tuple , a : Dict ): '''simple docstring''' return self.decoder.get(lowerCAmelCase__ ) def _lowerCamelCase ( self : Dict , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(lowerCAmelCase__ ): logger.error('Vocabulary path ({}) should be a directory'.format(lowerCAmelCase__ ) ) return lowerCAmelCase__ : int = os.path.join( lowerCAmelCase__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) with open(lowerCAmelCase__ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=lowerCAmelCase__ , ensure_ascii=lowerCAmelCase__ ) + '\n' ) return (vocab_file,)
701
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
69
0
lowerCamelCase__ = 'Input must be a string of 8 numbers plus letter' lowerCamelCase__ = 'TRWAGMYFPDXBNJZSQVHLCKE' def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : List[Any] = F'''Expected string as input, found {type(_SCREAMING_SNAKE_CASE ).__name__}''' raise TypeError(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = spanish_id.replace('-' , '' ).upper() if len(_SCREAMING_SNAKE_CASE ) != 9: raise ValueError(_SCREAMING_SNAKE_CASE ) try: lowerCAmelCase__ : List[str] = int(spanish_id_clean[0:8] ) lowerCAmelCase__ : List[str] = spanish_id_clean[8] except ValueError as ex: raise ValueError(_SCREAMING_SNAKE_CASE ) from ex if letter.isdigit(): raise ValueError(_SCREAMING_SNAKE_CASE ) return letter == LOOKUP_LETTERS[number % 23] if __name__ == "__main__": import doctest doctest.testmod()
702
lowerCamelCase__ = """Alexander Joslin""" import operator as op from .stack import Stack def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} lowerCAmelCase__ : Stack[int] = Stack() lowerCAmelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE_ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE_ ) elif i == ")": # RULE 4 lowerCAmelCase__ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : List[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : Tuple = operators[opr](SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) operand_stack.push(SCREAMING_SNAKE_CASE_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
69
0
from collections.abc import Generator def lowerCAmelCase__ ( ) -> Generator[int, None, None]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = 0, 1 while True: lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = b, a + b yield b def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 1_000 ) -> int: lowerCAmelCase__ : str = 1 lowerCAmelCase__ : List[str] = fibonacci_generator() while len(str(next(SCREAMING_SNAKE_CASE_ ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
703
import numpy class A__ : def __init__( self : Tuple , a : numpy.ndarray , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : int = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Dict = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : List[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowerCAmelCase__ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : List[Any] = numpy.zeros(output_array.shape ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowerCAmelCase__ : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowerCAmelCase__ : int = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowerCamelCase ( self : Optional[int] , a : numpy.ndarray , a : int , a : bool ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): lowerCAmelCase__ : Any = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Tuple = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def _lowerCamelCase ( self : Optional[Any] , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : Dict = input_arr lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return (value) * (1 - (value)) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : List[str] = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ , output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ , iterations=10 , give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> list: lowerCAmelCase__ : List[str] = word.split() def justify(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: lowerCAmelCase__ : Dict = max_width - width lowerCAmelCase__ : Tuple = len(_lowercase ) if len(_lowercase ) == 1: # if there is only word in line # just insert overall_spaces_count for the remainder of line return line[0] + " " * overall_spaces_count else: lowerCAmelCase__ : Tuple = words_count - 1 # num_spaces_between_words_list[i] : tells you to insert # num_spaces_between_words_list[i] spaces # after word on line[i] lowerCAmelCase__ : str = spaces_to_insert_between_words * [ overall_spaces_count // spaces_to_insert_between_words ] lowerCAmelCase__ : Optional[int] = ( overall_spaces_count % spaces_to_insert_between_words ) # distribute spaces via round robin to the left words for i in range(_lowercase ): num_spaces_between_words_list[i] += 1 lowerCAmelCase__ : Union[str, Any] = [] for i in range(_lowercase ): # add the word aligned_words_list.append(line[i] ) # add the spaces to insert aligned_words_list.append(num_spaces_between_words_list[i] * ' ' ) # just add the last word to the sentence aligned_words_list.append(line[-1] ) # join the aligned words list to form a justified line return "".join(_lowercase ) lowerCAmelCase__ : str = [] lowerCAmelCase__ : list[str] = [] lowerCAmelCase__ : Union[str, Any] = 0 for word in words: if width + len(_lowercase ) + len(_lowercase ) <= max_width: # keep adding words until we can fill out max_width # width = sum of length of all words (without overall_spaces_count) # len(word) = length of current word # len(line) = number of overall_spaces_count to insert between words line.append(_lowercase ) width += len(_lowercase ) else: # justify the line and add it to result answer.append(justify(_lowercase , _lowercase , _lowercase ) ) # reset new line and new width lowerCAmelCase__ : Optional[Any] = [word], len(_lowercase ) lowerCAmelCase__ : Optional[int] = max_width - width - len(_lowercase ) answer.append(' '.join(_lowercase ) + (remaining_spaces + 1) * ' ' ) return answer if __name__ == "__main__": from doctest import testmod testmod()
704
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A__ : def __init__( self : int , a : str , a : Union[str, Any]=13 , a : int=32 , a : Optional[Any]=2 , a : Tuple=3 , a : List[Any]=16 , a : List[str]=[1, 2, 1] , a : int=[2, 2, 4] , a : int=2 , a : Optional[Any]=2.0 , a : Optional[int]=True , a : Dict=0.0 , a : Any=0.0 , a : int=0.1 , a : List[str]="gelu" , a : Optional[Any]=False , a : str=True , a : Dict=0.0_2 , a : Any=1E-5 , a : Optional[int]=True , a : str=None , a : str=True , a : int=10 , a : str=8 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : Optional[Any] = patch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = embed_dim lowerCAmelCase__ : Tuple = depths lowerCAmelCase__ : List[str] = num_heads lowerCAmelCase__ : List[Any] = window_size lowerCAmelCase__ : Any = mlp_ratio lowerCAmelCase__ : Optional[Any] = qkv_bias lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : int = drop_path_rate lowerCAmelCase__ : Optional[Any] = hidden_act lowerCAmelCase__ : int = use_absolute_embeddings lowerCAmelCase__ : List[str] = patch_norm lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Dict = use_labels lowerCAmelCase__ : List[Any] = type_sequence_label_size lowerCAmelCase__ : Optional[Any] = encoder_stride def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = None if self.use_labels: lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : int = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : List[str] , a : Any , a : str , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = SwinvaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model(a ) lowerCAmelCase__ : str = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase__ : Tuple = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SwinvaForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : Dict = SwinvaForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Union[str, Any] , a : int , a : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.type_sequence_label_size lowerCAmelCase__ : List[Any] = SwinvaForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = config_and_inputs lowerCAmelCase__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = SwinvaModelTester(self ) lowerCAmelCase__ : int = ConfigTester(self , config_class=a , embed_dim=37 ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[int] = model_class(a ) lowerCAmelCase__ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Tuple = [*signature.parameters.keys()] lowerCAmelCase__ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = True for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = True lowerCAmelCase__ : str = False lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Dict = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : int = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Dict = outputs.attentions lowerCAmelCase__ : Dict = len(self.model_tester.depths ) self.assertEqual(len(a ) , a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Optional[int] = config.window_size**2 lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowerCAmelCase__ : Tuple = len(a ) # Check attention is always last and order is fine lowerCAmelCase__ : str = True lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(**self._prepare_for_class(a , a ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): lowerCAmelCase__ : Optional[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase__ : Any = 2 self.assertEqual(out_len + added_hidden_states , len(a ) ) lowerCAmelCase__ : Dict = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _lowerCamelCase ( self : int , a : Optional[int] , a : int , a : Optional[Any] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.hidden_states lowerCAmelCase__ : str = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a ) , a ) # Swinv2 has a different seq_length lowerCAmelCase__ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowerCAmelCase__ : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(a ) , a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = reshaped_hidden_states[0].shape lowerCAmelCase__ : List[str] = ( reshaped_hidden_states[0].view(a , a , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : str = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : List[str] = True self.check_hidden_states_output(a , a , a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Any = 3 lowerCAmelCase__ : int = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase__ : str = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : str = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase__ : Any = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase__ : str = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[str] = SwinvaModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = _config_zero_init(a ) for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(config=a ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( a ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Any = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**a ) # verify the logits lowerCAmelCase__ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Optional[Any] = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) )
69
0
from manim import * class A__ ( lowercase__ ): def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : str = Rectangle(height=0.5 , width=0.5 ) lowerCAmelCase__ : Union[str, Any] = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) lowerCAmelCase__ : Any = [mem.copy() for i in range(6 )] lowerCAmelCase__ : List[Any] = [mem.copy() for i in range(6 )] lowerCAmelCase__ : List[Any] = VGroup(*__lowerCamelCase ).arrange(__lowerCamelCase , buff=0 ) lowerCAmelCase__ : Optional[int] = VGroup(*__lowerCamelCase ).arrange(__lowerCamelCase , buff=0 ) lowerCAmelCase__ : Tuple = VGroup(__lowerCamelCase , __lowerCamelCase ).arrange(__lowerCamelCase , buff=0 ) lowerCAmelCase__ : Optional[int] = Text('CPU' , font_size=24 ) lowerCAmelCase__ : List[Any] = Group(__lowerCamelCase , __lowerCamelCase ).arrange(__lowerCamelCase , buff=0.5 , aligned_edge=__lowerCamelCase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__lowerCamelCase ) lowerCAmelCase__ : List[Any] = [mem.copy() for i in range(1 )] lowerCAmelCase__ : Any = VGroup(*__lowerCamelCase ).arrange(__lowerCamelCase , buff=0 ) lowerCAmelCase__ : int = Text('GPU' , font_size=24 ) lowerCAmelCase__ : Union[str, Any] = Group(__lowerCamelCase , __lowerCamelCase ).arrange(__lowerCamelCase , buff=0.5 , aligned_edge=__lowerCamelCase ) gpu.align_to(__lowerCamelCase , __lowerCamelCase ) gpu.set_x(gpu.get_x() - 1 ) self.add(__lowerCamelCase ) lowerCAmelCase__ : List[Any] = [mem.copy() for i in range(6 )] lowerCAmelCase__ : List[Any] = VGroup(*__lowerCamelCase ).arrange(__lowerCamelCase , buff=0 ) lowerCAmelCase__ : int = Text('Model' , font_size=24 ) lowerCAmelCase__ : Optional[Any] = Group(__lowerCamelCase , __lowerCamelCase ).arrange(__lowerCamelCase , buff=0.5 , aligned_edge=__lowerCamelCase ) model.move_to([3, -1.0, 0] ) self.play( Create(__lowerCamelCase , run_time=1 ) , Create(__lowerCamelCase , run_time=1 ) , Create(__lowerCamelCase , run_time=1 ) , ) lowerCAmelCase__ : Union[str, Any] = MarkupText( f'''First, an empty model skeleton is loaded\ninto <span fgcolor=\'{YELLOW}\'>memory</span> without using much RAM.''' , font_size=24 , ) lowerCAmelCase__ : Union[str, Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowerCAmelCase__ : Tuple = MarkupText( f'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(__lowerCamelCase , run_time=2.5 ) , Write(__lowerCamelCase ) , Write(__lowerCamelCase ) ) self.add(__lowerCamelCase ) lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : List[Any] = [] for i, rect in enumerate(__lowerCamelCase ): lowerCAmelCase__ : Dict = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0.0 ).set_fill(__lowerCamelCase , opacity=0.7 ) cpu_target.move_to(__lowerCamelCase ) cpu_target.generate_target() lowerCAmelCase__ : Any = 0.4_6 / 4 lowerCAmelCase__ : str = 0.4_6 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=__lowerCamelCase ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target , direction=__lowerCamelCase , buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target , direction=__lowerCamelCase , buff=0.0 ) cpu_targs.append(__lowerCamelCase ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(__lowerCamelCase ) ) second_animations.append(MoveToTarget(__lowerCamelCase , run_time=1.5 ) ) self.play(*__lowerCamelCase ) self.play(*__lowerCamelCase ) self.wait()
705
from itertools import permutations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase__ : str = [7, 11, 13, 17] for i, test in enumerate(SCREAMING_SNAKE_CASE_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 10 ) -> int: return sum( int(''.join(map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) ) for num in permutations(range(SCREAMING_SNAKE_CASE_ ) ) if is_substring_divisible(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
69
0
import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class A__ ( __lowerCamelCase ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'BlipImageProcessor' lowercase = 'AutoTokenizer' def __init__( self : Tuple , a : Tuple , a : int , a : Tuple ): '''simple docstring''' super().__init__(UpperCAmelCase_ , UpperCAmelCase_ ) # add QFormer tokenizer lowerCAmelCase__ : Tuple = qformer_tokenizer def __call__( self : List[Any] , a : ImageInput = None , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : List[str] , ): '''simple docstring''' if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) lowerCAmelCase__ : Tuple = BatchFeature() if text is not None: lowerCAmelCase__ : List[str] = self.tokenizer( text=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=UpperCAmelCase_ , stride=UpperCAmelCase_ , pad_to_multiple_of=UpperCAmelCase_ , return_attention_mask=UpperCAmelCase_ , return_overflowing_tokens=UpperCAmelCase_ , return_special_tokens_mask=UpperCAmelCase_ , return_offsets_mapping=UpperCAmelCase_ , return_token_type_ids=UpperCAmelCase_ , return_length=UpperCAmelCase_ , verbose=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ , ) encoding.update(UpperCAmelCase_ ) lowerCAmelCase__ : int = self.qformer_tokenizer( text=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=UpperCAmelCase_ , stride=UpperCAmelCase_ , pad_to_multiple_of=UpperCAmelCase_ , return_attention_mask=UpperCAmelCase_ , return_overflowing_tokens=UpperCAmelCase_ , return_special_tokens_mask=UpperCAmelCase_ , return_offsets_mapping=UpperCAmelCase_ , return_token_type_ids=UpperCAmelCase_ , return_length=UpperCAmelCase_ , verbose=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase__ : List[Any] = qformer_text_encoding.pop('input_ids' ) lowerCAmelCase__ : Any = qformer_text_encoding.pop('attention_mask' ) if images is not None: lowerCAmelCase__ : Optional[int] = self.image_processor(UpperCAmelCase_ , return_tensors=UpperCAmelCase_ ) encoding.update(UpperCAmelCase_ ) return encoding def _lowerCamelCase ( self : List[str] , *a : Union[str, Any] , **a : Union[str, Any] ): '''simple docstring''' return self.tokenizer.batch_decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) def _lowerCamelCase ( self : Optional[Any] , *a : List[str] , **a : Union[str, Any] ): '''simple docstring''' return self.tokenizer.decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.tokenizer.model_input_names lowerCAmelCase__ : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _lowerCamelCase ( self : List[Any] , a : str , **a : Union[str, Any] ): '''simple docstring''' if os.path.isfile(UpperCAmelCase_ ): raise ValueError(f'''Provided path ({save_directory}) should be a directory, not a file''' ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) lowerCAmelCase__ : Union[str, Any] = os.path.join(UpperCAmelCase_ , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(UpperCAmelCase_ ) return super().save_pretrained(UpperCAmelCase_ , **UpperCAmelCase_ ) @classmethod def _lowerCamelCase ( cls : Dict , a : str , **a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = AutoTokenizer.from_pretrained(UpperCAmelCase_ , subfolder='qformer_tokenizer' ) lowerCAmelCase__ : Union[str, Any] = cls._get_arguments_from_pretrained(UpperCAmelCase_ , **UpperCAmelCase_ ) args.append(UpperCAmelCase_ ) return cls(*UpperCAmelCase_ )
706
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
69
0
lowerCamelCase__ = { 0: "0", 1: "1", 2: "2", 3: "3", 4: "4", 5: "5", 6: "6", 7: "7", 8: "8", 9: "9", 10: "a", 11: "b", 12: "c", 13: "d", 14: "e", 15: "f", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: assert type(SCREAMING_SNAKE_CASE_ ) in (int, float) and decimal == int(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Tuple = int(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = '' lowerCAmelCase__ : Tuple = False if decimal < 0: lowerCAmelCase__ : Optional[Any] = True decimal *= -1 while decimal > 0: lowerCAmelCase__ : Dict = divmod(SCREAMING_SNAKE_CASE_ , 16 ) lowerCAmelCase__ : Tuple = values[remainder] + hexadecimal lowerCAmelCase__ : Optional[Any] = '0x' + hexadecimal if negative: lowerCAmelCase__ : str = '-' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
707
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ , lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
69
0
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: lowerCAmelCase__ : List[str] = TapasConfig.from_json_file(__A ) # set absolute/relative position embeddings parameter lowerCAmelCase__ : List[str] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": lowerCAmelCase__ : Optional[int] = TapasForQuestionAnswering(config=__A ) elif task == "WTQ": # run_task_main.py hparams lowerCAmelCase__ : List[Any] = 4 lowerCAmelCase__ : List[Any] = True # hparam_utils.py hparams lowerCAmelCase__ : Dict = 0.664694 lowerCAmelCase__ : Any = 0.207951 lowerCAmelCase__ : List[str] = 0.121194 lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Any = True lowerCAmelCase__ : Dict = False lowerCAmelCase__ : Union[str, Any] = 0.0352513 lowerCAmelCase__ : Any = TapasForQuestionAnswering(config=__A ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams lowerCAmelCase__ : str = 4 lowerCAmelCase__ : List[Any] = False # hparam_utils.py hparams lowerCAmelCase__ : Tuple = 36.4519 lowerCAmelCase__ : Optional[Any] = 0.903421 lowerCAmelCase__ : Tuple = 222.088 lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = True lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : str = 0.763141 lowerCAmelCase__ : Union[str, Any] = TapasForQuestionAnswering(config=__A ) elif task == "TABFACT": lowerCAmelCase__ : Any = TapasForSequenceClassification(config=__A ) elif task == "MLM": lowerCAmelCase__ : Optional[Any] = TapasForMaskedLM(config=__A ) elif task == "INTERMEDIATE_PRETRAINING": lowerCAmelCase__ : List[str] = TapasModel(config=__A ) else: raise ValueError(F'''Task {task} not supported.''' ) print(F'''Building PyTorch model from configuration: {config}''' ) # Load weights from tf checkpoint load_tf_weights_in_tapas(__A , __A , __A ) # Save pytorch-model (weights and configuration) print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__A ) # Save tokenizer files print(F'''Save tokenizer files to {pytorch_dump_path}''' ) lowerCAmelCase__ : int = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' , model_max_length=512 ) tokenizer.save_pretrained(__A ) print('Used relative position embeddings:' , model.config.reset_position_index_per_cell ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""SQA""", type=str, help="""Model task for which to convert a checkpoint. Defaults to SQA.""" ) parser.add_argument( """--reset_position_index_per_cell""", default=False, action="""store_true""", help="""Whether to use relative position embeddings or not. Defaults to True.""", ) parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--tapas_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained TAPAS model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowerCamelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
708
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
0
import inspect import os import sys import unittest import accelerate from accelerate.test_utils import execute_subprocess_async, require_tpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : str = inspect.getfile(accelerate.test_utils ) lowerCAmelCase__ : Optional[Any] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['scripts', 'test_script.py'] ) lowerCAmelCase__ : int = os.path.sep.join(inspect.getfile(self.__class__ ).split(os.path.sep )[:-1] ) @require_tpu def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = f''' {self.test_dir}/xla_spawn.py --num_cores 8 {self.test_file_path} '''.split() lowerCAmelCase__ : Dict = [sys.executable] + distributed_args execute_subprocess_async(_lowerCAmelCase , env=os.environ.copy() )
709
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """google/rembert""": 256, } lowerCamelCase__ = """▁""" class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = RemBertTokenizer def __init__( self : Optional[Any] , a : str=None , a : Any=None , a : List[Any]=True , a : str=True , a : Dict=False , a : Dict="[CLS]" , a : int="[SEP]" , a : Tuple="<unk>" , a : Optional[Any]="[SEP]" , a : Tuple="<pad>" , a : Dict="[CLS]" , a : Optional[Any]="[MASK]" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( a , tokenizer_file=a , do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : int = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : List[Any] = keep_accents lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
69
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = "upernet" def __init__( self : Any , a : List[str]=None , a : Any=512 , a : Dict=0.0_2 , a : Any=[1, 2, 3, 6] , a : Optional[Any]=True , a : List[Any]=0.4 , a : Any=384 , a : List[str]=256 , a : Any=1 , a : int=False , a : Tuple=255 , **a : Optional[Any] , ): '''simple docstring''' super().__init__(**UpperCamelCase_ ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) lowerCAmelCase__ : int = CONFIG_MAPPING["resnet"](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(UpperCamelCase_ , UpperCamelCase_ ): lowerCAmelCase__ : Any = backbone_config.get('model_type' ) lowerCAmelCase__ : str = CONFIG_MAPPING[backbone_model_type] lowerCAmelCase__ : Union[str, Any] = config_class.from_dict(UpperCamelCase_ ) lowerCAmelCase__ : List[Any] = backbone_config lowerCAmelCase__ : Dict = hidden_size lowerCAmelCase__ : Optional[Any] = initializer_range lowerCAmelCase__ : str = pool_scales lowerCAmelCase__ : Tuple = use_auxiliary_head lowerCAmelCase__ : int = auxiliary_loss_weight lowerCAmelCase__ : List[Any] = auxiliary_in_channels lowerCAmelCase__ : Tuple = auxiliary_channels lowerCAmelCase__ : Any = auxiliary_num_convs lowerCAmelCase__ : Tuple = auxiliary_concat_input lowerCAmelCase__ : List[Any] = loss_ignore_index def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = copy.deepcopy(self.__dict__ ) lowerCAmelCase__ : Optional[Any] = self.backbone_config.to_dict() lowerCAmelCase__ : Any = self.__class__.model_type return output
710
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : int = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : List[Any] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : int = None lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Dict = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Union[str, Any] = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
69
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class A__ ( __magic_name__ ): lowercase = 'yolos' def __init__( self : str , a : Optional[Any]=768 , a : int=12 , a : Any=12 , a : Union[str, Any]=3_072 , a : List[Any]="gelu" , a : Dict=0.0 , a : str=0.0 , a : str=0.0_2 , a : int=1E-12 , a : Union[str, Any]=[512, 864] , a : Any=16 , a : str=3 , a : Optional[int]=True , a : str=100 , a : Tuple=True , a : Optional[int]=False , a : str=1 , a : Union[str, Any]=5 , a : str=2 , a : Dict=5 , a : Optional[Any]=2 , a : List[Any]=0.1 , **a : Tuple , ): '''simple docstring''' super().__init__(**__A ) lowerCAmelCase__ : List[Any] = hidden_size lowerCAmelCase__ : List[str] = num_hidden_layers lowerCAmelCase__ : Any = num_attention_heads lowerCAmelCase__ : str = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase__ : Tuple = attention_probs_dropout_prob lowerCAmelCase__ : Any = initializer_range lowerCAmelCase__ : Optional[Any] = layer_norm_eps lowerCAmelCase__ : Dict = image_size lowerCAmelCase__ : Dict = patch_size lowerCAmelCase__ : Dict = num_channels lowerCAmelCase__ : List[str] = qkv_bias lowerCAmelCase__ : Optional[int] = num_detection_tokens lowerCAmelCase__ : Tuple = use_mid_position_embeddings lowerCAmelCase__ : List[str] = auxiliary_loss # Hungarian matcher lowerCAmelCase__ : List[Any] = class_cost lowerCAmelCase__ : List[str] = bbox_cost lowerCAmelCase__ : str = giou_cost # Loss coefficients lowerCAmelCase__ : Optional[int] = bbox_loss_coefficient lowerCAmelCase__ : List[str] = giou_loss_coefficient lowerCAmelCase__ : Tuple = eos_coefficient class A__ ( __magic_name__ ): lowercase = version.parse('1.11' ) @property def _lowerCamelCase ( self : str ): '''simple docstring''' return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def _lowerCamelCase ( self : str ): '''simple docstring''' return 1E-4 @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' return 12
711
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
0
from typing import TYPE_CHECKING from ...utils import _LazyModule lowerCamelCase__ = {"processing_wav2vec2_with_lm": ["Wav2Vec2ProcessorWithLM"]} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
712
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> float: if not nums: # Makes sure that the list is not empty raise ValueError('List is empty' ) lowerCAmelCase__ : int = sum(__UpperCAmelCase ) / len(__UpperCAmelCase ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(__UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
713
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: if not all(x.isalpha() for x in string ): raise ValueError('String must only contain alphabetic characters.' ) lowerCAmelCase__ : str = sorted(string.lower() ) return len(lowerCAmelCase__ ) == len(set(lowerCAmelCase__ ) ) if __name__ == "__main__": lowerCamelCase__ = input("""Enter a string """).strip() lowerCamelCase__ = is_isogram(input_str) print(F"""{input_str} is {"an" if isogram else "not an"} isogram.""")
714
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
0
import qiskit def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 2 ) -> Optional[int]: lowerCAmelCase__ : str = qubits # Using Aer's simulator lowerCAmelCase__ : Optional[Any] = qiskit.Aer.get_backend('aer_simulator' ) # Creating a Quantum Circuit acting on the q register lowerCAmelCase__ : Dict = qiskit.QuantumCircuit(__A , __A ) # Adding a H gate on qubit 0 (now q0 in superposition) circuit.h(0 ) for i in range(1 , __A ): # Adding CX (CNOT) gate circuit.cx(i - 1 , __A ) # Mapping the quantum measurement to the classical bits circuit.measure(list(range(__A ) ) , list(range(__A ) ) ) # Now measuring any one qubit would affect other qubits to collapse # their super position and have same state as the measured one. # Executing the circuit on the simulator lowerCAmelCase__ : Dict = qiskit.execute(__A , __A , shots=1_000 ) return job.result().get_counts(__A ) if __name__ == "__main__": print(F"""Total count for various states are: {quantum_entanglement(3)}""")
715
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
0
'''simple docstring''' import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input lowerCamelCase__ = """Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine""" def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : int = _ask_options( 'In which compute environment are you running?' , ['This machine', 'AWS (Amazon SageMaker)'] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: lowerCAmelCase__ : Any = get_sagemaker_input() else: lowerCAmelCase__ : Optional[Any] = get_cluster_input() return config def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_=None ) -> Tuple: if subparsers is not None: lowerCAmelCase__ : List[Any] = subparsers.add_parser('config' , description=_lowerCamelCase ) else: lowerCAmelCase__ : Union[str, Any] = argparse.ArgumentParser('Accelerate config command' , description=_lowerCamelCase ) parser.add_argument( '--config_file' , default=_lowerCamelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=_lowerCamelCase ) return parser def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : int = get_user_input() if args.config_file is not None: lowerCAmelCase__ : str = args.config_file else: if not os.path.isdir(_lowerCamelCase ): os.makedirs(_lowerCamelCase ) lowerCAmelCase__ : int = default_yaml_config_file if config_file.endswith('.json' ): config.to_json_file(_lowerCamelCase ) else: config.to_yaml_file(_lowerCamelCase ) print(F'''accelerate configuration saved at {config_file}''' ) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Optional[int] = config_command_parser() lowerCAmelCase__ : Dict = parser.parse_args() config_command(_lowerCamelCase ) if __name__ == "__main__": main()
716
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
0
import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class A__ : def __init__( self : Tuple , a : int , a : Tuple=13 , a : List[str]=30 , a : Tuple=2 , a : Any=3 , a : Dict=True , a : Optional[int]=True , a : int=32 , a : List[str]=5 , a : Any=4 , a : Optional[int]=37 , a : List[Any]="gelu" , a : int=0.1 , a : Union[str, Any]=0.1 , a : Optional[Any]=10 , a : Dict=0.0_2 , a : Tuple=3 , a : Union[str, Any]=None , a : List[Any]=2 , ): '''simple docstring''' lowerCAmelCase__ : List[Any] = parent lowerCAmelCase__ : Optional[int] = batch_size lowerCAmelCase__ : Any = image_size lowerCAmelCase__ : Optional[Any] = patch_size lowerCAmelCase__ : Optional[Any] = num_channels lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : Optional[Any] = use_labels lowerCAmelCase__ : Union[str, Any] = hidden_size lowerCAmelCase__ : Tuple = num_hidden_layers lowerCAmelCase__ : Union[str, Any] = num_attention_heads lowerCAmelCase__ : Union[str, Any] = intermediate_size lowerCAmelCase__ : Any = hidden_act lowerCAmelCase__ : Tuple = hidden_dropout_prob lowerCAmelCase__ : Dict = attention_probs_dropout_prob lowerCAmelCase__ : Any = type_sequence_label_size lowerCAmelCase__ : Tuple = initializer_range lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Optional[int] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) lowerCAmelCase__ : List[str] = (image_size // patch_size) ** 2 lowerCAmelCase__ : List[str] = num_patches + 2 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : List[Any] = None if self.use_labels: lowerCAmelCase__ : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Dict = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : str ): '''simple docstring''' return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : int , a : Optional[Any] , a : int , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = DeiTModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() lowerCAmelCase__ : List[Any] = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : List[Any] , a : Optional[Any] , a : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = DeiTForMaskedImageModeling(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() lowerCAmelCase__ : Optional[int] = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : Optional[int] = DeiTForMaskedImageModeling(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() lowerCAmelCase__ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[int] = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Any , a : str , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : str = self.type_sequence_label_size lowerCAmelCase__ : List[str] = DeiTForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() lowerCAmelCase__ : Tuple = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : str = DeiTForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() lowerCAmelCase__ : Tuple = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : Tuple = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ( lowerCAmelCase__ ) : Tuple = config_and_inputs lowerCAmelCase__ : int = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A__ ( _A , _A , unittest.TestCase ): lowercase = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowercase = ( { 'feature-extraction': DeiTModel, 'image-classification': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : str = DeiTModelTester(self ) lowerCAmelCase__ : Optional[int] = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=37 ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='DeiT does not use inputs_embeds' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' pass def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Dict = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , nn.Linear ) ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Union[str, Any] = model_class(__lowerCamelCase ) lowerCAmelCase__ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Any = [*signature.parameters.keys()] lowerCAmelCase__ : Any = ["pixel_values"] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowerCamelCase ( self : Tuple , a : int , a : Optional[Any] , a : List[Any]=False ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' if not self.model_tester.is_training: return lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Dict = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(__lowerCamelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue lowerCAmelCase__ : Union[str, Any] = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.train() lowerCAmelCase__ : Union[str, Any] = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) lowerCAmelCase__ : Dict = model(**__lowerCamelCase ).loss loss.backward() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return lowerCAmelCase__ : Tuple = False lowerCAmelCase__ : Any = True for model_class in self.all_model_classes: if model_class in get_values(__lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue lowerCAmelCase__ : List[str] = model_class(__lowerCamelCase ) model.gradient_checkpointing_enable() model.to(__lowerCamelCase ) model.train() lowerCAmelCase__ : Dict = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) lowerCAmelCase__ : Tuple = model(**__lowerCamelCase ).loss loss.backward() def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : int = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(__lowerCamelCase ), *get_values(__lowerCamelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f'''Testing {model_class} with {problem_type["title"]}''' ): lowerCAmelCase__ : Tuple = problem_type["title"] lowerCAmelCase__ : Optional[Any] = problem_type["num_labels"] lowerCAmelCase__ : List[str] = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.train() lowerCAmelCase__ : List[Any] = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if problem_type["num_labels"] > 1: lowerCAmelCase__ : List[str] = inputs["labels"].unsqueeze(1 ).repeat(1 , problem_type['num_labels'] ) lowerCAmelCase__ : int = inputs["labels"].to(problem_type['dtype'] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=__lowerCamelCase ) as warning_list: lowerCAmelCase__ : Optional[Any] = model(**__lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[Any] = DeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def lowerCAmelCase__ ( ) -> Tuple: lowerCAmelCase__ : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return ( DeiTImageProcessor.from_pretrained('facebook/deit-base-distilled-patch16-224' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = DeiTForImageClassificationWithTeacher.from_pretrained('facebook/deit-base-distilled-patch16-224' ).to( __lowerCamelCase ) lowerCAmelCase__ : List[Any] = self.default_image_processor lowerCAmelCase__ : List[Any] = prepare_img() lowerCAmelCase__ : Optional[Any] = image_processor(images=__lowerCamelCase , return_tensors='pt' ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): lowerCAmelCase__ : List[str] = model(**__lowerCamelCase ) # verify the logits lowerCAmelCase__ : str = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) lowerCAmelCase__ : List[str] = torch.tensor([-1.0_2_6_6, 0.1_9_1_2, -1.2_8_6_1] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = DeiTModel.from_pretrained( 'facebook/deit-base-distilled-patch16-224' , torch_dtype=torch.floataa , device_map='auto' ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : Optional[int] = prepare_img() lowerCAmelCase__ : Union[str, Any] = image_processor(images=__lowerCamelCase , return_tensors='pt' ) lowerCAmelCase__ : Union[str, Any] = inputs.pixel_values.to(__lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): lowerCAmelCase__ : List[str] = model(__lowerCamelCase )
717
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class A__ : def __init__( self : Dict , a : Optional[int] , a : Tuple=13 , a : List[str]=30 , a : Union[str, Any]=2 , a : Optional[int]=3 , a : Dict=True , a : Optional[int]=True , a : Tuple=32 , a : List[Any]=5 , a : Any=4 , a : Tuple=37 , a : List[str]="gelu" , a : Dict=0.1 , a : str=0.1 , a : Optional[Any]=10 , a : Union[str, Any]=0.0_2 , a : Tuple=3 , a : str=None , a : Any=2 , ): '''simple docstring''' lowerCAmelCase__ : Any = parent lowerCAmelCase__ : Optional[Any] = batch_size lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : int = patch_size lowerCAmelCase__ : Union[str, Any] = num_channels lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : str = use_labels lowerCAmelCase__ : List[str] = hidden_size lowerCAmelCase__ : Union[str, Any] = num_hidden_layers lowerCAmelCase__ : Optional[Any] = num_attention_heads lowerCAmelCase__ : Any = intermediate_size lowerCAmelCase__ : Union[str, Any] = hidden_act lowerCAmelCase__ : Tuple = hidden_dropout_prob lowerCAmelCase__ : List[Any] = attention_probs_dropout_prob lowerCAmelCase__ : str = type_sequence_label_size lowerCAmelCase__ : List[Any] = initializer_range lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Any = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) lowerCAmelCase__ : int = (image_size // patch_size) ** 2 lowerCAmelCase__ : Optional[Any] = num_patches + 2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : List[Any] = None if self.use_labels: lowerCAmelCase__ : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase_ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : Optional[Any] , a : Dict , a : Optional[Any] , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = DeiTModel(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() lowerCAmelCase__ : Optional[Any] = model(lowerCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : Tuple , a : Dict , a : List[Any] , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : str = DeiTForMaskedImageModeling(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(lowerCamelCase_ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : str = 1 lowerCAmelCase__ : List[Any] = DeiTForMaskedImageModeling(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() lowerCAmelCase__ : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = model(lowerCamelCase_ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Any , a : List[Any] , a : Tuple , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.type_sequence_label_size lowerCAmelCase__ : Union[str, Any] = DeiTForImageClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() lowerCAmelCase__ : int = model(lowerCamelCase_ , labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCAmelCase__ : List[Any] = 1 lowerCAmelCase__ : Tuple = DeiTForImageClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() lowerCAmelCase__ : Tuple = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = model(lowerCamelCase_ , labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.prepare_config_and_inputs() ( lowerCAmelCase__ ) : List[str] = config_and_inputs lowerCAmelCase__ : Any = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class A__ ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): lowercase = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowercase = ( { 'feature-extraction': DeiTModel, 'image-classification': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = DeiTModelTester(self ) lowerCAmelCase__ : Union[str, Any] = ConfigTester(self , config_class=lowerCamelCase_ , has_text_modality=lowerCamelCase_ , hidden_size=37 ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='DeiT does not use inputs_embeds' ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' pass def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Dict = model_class(lowerCamelCase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase_ , nn.Linear ) ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = model_class(lowerCamelCase_ ) lowerCAmelCase__ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : List[Any] = [*signature.parameters.keys()] lowerCAmelCase__ : Tuple = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase_ ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase_ ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase_ ) def _lowerCamelCase ( self : int , a : List[str] , a : Optional[Any] , a : Optional[int]=False ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = super()._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ , return_labels=lowerCamelCase_ ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' if not self.model_tester.is_training: return lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : str = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowerCamelCase_ ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue lowerCAmelCase__ : int = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.train() lowerCAmelCase__ : Tuple = self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ , return_labels=lowerCamelCase_ ) lowerCAmelCase__ : Any = model(**lowerCamelCase_ ).loss loss.backward() def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return lowerCAmelCase__ : Union[str, Any] = False lowerCAmelCase__ : Optional[Any] = True for model_class in self.all_model_classes: if model_class in get_values(lowerCamelCase_ ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue lowerCAmelCase__ : Dict = model_class(lowerCamelCase_ ) model.gradient_checkpointing_enable() model.to(lowerCamelCase_ ) model.train() lowerCAmelCase__ : Optional[int] = self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ , return_labels=lowerCamelCase_ ) lowerCAmelCase__ : Tuple = model(**lowerCamelCase_ ).loss loss.backward() def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Tuple = [ {'''title''': '''multi_label_classification''', '''num_labels''': 2, '''dtype''': torch.float}, {'''title''': '''single_label_classification''', '''num_labels''': 1, '''dtype''': torch.long}, {'''title''': '''regression''', '''num_labels''': 1, '''dtype''': torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowerCamelCase_ ), *get_values(lowerCamelCase_ ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f'''Testing {model_class} with {problem_type["title"]}''' ): lowerCAmelCase__ : str = problem_type['''title'''] lowerCAmelCase__ : Dict = problem_type['''num_labels'''] lowerCAmelCase__ : List[Any] = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.train() lowerCAmelCase__ : str = self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ , return_labels=lowerCamelCase_ ) if problem_type["num_labels"] > 1: lowerCAmelCase__ : Any = inputs['''labels'''].unsqueeze(1 ).repeat(1 , problem_type['num_labels'] ) lowerCAmelCase__ : Union[str, Any] = inputs['''labels'''].to(problem_type['dtype'] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowerCamelCase_ ) as warning_list: lowerCAmelCase__ : List[str] = model(**lowerCamelCase_ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[str] = DeiTModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) def lowerCAmelCase__ ( ) -> List[Any]: lowerCAmelCase__ : str = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : int ): '''simple docstring''' return ( DeiTImageProcessor.from_pretrained('facebook/deit-base-distilled-patch16-224' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = DeiTForImageClassificationWithTeacher.from_pretrained('facebook/deit-base-distilled-patch16-224' ).to( lowerCamelCase_ ) lowerCAmelCase__ : Optional[Any] = self.default_image_processor lowerCAmelCase__ : List[str] = prepare_img() lowerCAmelCase__ : str = image_processor(images=lowerCamelCase_ , return_tensors='pt' ).to(lowerCamelCase_ ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Tuple = model(**lowerCamelCase_ ) # verify the logits lowerCAmelCase__ : str = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase_ ) lowerCAmelCase__ : Tuple = torch.tensor([-1.0_2_6_6, 0.1_9_1_2, -1.2_8_6_1] ).to(lowerCamelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase_ , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = DeiTModel.from_pretrained( 'facebook/deit-base-distilled-patch16-224' , torch_dtype=torch.floataa , device_map='auto' ) lowerCAmelCase__ : Optional[int] = self.default_image_processor lowerCAmelCase__ : List[str] = prepare_img() lowerCAmelCase__ : List[Any] = image_processor(images=lowerCamelCase_ , return_tensors='pt' ) lowerCAmelCase__ : Tuple = inputs.pixel_values.to(lowerCamelCase_ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(lowerCamelCase_ )
718
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : Optional[Any] = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(_A , _A ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Tuple = emb.weight.shape lowerCAmelCase__ : Any = nn.Linear(_A , _A , bias=_A ) lowerCAmelCase__ : List[str] = emb.weight.data return lin_layer def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Dict: lowerCAmelCase__ : Optional[Any] = torch.load(_A , map_location='cpu' ) lowerCAmelCase__ : str = mam_aaa['args'] or mam_aaa['cfg']['model'] lowerCAmelCase__ : Optional[Any] = mam_aaa['model'] remove_ignore_keys_(_A ) lowerCAmelCase__ : List[Any] = state_dict['encoder.embed_tokens.weight'].shape[0] lowerCAmelCase__ : Dict = MaMaaaConfig( vocab_size=_A , max_position_embeddings=1_024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='relu' , ) lowerCAmelCase__ : Dict = state_dict['decoder.embed_tokens.weight'] lowerCAmelCase__ : Dict = MaMaaaForConditionalGeneration(_A ) model.model.load_state_dict(_A , strict=_A ) lowerCAmelCase__ : List[Any] = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") lowerCamelCase__ = parser.parse_args() lowerCamelCase__ = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
719
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
720
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
0