code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
711
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
0
import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class A__ : def __init__( self : Any , a : Tuple , a : Optional[int]=13 , a : Any=7 , a : str=6 , a : Optional[Any]=17 , a : List[Any]=23 , a : Any=11 , a : str=True , ): '''simple docstring''' lowerCAmelCase__ : Dict = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : Any = seq_length lowerCAmelCase__ : List[Any] = act_dim lowerCAmelCase__ : Dict = state_dim lowerCAmelCase__ : Union[str, Any] = hidden_size lowerCAmelCase__ : Any = max_length lowerCAmelCase__ : str = is_training def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowerCAmelCase__ : Any = floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowerCAmelCase__ : List[str] = floats_tensor((self.batch_size, self.seq_length, 1) ) lowerCAmelCase__ : List[Any] = floats_tensor((self.batch_size, self.seq_length, 1) ) lowerCAmelCase__ : Optional[int] = ids_tensor((self.batch_size, self.seq_length) , vocab_size=1_000 ) lowerCAmelCase__ : Union[str, Any] = random_attention_mask((self.batch_size, self.seq_length) ) lowerCAmelCase__ : Dict = self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _lowerCamelCase ( self : List[Any] , a : Tuple , a : str , a : Optional[Any] , a : Any , a : List[Any] , a : Any , a : Optional[int] , ): '''simple docstring''' lowerCAmelCase__ : Dict = DecisionTransformerModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , a , a , a , a , a ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.prepare_config_and_inputs() ( lowerCAmelCase__ ) : Any = config_and_inputs lowerCAmelCase__ : Any = { 'states': states, 'actions': actions, 'rewards': rewards, 'returns_to_go': returns_to_go, 'timesteps': timesteps, 'attention_mask': attention_mask, } return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = (DecisionTransformerModel,) if is_torch_available() else () lowercase = () lowercase = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids lowercase = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features lowercase = False lowercase = False lowercase = False lowercase = False lowercase = False lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = DecisionTransformerModelTester(self ) lowerCAmelCase__ : Optional[int] = ConfigTester(self , config_class=a , hidden_size=37 ) def _lowerCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) @slow def _lowerCamelCase ( self : int ): '''simple docstring''' for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : str = DecisionTransformerModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : str = model_class(a ) lowerCAmelCase__ : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Dict = [*signature.parameters.keys()] lowerCAmelCase__ : Any = [ 'states', 'actions', 'rewards', 'returns_to_go', 'timesteps', 'attention_mask', ] self.assertListEqual(arg_names[: len(a )] , a ) @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = 2 # number of steps of autoregressive prediction we will perform lowerCAmelCase__ : List[str] = 10 # defined by the RL environment, may be normalized lowerCAmelCase__ : Union[str, Any] = DecisionTransformerModel.from_pretrained('edbeeching/decision-transformer-gym-hopper-expert' ) lowerCAmelCase__ : str = model.to(a ) lowerCAmelCase__ : Union[str, Any] = model.config torch.manual_seed(0 ) lowerCAmelCase__ : Dict = torch.randn(1 , 1 , config.state_dim ).to(device=a , dtype=torch.floataa ) # env.reset() lowerCAmelCase__ : int = torch.tensor( [[0.2_4_2_7_9_3, -0.2_8_6_9_3_0_7_4, 0.8_7_4_2_6_1_3], [0.6_7_8_1_5_2_7_4, -0.0_8_1_0_1_0_8_5, -0.1_2_9_5_2_1_4_7]] , device=a ) lowerCAmelCase__ : Any = torch.tensor(a , device=a , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowerCAmelCase__ : List[Any] = state lowerCAmelCase__ : Tuple = torch.zeros(1 , 0 , config.act_dim , device=a , dtype=torch.floataa ) lowerCAmelCase__ : Optional[int] = torch.zeros(1 , 0 , device=a , dtype=torch.floataa ) lowerCAmelCase__ : List[str] = torch.tensor(0 , device=a , dtype=torch.long ).reshape(1 , 1 ) for step in range(a ): lowerCAmelCase__ : Optional[int] = torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=a )] , dim=1 ) lowerCAmelCase__ : Optional[Any] = torch.cat([rewards, torch.zeros(1 , 1 , device=a )] , dim=1 ) lowerCAmelCase__ : int = torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowerCAmelCase__ : Tuple = model( states=a , actions=a , rewards=a , returns_to_go=a , timesteps=a , attention_mask=a , return_dict=a , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowerCAmelCase__ : Union[str, Any] = ( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=a , dtype=torch.floataa ), 1.0, False, {}, ) lowerCAmelCase__ : Optional[Any] = action_pred[0, -1] lowerCAmelCase__ : str = torch.cat([states, state] , dim=1 ) lowerCAmelCase__ : Optional[Any] = returns_to_go[0, -1] - reward lowerCAmelCase__ : List[str] = torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowerCAmelCase__ : Tuple = torch.cat( [timesteps, torch.ones((1, 1) , device=a , dtype=torch.long ) * (step + 1)] , dim=1 )
712
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
0
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class A__ : def __init__( self : List[Any] , a : Optional[int] , a : List[str]=12 , a : List[Any]=7 , a : Dict=True , a : str=True , a : Union[str, Any]=True , a : Optional[int]=99 , a : str=32 , a : List[Any]=32 , a : Optional[Any]=2 , a : Any=4 , a : List[str]=37 , a : List[Any]=0.1 , a : Any=0.1 , a : str=512 , a : Optional[int]=0.0_2 , a : Tuple=0 , a : Dict=None , ): '''simple docstring''' lowerCAmelCase__ : List[Any] = parent lowerCAmelCase__ : Tuple = batch_size lowerCAmelCase__ : List[str] = seq_length lowerCAmelCase__ : str = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_labels lowerCAmelCase__ : Union[str, Any] = vocab_size lowerCAmelCase__ : int = hidden_size lowerCAmelCase__ : Optional[Any] = projection_dim lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : Dict = num_attention_heads lowerCAmelCase__ : Union[str, Any] = intermediate_size lowerCAmelCase__ : Optional[Any] = dropout lowerCAmelCase__ : Tuple = attention_dropout lowerCAmelCase__ : List[str] = max_position_embeddings lowerCAmelCase__ : Union[str, Any] = initializer_range lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : List[Any] = bos_token_id def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Optional[int] = None if self.use_input_mask: lowerCAmelCase__ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowerCAmelCase__ : Any = input_mask.numpy() lowerCAmelCase__ : Optional[Any] = input_mask.shape lowerCAmelCase__ : Union[str, Any] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a ): lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[Any] = 0 lowerCAmelCase__ : str = self.get_config() return config, input_ids, tf.convert_to_tensor(a ) def _lowerCamelCase ( self : str ): '''simple docstring''' return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _lowerCamelCase ( self : List[str] , a : str , a : Dict , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = TFBlipTextModel(config=a ) lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , training=a ) lowerCAmelCase__ : str = model(a , training=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ : int = config_and_inputs lowerCAmelCase__ : str = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class A__ ( __magic_name__ , unittest.TestCase ): lowercase = (TFBlipTextModel,) if is_tf_available() else () lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Dict = BlipTextModelTester(self ) lowerCAmelCase__ : Tuple = ConfigTester(self , config_class=a , hidden_size=37 ) def _lowerCamelCase ( self : str ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' pass def _lowerCamelCase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Union[str, Any] = TFBlipTextModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Union[str, Any] , a : Any=True ): '''simple docstring''' super().test_pt_tf_model_equivalence(allow_missing_keys=a )
713
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
0
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) lowerCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={'help': 'Model type selected in the list: ' + ', '.join(__magic_name__ )} ) lowercase = field( default=__magic_name__ , metadata={'help': 'The input data dir. Should contain the .json files for the SQuAD task.'} ) lowercase = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) lowercase = field( default=128 , metadata={'help': 'When splitting up a long document into chunks, how much stride to take between chunks.'} , ) lowercase = field( default=64 , metadata={ 'help': ( 'The maximum number of tokens for the question. Questions longer than this will ' 'be truncated to this length.' ) } , ) lowercase = field( default=30 , metadata={ 'help': ( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ) } , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) lowercase = field( default=__magic_name__ , metadata={'help': 'If true, the SQuAD examples contain some that do not have an answer.'} ) lowercase = field( default=0.0 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) lowercase = field( default=20 , metadata={'help': 'If null_score - best_non_null is greater than the threshold predict null.'} ) lowercase = field( default=0 , metadata={ 'help': ( 'language id of input for language-specific xlm models (see' ' tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)' ) } , ) lowercase = field(default=1 , metadata={'help': 'multiple threads for converting example to features'} ) class A__ ( __magic_name__ ): lowercase = 'train' lowercase = 'dev' class A__ ( __magic_name__ ): lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 def __init__( self : Tuple , a : SquadDataTrainingArguments , a : PreTrainedTokenizer , a : Optional[int] = None , a : Union[str, Split] = Split.train , a : Optional[bool] = False , a : Optional[str] = None , a : Optional[str] = "pt" , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = args lowerCAmelCase__ : str = is_language_sensitive lowerCAmelCase__ : Optional[Any] = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(a , a ): try: lowerCAmelCase__ : List[str] = Split[mode] except KeyError: raise KeyError('mode is not a valid split name' ) lowerCAmelCase__ : Union[str, Any] = mode # Load data features from cache or dataset file lowerCAmelCase__ : Tuple = 'v2' if args.version_2_with_negative else 'v1' lowerCAmelCase__ : Union[str, Any] = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}''' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCAmelCase__ : Any = cached_features_file + '.lock' with FileLock(a ): if os.path.exists(a ) and not args.overwrite_cache: lowerCAmelCase__ : Any = time.time() lowerCAmelCase__ : Optional[Any] = torch.load(a ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. lowerCAmelCase__ : Optional[Any] = self.old_features['features'] lowerCAmelCase__ : Union[str, Any] = self.old_features.get('dataset' , a ) lowerCAmelCase__ : Optional[Any] = self.old_features.get('examples' , a ) logger.info( f'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f'''Deleting cached file {cached_features_file} will allow dataset and examples to be cached in''' ' future run' ) else: if mode == Split.dev: lowerCAmelCase__ : List[str] = self.processor.get_dev_examples(args.data_dir ) else: lowerCAmelCase__ : str = self.processor.get_train_examples(args.data_dir ) lowerCAmelCase__ : Union[str, Any] = squad_convert_examples_to_features( examples=self.examples , tokenizer=a , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=a , ) lowerCAmelCase__ : Optional[int] = time.time() torch.save( {'features': self.features, 'dataset': self.dataset, 'examples': self.examples} , a , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self : Union[str, Any] ): '''simple docstring''' return len(self.features ) def __getitem__( self : str , a : int ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.features[i] lowerCAmelCase__ : Any = torch.tensor(feature.input_ids , dtype=torch.long ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(feature.attention_mask , dtype=torch.long ) lowerCAmelCase__ : str = torch.tensor(feature.token_type_ids , dtype=torch.long ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(feature.cls_index , dtype=torch.long ) lowerCAmelCase__ : int = torch.tensor(feature.p_mask , dtype=torch.float ) lowerCAmelCase__ : Dict = torch.tensor(feature.is_impossible , dtype=torch.float ) lowerCAmelCase__ : Dict = { 'input_ids': input_ids, 'attention_mask': attention_mask, 'token_type_ids': token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({'cls_index': cls_index, 'p_mask': p_mask} ) if self.args.version_2_with_negative: inputs.update({'is_impossible': is_impossible} ) if self.is_language_sensitive: inputs.update({'langs': (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: lowerCAmelCase__ : Any = torch.tensor(feature.start_position , dtype=torch.long ) lowerCAmelCase__ : str = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({'start_positions': start_positions, 'end_positions': end_positions} ) return inputs
714
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
0
import mpmath # for roots of unity import numpy as np class A__ : def __init__( self : Optional[Any] , a : Tuple=None , a : int=None ): '''simple docstring''' lowerCAmelCase__ : Tuple = list(poly_a or [0] )[:] lowerCAmelCase__ : List[Any] = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() lowerCAmelCase__ : Any = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() lowerCAmelCase__ : Tuple = len(self.polyB ) # Add 0 to make lengths equal a power of 2 lowerCAmelCase__ : Tuple = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform lowerCAmelCase__ : List[Any] = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product lowerCAmelCase__ : Optional[int] = self.__multiply() def _lowerCamelCase ( self : Tuple , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Dict = [[x] for x in self.polyA] if which == 'A' else [[x] for x in self.polyB] # Corner case if len(a ) <= 1: return dft[0] # lowerCAmelCase__ : Dict = self.c_max_length // 2 while next_ncol > 0: lowerCAmelCase__ : Optional[int] = [[] for i in range(a )] lowerCAmelCase__ : int = self.root**next_ncol # First half of next step lowerCAmelCase__ : Tuple = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step lowerCAmelCase__ : Optional[Any] = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update lowerCAmelCase__ : List[str] = new_dft lowerCAmelCase__ : Optional[int] = next_ncol // 2 return dft[0] def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.__dft('A' ) lowerCAmelCase__ : str = self.__dft('B' ) lowerCAmelCase__ : Tuple = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT lowerCAmelCase__ : List[str] = 2 while next_ncol <= self.c_max_length: lowerCAmelCase__ : Optional[int] = [[] for i in range(a )] lowerCAmelCase__ : Union[str, Any] = self.root ** (next_ncol // 2) lowerCAmelCase__ : List[Any] = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update lowerCAmelCase__ : Optional[int] = new_inverse_c next_ncol *= 2 # Unpack lowerCAmelCase__ : Optional[int] = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1J for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = 'A = ' + ' + '.join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) lowerCAmelCase__ : Optional[int] = 'B = ' + ' + '.join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) lowerCAmelCase__ : Any = 'A*B = ' + ' + '.join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return f'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
715
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
0
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
716
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) lowerCAmelCase__ : List[Any] = get_activation('gelu' ) self.assertTrue(torch.allclose(gelu_python(a ) , torch_builtin(a ) ) ) self.assertFalse(torch.allclose(gelu_python(a ) , gelu_new(a ) ) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) lowerCAmelCase__ : Dict = get_activation('gelu' ) lowerCAmelCase__ : Tuple = get_activation('gelu_10' ) lowerCAmelCase__ : Dict = torch_builtin(a ) lowerCAmelCase__ : List[str] = geluaa(a ) lowerCAmelCase__ : List[Any] = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(a ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' get_activation('gelu' ) get_activation('gelu_10' ) get_activation('gelu_fast' ) get_activation('gelu_new' ) get_activation('gelu_python' ) get_activation('gelu_pytorch_tanh' ) get_activation('linear' ) get_activation('mish' ) get_activation('quick_gelu' ) get_activation('relu' ) get_activation('sigmoid' ) get_activation('silu' ) get_activation('swish' ) get_activation('tanh' ) with self.assertRaises(a ): get_activation('bogus' ) with self.assertRaises(a ): get_activation(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = get_activation('gelu' ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Any = get_activation('gelu' ) self.assertEqual(acta.a , 1 ) with self.assertRaises(a ): lowerCAmelCase__ : Optional[int] = acta.a
717
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_git""": ["""GIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GitConfig""", """GitVisionConfig"""], """processing_git""": ["""GitProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """GIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """GitForCausalLM""", """GitModel""", """GitPreTrainedModel""", """GitVisionModel""", ] if TYPE_CHECKING: from .configuration_git import GIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GitConfig, GitVisionConfig from .processing_git import GitProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_git import ( GIT_PRETRAINED_MODEL_ARCHIVE_LIST, GitForCausalLM, GitModel, GitPreTrainedModel, GitVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
718
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
from __future__ import annotations from typing import Any class A__ : def __init__( self : List[str] , a : int , a : int , a : float = 0 ): '''simple docstring''' lowerCAmelCase__ : Dict = row, column lowerCAmelCase__ : int = [[default_value for c in range(a )] for r in range(a )] def __str__( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = f'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier lowerCAmelCase__ : Tuple = 0 for row_vector in self.array: for obj in row_vector: lowerCAmelCase__ : Any = max(a , len(str(a ) ) ) lowerCAmelCase__ : int = f'''%{max_element_length}s''' # Make string and return def single_line(a : list[float] ) -> str: nonlocal string_format_identifier lowerCAmelCase__ : int = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(a ) for row_vector in self.array ) return s def __repr__( self : int ): '''simple docstring''' return str(self ) def _lowerCamelCase ( self : Optional[Any] , a : tuple[int, int] ): '''simple docstring''' if not (isinstance(a , (list, tuple) ) and len(a ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : List[str] , a : tuple[int, int] ): '''simple docstring''' assert self.validate_indicies(a ) return self.array[loc[0]][loc[1]] def __setitem__( self : Any , a : tuple[int, int] , a : float ): '''simple docstring''' assert self.validate_indicies(a ) lowerCAmelCase__ : Optional[Any] = value def __add__( self : Tuple , a : Matrix ): '''simple docstring''' assert isinstance(a , a ) assert self.row == another.row and self.column == another.column # Add lowerCAmelCase__ : List[Any] = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase__ : Tuple = self[r, c] + another[r, c] return result def __neg__( self : Dict ): '''simple docstring''' lowerCAmelCase__ : str = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase__ : List[Any] = -self[r, c] return result def __sub__( self : Dict , a : Matrix ): '''simple docstring''' return self + (-another) def __mul__( self : Tuple , a : int | float | Matrix ): '''simple docstring''' if isinstance(a , (int, float) ): # Scalar multiplication lowerCAmelCase__ : Any = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase__ : Optional[Any] = self[r, c] * another return result elif isinstance(a , a ): # Matrix multiplication assert self.column == another.row lowerCAmelCase__ : List[str] = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: lowerCAmelCase__ : Union[str, Any] = f'''Unsupported type given for another ({type(a )})''' raise TypeError(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase__ : Optional[int] = self[r, c] return result def _lowerCamelCase ( self : List[str] , a : Matrix , a : Matrix ): '''simple docstring''' assert isinstance(a , a ) and isinstance(a , a ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate lowerCAmelCase__ : Any = v.transpose() lowerCAmelCase__ : Any = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def lowerCAmelCase__ ( ) -> None: # a^(-1) lowerCAmelCase__ : Any = Matrix(3 , 3 , 0 ) for i in range(3 ): lowerCAmelCase__ : List[str] = 1 print(F'''a^(-1) is {ainv}''' ) # u, v lowerCAmelCase__ : Dict = Matrix(3 , 1 , 0 ) lowerCAmelCase__ : Any = 1, 2, -3 lowerCAmelCase__ : Any = Matrix(3 , 1 , 0 ) lowerCAmelCase__ : List[Any] = 4, -2, 5 print(F'''u is {u}''' ) print(F'''v is {v}''' ) print(F'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )}''' ) def lowerCAmelCase__ ( ) -> None: import doctest doctest.testmod() testa()
719
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) lowerCamelCase__ = logging.getLogger(__name__) @dataclass(frozen=__magic_name__ ) class A__ : lowercase = 42 lowercase = 42 lowercase = None lowercase = None lowercase = None @dataclass(frozen=__magic_name__ ) class A__ : lowercase = 42 lowercase = None lowercase = None lowercase = None lowercase = None if is_torch_available(): import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): lowercase = 42 def __init__( self : List[str] , a : str , a : PreTrainedTokenizer , a : str , a : Optional[int] = None , a : str=False , a : bool = False , ): '''simple docstring''' lowerCAmelCase__ : int = hans_processors[task]() lowerCAmelCase__ : Dict = os.path.join( a , 'cached_{}_{}_{}_{}'.format( 'dev' if evaluate else 'train' , tokenizer.__class__.__name__ , str(a ) , a , ) , ) lowerCAmelCase__ : Any = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase__ : int = label_list[2], label_list[1] lowerCAmelCase__ : int = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCAmelCase__ : List[str] = cached_features_file + '.lock' with FileLock(a ): if os.path.exists(a ) and not overwrite_cache: logger.info(f'''Loading features from cached file {cached_features_file}''' ) lowerCAmelCase__ : Optional[Any] = torch.load(a ) else: logger.info(f'''Creating features from dataset file at {data_dir}''' ) lowerCAmelCase__ : Tuple = ( processor.get_dev_examples(a ) if evaluate else processor.get_train_examples(a ) ) logger.info('Training examples: %s' , len(a ) ) lowerCAmelCase__ : Optional[Any] = hans_convert_examples_to_features(a , a , a , a ) logger.info('Saving features into cached file %s' , a ) torch.save(self.features , a ) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.features ) def __getitem__( self : Optional[Any] , a : int ): '''simple docstring''' return self.features[i] def _lowerCamelCase ( self : str ): '''simple docstring''' return self.label_list if is_tf_available(): import tensorflow as tf class A__ : lowercase = 42 def __init__( self : Any , a : str , a : PreTrainedTokenizer , a : str , a : Optional[int] = 128 , a : Union[str, Any]=False , a : bool = False , ): '''simple docstring''' lowerCAmelCase__ : List[str] = hans_processors[task]() lowerCAmelCase__ : List[str] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase__ : Dict = label_list[2], label_list[1] lowerCAmelCase__ : Any = label_list lowerCAmelCase__ : Optional[Any] = processor.get_dev_examples(a ) if evaluate else processor.get_train_examples(a ) lowerCAmelCase__ : Optional[Any] = hans_convert_examples_to_features(a , a , a , a ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='convert examples to features' ): if ex_index % 10_000 == 0: logger.info('Writing example %d of %d' % (ex_index, len(a )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) lowerCAmelCase__ : Optional[Any] = tf.data.Dataset.from_generator( a , ( { 'example_id': tf.intaa, 'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa, }, tf.intaa, ) , ( { 'example_id': tf.TensorShape([] ), 'input_ids': tf.TensorShape([None, None] ), 'attention_mask': tf.TensorShape([None, None] ), 'token_type_ids': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' return self.dataset def __len__( self : Tuple ): '''simple docstring''' return len(self.features ) def __getitem__( self : Dict , a : Union[str, Any] ): '''simple docstring''' return self.features[i] def _lowerCamelCase ( self : str ): '''simple docstring''' return self.label_list class A__ ( __magic_name__ ): def _lowerCamelCase ( self : str , a : Dict ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(a , 'heuristics_train_set.txt' ) ) , 'train' ) def _lowerCamelCase ( self : Dict , a : Optional[Any] ): '''simple docstring''' return self._create_examples(self._read_tsv(os.path.join(a , 'heuristics_evaluation_set.txt' ) ) , 'dev' ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return ["contradiction", "entailment", "neutral"] def _lowerCamelCase ( self : Dict , a : Dict , a : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = [] for i, line in enumerate(a ): if i == 0: continue lowerCAmelCase__ : Optional[int] = '%s-%s' % (set_type, line[0]) lowerCAmelCase__ : Tuple = line[5] lowerCAmelCase__ : Any = line[6] lowerCAmelCase__ : Union[str, Any] = line[7][2:] if line[7].startswith('ex' ) else line[7] lowerCAmelCase__ : List[str] = line[0] examples.append(InputExample(guid=a , text_a=a , text_b=a , label=a , pairID=a ) ) return examples def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[int]: lowerCAmelCase__ : Dict = {label: i for i, label in enumerate(SCREAMING_SNAKE_CASE_ )} lowerCAmelCase__ : Dict = [] for ex_index, example in tqdm.tqdm(enumerate(SCREAMING_SNAKE_CASE_ ) , desc='convert examples to features' ): if ex_index % 10_000 == 0: logger.info('Writing example %d' % (ex_index) ) lowerCAmelCase__ : Union[str, Any] = tokenizer( example.text_a , example.text_b , add_special_tokens=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , padding='max_length' , truncation=SCREAMING_SNAKE_CASE_ , return_overflowing_tokens=SCREAMING_SNAKE_CASE_ , ) lowerCAmelCase__ : List[str] = label_map[example.label] if example.label in label_map else 0 lowerCAmelCase__ : Any = int(example.pairID ) features.append(InputFeatures(**SCREAMING_SNAKE_CASE_ , label=SCREAMING_SNAKE_CASE_ , pairID=SCREAMING_SNAKE_CASE_ ) ) for i, example in enumerate(examples[:5] ): logger.info('*** Example ***' ) logger.info(F'''guid: {example}''' ) logger.info(F'''features: {features[i]}''' ) return features lowerCamelCase__ = { """hans""": 3, } lowerCamelCase__ = { """hans""": HansProcessor, }
720
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> float: if principal <= 0: raise Exception('Principal borrowed must be > 0' ) if rate_per_annum < 0: raise Exception('Rate of interest must be >= 0' ) if years_to_repay <= 0 or not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise Exception('Years to repay must be an integer > 0' ) # Yearly rate is divided by 12 to get monthly rate lowerCAmelCase__ : Optional[int] = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly lowerCAmelCase__ : str = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
721
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
69
0
import argparse from collections import defaultdict def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : int = F'''{file}_{class_name}_{test_name}''' done_test[_id] += 1 with open(SCREAMING_SNAKE_CASE_ , 'r' ) as f: lowerCAmelCase__ : Optional[Any] = f.readlines() lowerCAmelCase__ : List[str] = F'''class {class_name}(''' lowerCAmelCase__ : Union[str, Any] = F'''{4 * " "}def {test_name}(''' lowerCAmelCase__ : List[str] = F'''{8 * " "}{correct_line.split()[0]}''' lowerCAmelCase__ : Optional[Any] = F'''{16 * " "}{correct_line.split()[0]}''' lowerCAmelCase__ : Any = False lowerCAmelCase__ : Tuple = False lowerCAmelCase__ : List[Any] = False lowerCAmelCase__ : List[str] = False lowerCAmelCase__ : Dict = 0 lowerCAmelCase__ : List[str] = 0 lowerCAmelCase__ : Tuple = [] for line in lines: if line.startswith(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Any = True elif in_class and line.startswith(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Any = True elif in_class and in_func and (line.startswith(SCREAMING_SNAKE_CASE_ ) or line.startswith(SCREAMING_SNAKE_CASE_ )): lowerCAmelCase__ : Optional[Any] = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: lowerCAmelCase__ : Tuple = True if in_class and in_func and in_line: if ")" not in line: continue else: lowerCAmelCase__ : Optional[Any] = True if in_class and in_func and in_line and insert_line: new_lines.append(F'''{spaces * " "}{correct_line}''' ) lowerCAmelCase__ : Dict = False else: new_lines.append(SCREAMING_SNAKE_CASE_ ) with open(SCREAMING_SNAKE_CASE_ , 'w' ) as f: for line in new_lines: f.write(SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None ): if fail is not None: with open(SCREAMING_SNAKE_CASE_ , 'r' ) as f: lowerCAmelCase__ : Union[str, Any] = {l.strip() for l in f.readlines()} else: lowerCAmelCase__ : Tuple = None with open(SCREAMING_SNAKE_CASE_ , 'r' ) as f: lowerCAmelCase__ : Dict = f.readlines() lowerCAmelCase__ : Optional[int] = defaultdict(SCREAMING_SNAKE_CASE_ ) for line in correct_lines: lowerCAmelCase__ : Tuple = line.split(';' ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument("""--correct_filename""", help="""filename of tests with expected result""") parser.add_argument("""--fail_filename""", help="""filename of test failures""", type=str, default=None) lowerCamelCase__ = parser.parse_args() main(args.correct_filename, args.fail_filename)
700
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __magic_name__ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'LayoutLMv3ImageProcessor' lowercase = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self : Optional[int] , a : Union[str, Any]=None , a : Optional[Any]=None , **a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , a , ) lowerCAmelCase__ : int = kwargs.pop('feature_extractor' ) lowerCAmelCase__ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(a , a ) def __call__( self : List[Any] , a : List[Any] , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a : Union[List[List[int]], List[List[List[int]]]] = None , a : Optional[Union[List[int], List[List[int]]]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : str , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor lowerCAmelCase__ : List[str] = self.image_processor(images=a , return_tensors=a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a , a ): lowerCAmelCase__ : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ : List[str] = features['words'] lowerCAmelCase__ : List[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_token_type_ids=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_length=a , verbose=a , return_tensors=a , **a , ) # add pixel values lowerCAmelCase__ : Tuple = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase__ : List[str] = self.get_overflowing_images(a , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase__ : List[str] = images return encoded_inputs def _lowerCamelCase ( self : Any , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a ) != len(a ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' f''' {len(a )} and {len(a )}''' ) return images_with_overflow def _lowerCamelCase ( self : Union[str, Any] , *a : Optional[Any] , **a : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Tuple , *a : List[str] , **a : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*a , **a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , a , ) return self.image_processor_class @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , a , ) return self.image_processor
69
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """microsoft/deberta-v2-xlarge""": """https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json""", """microsoft/deberta-v2-xxlarge""": """https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json""", """microsoft/deberta-v2-xlarge-mnli""": ( """https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json""" ), """microsoft/deberta-v2-xxlarge-mnli""": ( """https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json""" ), } class A__ ( __magic_name__ ): lowercase = 'deberta-v2' def __init__( self : Dict , a : Any=128_100 , a : List[Any]=1_536 , a : List[str]=24 , a : List[Any]=24 , a : int=6_144 , a : List[Any]="gelu" , a : int=0.1 , a : Union[str, Any]=0.1 , a : Dict=512 , a : Optional[int]=0 , a : Tuple=0.0_2 , a : Optional[int]=1E-7 , a : Optional[int]=False , a : Tuple=-1 , a : Dict=0 , a : Dict=True , a : Optional[int]=None , a : Optional[Any]=0 , a : Optional[int]="gelu" , **a : Tuple , ): '''simple docstring''' super().__init__(**a ) lowerCAmelCase__ : int = hidden_size lowerCAmelCase__ : List[str] = num_hidden_layers lowerCAmelCase__ : List[Any] = num_attention_heads lowerCAmelCase__ : Any = intermediate_size lowerCAmelCase__ : str = hidden_act lowerCAmelCase__ : List[Any] = hidden_dropout_prob lowerCAmelCase__ : List[str] = attention_probs_dropout_prob lowerCAmelCase__ : List[str] = max_position_embeddings lowerCAmelCase__ : Union[str, Any] = type_vocab_size lowerCAmelCase__ : Optional[Any] = initializer_range lowerCAmelCase__ : List[Any] = relative_attention lowerCAmelCase__ : Optional[Any] = max_relative_positions lowerCAmelCase__ : Union[str, Any] = pad_token_id lowerCAmelCase__ : str = position_biased_input # Backwards compatibility if type(a ) == str: lowerCAmelCase__ : Optional[Any] = [x.strip() for x in pos_att_type.lower().split('|' )] lowerCAmelCase__ : Dict = pos_att_type lowerCAmelCase__ : Tuple = vocab_size lowerCAmelCase__ : Dict = layer_norm_eps lowerCAmelCase__ : str = kwargs.get('pooler_hidden_size' , a ) lowerCAmelCase__ : Any = pooler_dropout lowerCAmelCase__ : Dict = pooler_hidden_act class A__ ( __magic_name__ ): @property def _lowerCamelCase ( self : Any ): '''simple docstring''' if self.task == "multiple-choice": lowerCAmelCase__ : Optional[Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: lowerCAmelCase__ : int = {0: 'batch', 1: 'sequence'} if self._config.type_vocab_size > 0: return OrderedDict( [('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis)] ) else: return OrderedDict([('input_ids', dynamic_axis), ('attention_mask', dynamic_axis)] ) @property def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return 12 def _lowerCamelCase ( self : Optional[Any] , a : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , a : int = -1 , a : int = -1 , a : int = -1 , a : bool = False , a : Optional["TensorType"] = None , a : int = 3 , a : int = 40 , a : int = 40 , a : "PreTrainedTokenizerBase" = None , ): '''simple docstring''' lowerCAmelCase__ : List[str] = super().generate_dummy_inputs(preprocessor=a , framework=a ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
701
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
69
0
from __future__ import annotations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list[int]: lowerCAmelCase__ : Tuple = 2 lowerCAmelCase__ : Any = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(SCREAMING_SNAKE_CASE_ ) if n > 1: factors.append(SCREAMING_SNAKE_CASE_ ) return factors if __name__ == "__main__": import doctest doctest.testmod()
702
lowerCamelCase__ = """Alexander Joslin""" import operator as op from .stack import Stack def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} lowerCAmelCase__ : Stack[int] = Stack() lowerCAmelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE_ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE_ ) elif i == ")": # RULE 4 lowerCAmelCase__ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : List[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : Tuple = operators[opr](SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) operand_stack.push(SCREAMING_SNAKE_CASE_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
69
0
import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['input_features', 'is_longer'] def __init__( self : Any , a : Union[str, Any]=64 , a : int=48_000 , a : int=480 , a : Any=10 , a : Union[str, Any]=1_024 , a : List[Any]=0.0 , a : Union[str, Any]=False , a : float = 0 , a : float = 14_000 , a : int = None , a : str = "fusion" , a : str = "repeatpad" , **a : Any , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , return_attention_mask=a , **a , ) lowerCAmelCase__ : List[str] = top_db lowerCAmelCase__ : Dict = truncation lowerCAmelCase__ : Dict = padding lowerCAmelCase__ : List[str] = fft_window_size lowerCAmelCase__ : Any = (fft_window_size >> 1) + 1 lowerCAmelCase__ : Optional[Any] = hop_length lowerCAmelCase__ : Optional[int] = max_length_s lowerCAmelCase__ : Any = max_length_s * sampling_rate lowerCAmelCase__ : Optional[Any] = sampling_rate lowerCAmelCase__ : Any = frequency_min lowerCAmelCase__ : Dict = frequency_max lowerCAmelCase__ : Optional[Any] = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=a , min_frequency=a , max_frequency=a , sampling_rate=a , norm=a , mel_scale='htk' , ) lowerCAmelCase__ : Tuple = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=a , min_frequency=a , max_frequency=a , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = copy.deepcopy(self.__dict__ ) lowerCAmelCase__ : Any = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def _lowerCamelCase ( self : Any , a : np.array , a : Optional[np.array] = None ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = spectrogram( a , window_function(self.fft_window_size , 'hann' ) , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=a , log_mel='dB' , ) return log_mel_spectrogram.T def _lowerCamelCase ( self : Any , a : Dict , a : Dict , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = np.array_split(list(range(0 , total_frames - chunk_frames + 1 ) ) , 3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk lowerCAmelCase__ : List[Any] = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk lowerCAmelCase__ : Optional[int] = [0] # randomly choose index for each part lowerCAmelCase__ : Dict = np.random.choice(ranges[0] ) lowerCAmelCase__ : Tuple = np.random.choice(ranges[1] ) lowerCAmelCase__ : Optional[int] = np.random.choice(ranges[2] ) lowerCAmelCase__ : Optional[int] = mel[idx_front : idx_front + chunk_frames, :] lowerCAmelCase__ : Optional[int] = mel[idx_middle : idx_middle + chunk_frames, :] lowerCAmelCase__ : Optional[int] = mel[idx_back : idx_back + chunk_frames, :] lowerCAmelCase__ : List[Any] = torch.tensor(mel[None, None, :] ) lowerCAmelCase__ : List[str] = torch.nn.functional.interpolate( a , size=[chunk_frames, 64] , mode='bilinear' , align_corners=a ) lowerCAmelCase__ : List[str] = mel_shrink[0][0].numpy() lowerCAmelCase__ : Any = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0 ) return mel_fusion def _lowerCamelCase ( self : List[Any] , a : np.array , a : List[str] , a : int , a : Optional[int] ): '''simple docstring''' if waveform.shape[0] > max_length: if truncation == "rand_trunc": lowerCAmelCase__ : Optional[Any] = True # random crop to max_length (for compatibility) -> this should be handled by self.pad lowerCAmelCase__ : Optional[int] = len(a ) - max_length lowerCAmelCase__ : int = np.random.randint(0 , overflow + 1 ) lowerCAmelCase__ : Union[str, Any] = waveform[idx : idx + max_length] lowerCAmelCase__ : Union[str, Any] = self._np_extract_fbank_features(a , self.mel_filters_slaney )[None, :] elif truncation == "fusion": lowerCAmelCase__ : Dict = self._np_extract_fbank_features(a , self.mel_filters ) lowerCAmelCase__ : List[str] = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed lowerCAmelCase__ : Union[str, Any] = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. lowerCAmelCase__ : List[Any] = np.stack([mel, mel, mel, mel] , axis=0 ) lowerCAmelCase__ : Any = False else: lowerCAmelCase__ : Union[str, Any] = self._random_mel_fusion(a , a , a ) lowerCAmelCase__ : Tuple = True else: raise NotImplementedError(f'''data_truncating {truncation} not implemented''' ) else: lowerCAmelCase__ : Dict = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": lowerCAmelCase__ : Optional[Any] = int(max_length / len(a ) ) lowerCAmelCase__ : Any = np.stack(np.tile(a , n_repeat + 1 ) )[:max_length] if padding == "repeatpad": lowerCAmelCase__ : str = int(max_length / len(a ) ) lowerCAmelCase__ : Tuple = np.stack(np.tile(a , a ) ) lowerCAmelCase__ : Optional[Any] = np.pad(a , (0, max_length - waveform.shape[0]) , mode='constant' , constant_values=0 ) if truncation == "fusion": lowerCAmelCase__ : Optional[int] = self._np_extract_fbank_features(a , self.mel_filters ) lowerCAmelCase__ : Tuple = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0 ) else: lowerCAmelCase__ : str = self._np_extract_fbank_features(a , self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self : Union[str, Any] , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : str = None , a : Optional[str] = None , a : Optional[int] = None , a : Optional[int] = None , a : Optional[Union[str, TensorType]] = None , **a : Optional[int] , ): '''simple docstring''' lowerCAmelCase__ : Any = truncation if truncation is not None else self.truncation lowerCAmelCase__ : Any = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a''' f''' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input''' f''' was sampled with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : str = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : Optional[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : List[Any] = [np.asarray(a , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Union[str, Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : Dict = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Optional[int] = [np.asarray(a )] # convert to mel spectrogram, truncate and pad if needed. lowerCAmelCase__ : Optional[Any] = [ self._get_input_mel(a , max_length if max_length else self.nb_max_samples , a , a ) for waveform in raw_speech ] lowerCAmelCase__ : Any = [] lowerCAmelCase__ : Any = [] for mel, longer in padded_inputs: input_mel.append(a ) is_longer.append(a ) if truncation == "fusion" and sum(a ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer lowerCAmelCase__ : str = np.random.randint(0 , len(a ) ) lowerCAmelCase__ : Any = True if isinstance(input_mel[0] , a ): lowerCAmelCase__ : List[str] = [np.asarray(a , dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool lowerCAmelCase__ : str = [[longer] for longer in is_longer] lowerCAmelCase__ : str = {'input_features': input_mel, 'is_longer': is_longer} lowerCAmelCase__ : Dict = BatchFeature(a ) if return_tensors is not None: lowerCAmelCase__ : Tuple = input_features.convert_to_tensors(a ) return input_features
703
import numpy class A__ : def __init__( self : Tuple , a : numpy.ndarray , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : int = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Dict = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : List[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowerCAmelCase__ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : List[Any] = numpy.zeros(output_array.shape ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowerCAmelCase__ : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowerCAmelCase__ : int = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowerCamelCase ( self : Optional[int] , a : numpy.ndarray , a : int , a : bool ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): lowerCAmelCase__ : Any = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Tuple = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def _lowerCamelCase ( self : Optional[Any] , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : Dict = input_arr lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return (value) * (1 - (value)) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : List[str] = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ , output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ , iterations=10 , give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
69
0
from datetime import datetime import requests def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bytes: lowerCAmelCase__ : Optional[int] = 'https://downloadgram.net/wp-json/wppress/video-downloader/video?url=' lowerCAmelCase__ : Tuple = requests.get(base_url + url ).json()[0]['urls'][0]['src'] return requests.get(SCREAMING_SNAKE_CASE_ ).content if __name__ == "__main__": lowerCamelCase__ = input("""Enter Video/IGTV url: """).strip() lowerCamelCase__ = F"""{datetime.now():%Y-%m-%d_%H:%M:%S}.mp4""" with open(file_name, """wb""") as fp: fp.write(download_video(url)) print(F"""Done. Video saved to disk as {file_name}.""")
704
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A__ : def __init__( self : int , a : str , a : Union[str, Any]=13 , a : int=32 , a : Optional[Any]=2 , a : Tuple=3 , a : List[Any]=16 , a : List[str]=[1, 2, 1] , a : int=[2, 2, 4] , a : int=2 , a : Optional[Any]=2.0 , a : Optional[int]=True , a : Dict=0.0 , a : Any=0.0 , a : int=0.1 , a : List[str]="gelu" , a : Optional[Any]=False , a : str=True , a : Dict=0.0_2 , a : Any=1E-5 , a : Optional[int]=True , a : str=None , a : str=True , a : int=10 , a : str=8 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : Optional[Any] = patch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = embed_dim lowerCAmelCase__ : Tuple = depths lowerCAmelCase__ : List[str] = num_heads lowerCAmelCase__ : List[Any] = window_size lowerCAmelCase__ : Any = mlp_ratio lowerCAmelCase__ : Optional[Any] = qkv_bias lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : int = drop_path_rate lowerCAmelCase__ : Optional[Any] = hidden_act lowerCAmelCase__ : int = use_absolute_embeddings lowerCAmelCase__ : List[str] = patch_norm lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Dict = use_labels lowerCAmelCase__ : List[Any] = type_sequence_label_size lowerCAmelCase__ : Optional[Any] = encoder_stride def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = None if self.use_labels: lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : int = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : List[str] , a : Any , a : str , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = SwinvaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model(a ) lowerCAmelCase__ : str = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase__ : Tuple = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SwinvaForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : Dict = SwinvaForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Union[str, Any] , a : int , a : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.type_sequence_label_size lowerCAmelCase__ : List[Any] = SwinvaForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = config_and_inputs lowerCAmelCase__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = SwinvaModelTester(self ) lowerCAmelCase__ : int = ConfigTester(self , config_class=a , embed_dim=37 ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[int] = model_class(a ) lowerCAmelCase__ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Tuple = [*signature.parameters.keys()] lowerCAmelCase__ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = True for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = True lowerCAmelCase__ : str = False lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Dict = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : int = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Dict = outputs.attentions lowerCAmelCase__ : Dict = len(self.model_tester.depths ) self.assertEqual(len(a ) , a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Optional[int] = config.window_size**2 lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowerCAmelCase__ : Tuple = len(a ) # Check attention is always last and order is fine lowerCAmelCase__ : str = True lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(**self._prepare_for_class(a , a ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): lowerCAmelCase__ : Optional[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase__ : Any = 2 self.assertEqual(out_len + added_hidden_states , len(a ) ) lowerCAmelCase__ : Dict = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _lowerCamelCase ( self : int , a : Optional[int] , a : int , a : Optional[Any] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.hidden_states lowerCAmelCase__ : str = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a ) , a ) # Swinv2 has a different seq_length lowerCAmelCase__ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowerCAmelCase__ : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(a ) , a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = reshaped_hidden_states[0].shape lowerCAmelCase__ : List[str] = ( reshaped_hidden_states[0].view(a , a , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : str = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : List[str] = True self.check_hidden_states_output(a , a , a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Any = 3 lowerCAmelCase__ : int = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase__ : str = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : str = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase__ : Any = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase__ : str = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[str] = SwinvaModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = _config_zero_init(a ) for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(config=a ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( a ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Any = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**a ) # verify the logits lowerCAmelCase__ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Optional[Any] = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) )
69
0
import requests from bsa import BeautifulSoup def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = "https://www.worldometers.info/coronavirus" ) -> dict: lowerCAmelCase__ : List[str] = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE_ ).text , 'html.parser' ) lowerCAmelCase__ : int = soup.findAll('h1' ) lowerCAmelCase__ : Optional[int] = soup.findAll('div' , {'class': 'maincounter-number'} ) keys += soup.findAll('span' , {'class': 'panel-title'} ) values += soup.findAll('div' , {'class': 'number-table-main'} ) return {key.text.strip(): value.text.strip() for key, value in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )} if __name__ == "__main__": print("""\033[1m""" + """COVID-19 Status of the World""" + """\033[0m\n""") for key, value in world_covidaa_stats().items(): print(F"""{key}\n{value}\n""")
705
from itertools import permutations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase__ : str = [7, 11, 13, 17] for i, test in enumerate(SCREAMING_SNAKE_CASE_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 10 ) -> int: return sum( int(''.join(map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) ) for num in permutations(range(SCREAMING_SNAKE_CASE_ ) ) if is_substring_divisible(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise TypeError('\'float\' object cannot be interpreted as an integer' ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise TypeError('\'str\' object cannot be interpreted as an integer' ) if num == 0: return "0b0" lowerCAmelCase__ : List[Any] = False if num < 0: lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : str = -num lowerCAmelCase__ : list[int] = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(SCREAMING_SNAKE_CASE_ ) for e in binary ) return "0b" + "".join(str(SCREAMING_SNAKE_CASE_ ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
706
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
69
0
from __future__ import annotations lowerCamelCase__ = """Muhammad Umer Farooq""" lowerCamelCase__ = """MIT""" lowerCamelCase__ = """1.0.0""" lowerCamelCase__ = """Muhammad Umer Farooq""" lowerCamelCase__ = """contact@muhammadumerfarooq.me""" lowerCamelCase__ = """Alpha""" import re from html.parser import HTMLParser from urllib import parse import requests class A__ ( __magic_name__ ): '''simple docstring''' def __init__( self : Optional[int] , a : str ): '''simple docstring''' super().__init__() lowerCAmelCase__ : list[str] = [] lowerCAmelCase__ : Union[str, Any] = domain def _lowerCamelCase ( self : int , a : str , a : list[tuple[str, str | None]] ): '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: lowerCAmelCase__ : int = parse.urljoin(self.domain , a ) self.urls.append(a ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: return ".".join(get_sub_domain_name(SCREAMING_SNAKE_CASE_ ).split('.' )[-2:] ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: return parse.urlparse(SCREAMING_SNAKE_CASE_ ).netloc def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = "https://github.com" ) -> list[str]: lowerCAmelCase__ : Optional[Any] = get_domain_name(SCREAMING_SNAKE_CASE_ ) # Initialize the parser lowerCAmelCase__ : Optional[int] = Parser(SCREAMING_SNAKE_CASE_ ) try: # Open URL lowerCAmelCase__ : Union[str, Any] = requests.get(SCREAMING_SNAKE_CASE_ ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through lowerCAmelCase__ : Optional[Any] = set() for link in parser.urls: # open URL. # read = requests.get(link) try: lowerCAmelCase__ : int = requests.get(SCREAMING_SNAKE_CASE_ ) # Get the valid email. lowerCAmelCase__ : Union[str, Any] = re.findall('[a-zA-Z0-9]+@' + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(SCREAMING_SNAKE_CASE_ ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowerCamelCase__ = emails_from_url("""https://github.com""") print(F"""{len(emails)} emails found:""") print("""\n""".join(sorted(emails)))
707
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ , lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
69
0
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem lowerCamelCase__ = importlib.util.find_spec("""s3fs""") is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 lowerCamelCase__ = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(F"""A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.""") fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: if "://" in dataset_path: lowerCAmelCase__ : int = dataset_path.split('://' )[1] return dataset_path def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if fs is not None and fs.protocol != "file": return True else: return False def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: lowerCAmelCase__ : int = not is_remote_filesystem(SCREAMING_SNAKE_CASE_ ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(SCREAMING_SNAKE_CASE_ ) , fs._strip_protocol(SCREAMING_SNAKE_CASE_ ) ) else: fs.mv(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , recursive=SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( ) -> None: if hasattr(fsspec.asyn , 'reset_lock' ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : Optional[int] = None lowerCAmelCase__ : List[str] = threading.Lock()
708
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
0
from __future__ import annotations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> set[str]: lowerCAmelCase__ : Union[str, Any] = set(SCREAMING_SNAKE_CASE_ ), [start] while stack: lowerCAmelCase__ : Optional[int] = stack.pop() explored.add(SCREAMING_SNAKE_CASE_ ) # Differences from BFS: # 1) pop last element instead of first one # 2) add adjacent elements to stack without exploring them for adj in reversed(graph[v] ): if adj not in explored: stack.append(SCREAMING_SNAKE_CASE_ ) return explored lowerCamelCase__ = { """A""": ["""B""", """C""", """D"""], """B""": ["""A""", """D""", """E"""], """C""": ["""A""", """F"""], """D""": ["""B""", """D"""], """E""": ["""B""", """F"""], """F""": ["""C""", """E""", """G"""], """G""": ["""F"""], } if __name__ == "__main__": import doctest doctest.testmod() print(depth_first_search(G, """A"""))
709
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """google/rembert""": 256, } lowerCamelCase__ = """▁""" class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = RemBertTokenizer def __init__( self : Optional[Any] , a : str=None , a : Any=None , a : List[Any]=True , a : str=True , a : Dict=False , a : Dict="[CLS]" , a : int="[SEP]" , a : Tuple="<unk>" , a : Optional[Any]="[SEP]" , a : Tuple="<pad>" , a : Dict="[CLS]" , a : Optional[Any]="[MASK]" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( a , tokenizer_file=a , do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : int = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : List[Any] = keep_accents lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
69
0
import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class A__ ( __magic_name__ ): lowercase = '' lowercase = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) lowercase = None # compression type in fsspec. ex: "gzip" lowercase = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self : Optional[Any] , a : str = "" , a : Optional[str] = None , a : Optional[dict] = None , **a : Any ): '''simple docstring''' super().__init__(self , **a ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode lowerCAmelCase__ : Tuple = fsspec.open( a , mode='rb' , protocol=a , compression=self.compression , client_kwargs={ 'requote_redirect_url': False, # see https://github.com/huggingface/datasets/pull/5459 'trust_env': True, # Enable reading proxy env variables. **(target_options or {}).pop('client_kwargs' , {} ), # To avoid issues if it was already passed. } , **(target_options or {}) , ) lowerCAmelCase__ : Optional[int] = os.path.basename(self.file.path.split('::' )[0] ) lowerCAmelCase__ : Optional[int] = ( self.compressed_name[: self.compressed_name.rindex('.' )] if '.' in self.compressed_name else self.compressed_name ) lowerCAmelCase__ : Optional[int] = None @classmethod def _lowerCamelCase ( cls : Dict , a : Optional[int] ): '''simple docstring''' return super()._strip_protocol(a ).lstrip('/' ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' if self.dir_cache is None: lowerCAmelCase__ : List[str] = {**self.file.fs.info(self.file.path ), 'name': self.uncompressed_name} lowerCAmelCase__ : Optional[int] = {f['name']: f} def _lowerCamelCase ( self : List[str] , a : str ): '''simple docstring''' return self.file.open().read() def _lowerCamelCase ( self : Optional[int] , a : str , a : str = "rb" , a : Any=None , a : str=True , a : List[Any]=None , **a : List[Any] , ): '''simple docstring''' lowerCAmelCase__ : List[str] = self._strip_protocol(a ) if mode != "rb": raise ValueError(f'''Tried to read with mode {mode} on file {self.file.path} opened with mode \'rb\'''' ) return self.file.open() class A__ ( __magic_name__ ): lowercase = 'bz2' lowercase = 'bz2' lowercase = '.bz2' class A__ ( __magic_name__ ): lowercase = 'gzip' lowercase = 'gzip' lowercase = '.gz' class A__ ( __magic_name__ ): lowercase = 'lz4' lowercase = 'lz4' lowercase = '.lz4' class A__ ( __magic_name__ ): lowercase = 'xz' lowercase = 'xz' lowercase = '.xz' class A__ ( __magic_name__ ): lowercase = 'zstd' lowercase = 'zstd' lowercase = '.zst' def __init__( self : Optional[Any] , a : str , a : str = "rb" , a : Optional[str] = None , a : Optional[dict] = None , a : int = DEFAULT_BLOCK_SIZE , **a : Optional[int] , ): '''simple docstring''' super().__init__( fo=a , mode=a , target_protocol=a , target_options=a , block_size=a , **a , ) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 lowerCAmelCase__ : Dict = self.file.__enter__ class A__ : def __init__( self : Dict , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = file_ def __enter__( self : Any ): '''simple docstring''' self._file.__enter__() return self def __exit__( self : List[Any] , *a : Optional[Any] , **a : str ): '''simple docstring''' self._file.__exit__(*a , **a ) def __iter__( self : Union[str, Any] ): '''simple docstring''' return iter(self._file ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return next(self._file ) def __getattr__( self : List[Any] , a : Optional[int] ): '''simple docstring''' return getattr(self._file , a ) def fixed_enter(*a : str , **a : int ): return WrappedFile(_enter(*a , **a ) ) lowerCAmelCase__ : List[Any] = fixed_enter
710
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : int = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : List[Any] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : int = None lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Dict = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Union[str, Any] = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
69
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter lowerCamelCase__ = """Create a default config file for Accelerate with only a few flags set.""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_="no" , SCREAMING_SNAKE_CASE_ = default_json_config_file , SCREAMING_SNAKE_CASE_ = False ) -> str: lowerCAmelCase__ : Optional[Any] = Path(SCREAMING_SNAKE_CASE_ ) path.parent.mkdir(parents=SCREAMING_SNAKE_CASE_ , exist_ok=SCREAMING_SNAKE_CASE_ ) if path.exists(): print( F'''Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.''' ) return False lowerCAmelCase__ : Tuple = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( F'''`mixed_precision` should be one of \'no\', \'fp16\', \'bf16\', or \'fp8\'. Received {mixed_precision}''' ) lowerCAmelCase__ : Any = { 'compute_environment': 'LOCAL_MACHINE', 'mixed_precision': mixed_precision, } if torch.cuda.is_available(): lowerCAmelCase__ : Union[str, Any] = torch.cuda.device_count() lowerCAmelCase__ : int = num_gpus lowerCAmelCase__ : Any = False if num_gpus > 1: lowerCAmelCase__ : Dict = 'MULTI_GPU' else: lowerCAmelCase__ : Any = 'NO' elif is_xpu_available() and use_xpu: lowerCAmelCase__ : Tuple = torch.xpu.device_count() lowerCAmelCase__ : int = num_xpus lowerCAmelCase__ : Any = False if num_xpus > 1: lowerCAmelCase__ : Dict = 'MULTI_XPU' else: lowerCAmelCase__ : Dict = 'NO' elif is_npu_available(): lowerCAmelCase__ : int = torch.npu.device_count() lowerCAmelCase__ : Optional[int] = num_npus lowerCAmelCase__ : List[str] = False if num_npus > 1: lowerCAmelCase__ : Union[str, Any] = 'MULTI_NPU' else: lowerCAmelCase__ : int = 'NO' else: lowerCAmelCase__ : str = 0 lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Optional[int] = 1 lowerCAmelCase__ : List[str] = 'NO' lowerCAmelCase__ : List[str] = ClusterConfig(**SCREAMING_SNAKE_CASE_ ) config.to_json_file(SCREAMING_SNAKE_CASE_ ) return path def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = parser.add_parser('default' , parents=SCREAMING_SNAKE_CASE_ , help=SCREAMING_SNAKE_CASE_ , formatter_class=SCREAMING_SNAKE_CASE_ ) parser.add_argument( '--config_file' , default=SCREAMING_SNAKE_CASE_ , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , dest='save_location' , ) parser.add_argument( '--mixed_precision' , choices=['no', 'fp16', 'bf16'] , type=SCREAMING_SNAKE_CASE_ , help='Whether or not to use mixed precision training. ' 'Choose between FP16 and BF16 (bfloat16) training. ' 'BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later.' , default='no' , ) parser.set_defaults(func=SCREAMING_SNAKE_CASE_ ) return parser def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Dict: lowerCAmelCase__ : Any = write_basic_config(args.mixed_precision , args.save_location ) if config_file: print(F'''accelerate configuration saved at {config_file}''' )
711
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: assert column_title.isupper() lowerCAmelCase__ : Optional[int] = 0 lowerCAmelCase__ : int = len(SCREAMING_SNAKE_CASE_ ) - 1 lowerCAmelCase__ : Union[str, Any] = 0 while index >= 0: lowerCAmelCase__ : int = (ord(column_title[index] ) - 64) * pow(26 , SCREAMING_SNAKE_CASE_ ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
712
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
0
import json import os import re import sys import urllib.request import requests from bsa import BeautifulSoup lowerCamelCase__ = { """User-Agent""": """Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36""" """ (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582""" } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = "dhaka" , SCREAMING_SNAKE_CASE_ = 5 ) -> int: lowerCAmelCase__ : List[str] = min(SCREAMING_SNAKE_CASE_ , 50 ) # Prevent abuse! lowerCAmelCase__ : Optional[Any] = { 'q': query, 'tbm': 'isch', 'hl': 'en', 'ijn': '0', } lowerCAmelCase__ : Optional[Any] = requests.get('https://www.google.com/search' , params=SCREAMING_SNAKE_CASE_ , headers=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = BeautifulSoup(html.text , 'html.parser' ) lowerCAmelCase__ : List[Any] = ''.join( re.findall(r'AF_initDataCallback\(([^<]+)\);' , str(soup.select('script' ) ) ) ) lowerCAmelCase__ : List[str] = json.dumps(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = json.loads(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = re.findall( r'\[\"GRID_STATE0\",null,\[\[1,\[0,\".*?\",(.*),\"All\",' , SCREAMING_SNAKE_CASE_ , ) if not matched_google_image_data: return 0 lowerCAmelCase__ : Dict = re.sub( r'\[\"(https\:\/\/encrypted-tbn0\.gstatic\.com\/images\?.*?)\",\d+,\d+\]' , '' , str(SCREAMING_SNAKE_CASE_ ) , ) lowerCAmelCase__ : Optional[int] = re.findall( r'(?:\'|,),\[\"(https:|http.*?)\",\d+,\d+\]' , SCREAMING_SNAKE_CASE_ , ) for index, fixed_full_res_image in enumerate(SCREAMING_SNAKE_CASE_ ): if index >= max_images: return index lowerCAmelCase__ : Any = bytes(SCREAMING_SNAKE_CASE_ , 'ascii' ).decode( 'unicode-escape' ) lowerCAmelCase__ : Tuple = bytes(SCREAMING_SNAKE_CASE_ , 'ascii' ).decode( 'unicode-escape' ) lowerCAmelCase__ : Optional[int] = urllib.request.build_opener() lowerCAmelCase__ : Optional[int] = [ ( 'User-Agent', 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36' ' (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582', ) ] urllib.request.install_opener(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = F'''query_{query.replace(" " , "_" )}''' if not os.path.exists(SCREAMING_SNAKE_CASE_ ): os.makedirs(SCREAMING_SNAKE_CASE_ ) urllib.request.urlretrieve( # noqa: S310 SCREAMING_SNAKE_CASE_ , F'''{path_name}/original_size_img_{index}.jpg''' ) return index if __name__ == "__main__": try: lowerCamelCase__ = download_images_from_google_query(sys.argv[1]) print(F"""{image_count} images were downloaded to disk.""") except IndexError: print("""Please provide a search term.""") raise
713
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
0
import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = mock.Mock() lowerCAmelCase__ : List[str] = 500 lowerCAmelCase__ : List[str] = {} lowerCAmelCase__ : Union[str, Any] = HTTPError lowerCAmelCase__ : Any = {} # Download this model to make sure it's in the cache. lowerCAmelCase__ : int = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=a ) as mock_head: lowerCAmelCase__ : Any = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = mock.Mock() lowerCAmelCase__ : Dict = 500 lowerCAmelCase__ : str = {} lowerCAmelCase__ : Tuple = HTTPError lowerCAmelCase__ : Optional[int] = {} # Download this model to make sure it's in the cache. lowerCAmelCase__ : int = GPTaTokenizerFast.from_pretrained('gpt2' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=a ) as mock_head: lowerCAmelCase__ : Union[str, Any] = GPTaTokenizerFast.from_pretrained('gpt2' ) # This check we did call the fake head request mock_head.assert_called() def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' try: lowerCAmelCase__ : Dict = tempfile.mktemp() with open(a , 'wb' ) as f: http_get('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' , a ) lowerCAmelCase__ : Any = AlbertTokenizer.from_pretrained(a ) finally: os.remove(a ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile('tokenizer.json' ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open('tokenizer.json' , 'wb' ) as f: http_get('https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json' , a ) lowerCAmelCase__ : int = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1_000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove('tokenizer.json' ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AlbertTokenizer.from_pretrained('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' ) @is_staging_test class A__ ( unittest.TestCase ): lowercase = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] @classmethod def _lowerCamelCase ( cls : str ): '''simple docstring''' lowerCAmelCase__ : str = TOKEN HfFolder.save_token(a ) @classmethod def _lowerCamelCase ( cls : List[str] ): '''simple docstring''' try: delete_repo(token=cls._token , repo_id='test-tokenizer' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-tokenizer-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-tokenizer' ) except HTTPError: pass def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : str = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : List[Any] = BertTokenizer(a ) tokenizer.push_to_hub('test-tokenizer' , use_auth_token=self._token ) lowerCAmelCase__ : Optional[Any] = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='test-tokenizer' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(a , repo_id='test-tokenizer' , push_to_hub=a , use_auth_token=self._token ) lowerCAmelCase__ : List[Any] = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def _lowerCamelCase ( self : int ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : Union[str, Any] = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : Optional[Any] = BertTokenizer(a ) tokenizer.push_to_hub('valid_org/test-tokenizer-org' , use_auth_token=self._token ) lowerCAmelCase__ : Union[str, Any] = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-tokenizer-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( a , repo_id='valid_org/test-tokenizer-org' , push_to_hub=a , use_auth_token=self._token ) lowerCAmelCase__ : Dict = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : Dict = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : List[str] = CustomTokenizer(a ) # No fast custom tokenizer tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) lowerCAmelCase__ : Any = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : int = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : str = BertTokenizerFast.from_pretrained(a ) bert_tokenizer.save_pretrained(a ) lowerCAmelCase__ : Union[str, Any] = CustomTokenizerFast.from_pretrained(a ) tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) lowerCAmelCase__ : Optional[int] = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizerFast' ) lowerCAmelCase__ : List[Any] = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=a , trust_remote_code=a ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = Trie() trie.add('Hello 友達' ) self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {' ': {'友': {'達': {'': 1}}}}}}}}} ) trie.add('Hello' ) trie.data self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {'': 1, ' ': {'友': {'達': {'': 1}}}}}}}}} ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = Trie() self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS] This is a extra_id_100'] ) trie.add('[CLS]' ) trie.add('extra_id_1' ) trie.add('extra_id_100' ) self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS]', ' This is a ', 'extra_id_100'] ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = Trie() trie.add('A' ) self.assertEqual(trie.split('ABC' ) , ['A', 'BC'] ) self.assertEqual(trie.split('BCA' ) , ['BC', 'A'] ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = Trie() trie.add('TOKEN]' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = Trie() trie.add('A' ) trie.add('P' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = Trie() trie.add('AB' ) trie.add('B' ) trie.add('C' ) self.assertEqual(trie.split('ABC' ) , ['AB', 'C'] ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = Trie() trie.add('ABC' ) trie.add('B' ) trie.add('CD' ) self.assertEqual(trie.split('ABCD' ) , ['ABC', 'D'] ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = Trie() lowerCAmelCase__ : str = trie.cut_text('ABC' , [0, 0, 2, 1, 2, 3] ) self.assertEqual(a , ['AB', 'C'] )
714
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
0
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """facebook/mbart-large-en-ro""": ( """https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model""" ), """facebook/mbart-large-cc25""": ( """https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/mbart-large-en-ro""": """https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json""", """facebook/mbart-large-cc25""": """https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """facebook/mbart-large-en-ro""": 1024, """facebook/mbart-large-cc25""": 1024, } # fmt: off lowerCamelCase__ = ["""ar_AR""", """cs_CZ""", """de_DE""", """en_XX""", """es_XX""", """et_EE""", """fi_FI""", """fr_XX""", """gu_IN""", """hi_IN""", """it_IT""", """ja_XX""", """kk_KZ""", """ko_KR""", """lt_LT""", """lv_LV""", """my_MM""", """ne_NP""", """nl_XX""", """ro_RO""", """ru_RU""", """si_LK""", """tr_TR""", """vi_VN""", """zh_CN"""] class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = ['input_ids', 'attention_mask'] lowercase = MBartTokenizer lowercase = [] lowercase = [] def __init__( self : int , a : Tuple=None , a : Optional[int]=None , a : int="<s>" , a : Optional[int]="</s>" , a : int="</s>" , a : Tuple="<s>" , a : Dict="<unk>" , a : int="<pad>" , a : List[Any]="<mask>" , a : Optional[int]=None , a : Optional[int]=None , a : List[Any]=None , **a : Union[str, Any] , ): '''simple docstring''' lowerCAmelCase__ : int = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( vocab_file=a , tokenizer_file=a , bos_token=a , eos_token=a , sep_token=a , cls_token=a , unk_token=a , pad_token=a , mask_token=a , src_lang=a , tgt_lang=a , additional_special_tokens=a , **a , ) lowerCAmelCase__ : Tuple = vocab_file lowerCAmelCase__ : List[Any] = False if not self.vocab_file else True lowerCAmelCase__ : int = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens} ) lowerCAmelCase__ : List[str] = { lang_code: self.convert_tokens_to_ids(a ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCAmelCase__ : Optional[int] = src_lang if src_lang is not None else 'en_XX' lowerCAmelCase__ : Union[str, Any] = self.convert_tokens_to_ids(self._src_lang ) lowerCAmelCase__ : str = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return self._src_lang @src_lang.setter def _lowerCamelCase ( self : Optional[int] , a : str ): '''simple docstring''' lowerCAmelCase__ : List[Any] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def _lowerCamelCase ( self : Any , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _lowerCamelCase ( self : int , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = [self.sep_token_id] lowerCAmelCase__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : str , a : Optional[str] , a : Optional[str] , **a : List[Any] ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowerCAmelCase__ : Union[str, Any] = src_lang lowerCAmelCase__ : Optional[Any] = self(a , add_special_tokens=a , return_tensors=a , **a ) lowerCAmelCase__ : List[str] = self.convert_tokens_to_ids(a ) lowerCAmelCase__ : List[Any] = tgt_lang_id return inputs def _lowerCamelCase ( self : Optional[Any] , a : List[str] , a : str = "en_XX" , a : Optional[List[str]] = None , a : str = "ro_RO" , **a : Tuple , ): '''simple docstring''' lowerCAmelCase__ : str = src_lang lowerCAmelCase__ : Tuple = tgt_lang return super().prepare_seqaseq_batch(a , a , **a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def _lowerCamelCase ( self : int ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def _lowerCamelCase ( self : Union[str, Any] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.convert_tokens_to_ids(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Union[str, Any] = [self.eos_token_id, self.cur_lang_code] lowerCAmelCase__ : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase__ : List[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase__ : Dict = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def _lowerCamelCase ( self : Tuple , a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.convert_tokens_to_ids(a ) lowerCAmelCase__ : str = [] lowerCAmelCase__ : Any = [self.eos_token_id, self.cur_lang_code] lowerCAmelCase__ : Optional[Any] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCAmelCase__ : List[str] = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCAmelCase__ : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def _lowerCamelCase ( self : Any , a : str , a : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory.''' ) return lowerCAmelCase__ : List[str] = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
715
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
0
'''simple docstring''' import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class A__ : def __init__( self : List[str] , a : Optional[int]=2 , a : Optional[Any]=3 , a : List[Any]=64 , a : Any=None ): '''simple docstring''' lowerCAmelCase__ : Dict = np.random.default_rng(a ) lowerCAmelCase__ : Tuple = length lowerCAmelCase__ : Tuple = rng.normal(size=(length,) ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self : List[str] ): '''simple docstring''' return self.length def __getitem__( self : Union[str, Any] , a : Dict ): '''simple docstring''' return {"x": self.x[i], "y": self.y[i]} class A__ ( torch.nn.Module ): def __init__( self : Dict , a : Tuple=0 , a : Optional[Any]=0 , a : Optional[int]=False ): '''simple docstring''' super().__init__() lowerCAmelCase__ : Union[str, Any] = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) lowerCAmelCase__ : List[Any] = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) lowerCAmelCase__ : Any = True def _lowerCamelCase ( self : Union[str, Any] , a : Optional[int]=None ): '''simple docstring''' if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) lowerCAmelCase__ : Optional[Any] = False return x * self.a[0] + self.b[0] class A__ ( torch.nn.Module ): def __init__( self : Any , a : List[Any]=0 , a : Optional[Any]=0 , a : Tuple=False ): '''simple docstring''' super().__init__() lowerCAmelCase__ : int = torch.nn.Parameter(torch.tensor(a ).float() ) lowerCAmelCase__ : str = torch.nn.Parameter(torch.tensor(a ).float() ) lowerCAmelCase__ : Tuple = True def _lowerCamelCase ( self : str , a : int=None ): '''simple docstring''' if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) lowerCAmelCase__ : Optional[int] = False return x * self.a + self.b def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 16 ) -> Dict: from datasets import load_dataset from transformers import AutoTokenizer lowerCAmelCase__ : List[str] = AutoTokenizer.from_pretrained('bert-base-cased' ) lowerCAmelCase__ : int = {'train': 'tests/test_samples/MRPC/train.csv', 'validation': 'tests/test_samples/MRPC/dev.csv'} lowerCAmelCase__ : str = load_dataset('csv' , data_files=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Tuple = datasets['train'].unique('label' ) lowerCAmelCase__ : Union[str, Any] = {v: i for i, v in enumerate(SCREAMING_SNAKE_CASE_ )} def tokenize_function(SCREAMING_SNAKE_CASE_ ): # max_length=None => use the model max length (it's actually the default) lowerCAmelCase__ : Union[str, Any] = tokenizer( examples['sentence1'] , examples['sentence2'] , truncation=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , padding='max_length' ) if "label" in examples: lowerCAmelCase__ : Optional[Any] = [label_to_id[l] for l in examples['label']] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset lowerCAmelCase__ : Union[str, Any] = datasets.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , remove_columns=['sentence1', 'sentence2', 'label'] , ) def collate_fn(SCREAMING_SNAKE_CASE_ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(SCREAMING_SNAKE_CASE_ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. lowerCAmelCase__ : int = DataLoader(tokenized_datasets['train'] , shuffle=SCREAMING_SNAKE_CASE_ , collate_fn=SCREAMING_SNAKE_CASE_ , batch_size=2 ) lowerCAmelCase__ : List[Any] = DataLoader(tokenized_datasets['validation'] , shuffle=SCREAMING_SNAKE_CASE_ , collate_fn=SCREAMING_SNAKE_CASE_ , batch_size=1 ) return train_dataloader, eval_dataloader
716
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} # See all BART models at https://huggingface.co/models?filter=bart lowerCamelCase__ = { """vocab_file""": { """facebook/bart-base""": """https://huggingface.co/facebook/bart-base/resolve/main/vocab.json""", """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/vocab.json""", """facebook/bart-large-mnli""": """https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json""", """facebook/bart-large-cnn""": """https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json""", """facebook/bart-large-xsum""": """https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json""", """yjernite/bart_eli5""": """https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json""", }, """merges_file""": { """facebook/bart-base""": """https://huggingface.co/facebook/bart-base/resolve/main/merges.txt""", """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/merges.txt""", """facebook/bart-large-mnli""": """https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt""", """facebook/bart-large-cnn""": """https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt""", """facebook/bart-large-xsum""": """https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt""", """yjernite/bart_eli5""": """https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt""", }, """tokenizer_file""": { """facebook/bart-base""": """https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json""", """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json""", """facebook/bart-large-mnli""": """https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json""", """facebook/bart-large-cnn""": """https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json""", """facebook/bart-large-xsum""": """https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json""", """yjernite/bart_eli5""": """https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """facebook/bart-base""": 1024, """facebook/bart-large""": 1024, """facebook/bart-large-mnli""": 1024, """facebook/bart-large-cnn""": 1024, """facebook/bart-large-xsum""": 1024, """yjernite/bart_eli5""": 1024, } class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = ['input_ids', 'attention_mask'] lowercase = BartTokenizer def __init__( self : Optional[Any] , a : Union[str, Any]=None , a : List[str]=None , a : Union[str, Any]=None , a : Optional[Any]="replace" , a : int="<s>" , a : Optional[int]="</s>" , a : int="</s>" , a : str="<s>" , a : Tuple="<unk>" , a : Tuple="<pad>" , a : str="<mask>" , a : List[str]=False , a : Optional[Any]=True , **a : int , ): '''simple docstring''' super().__init__( a , a , tokenizer_file=a , errors=a , bos_token=a , eos_token=a , sep_token=a , cls_token=a , unk_token=a , pad_token=a , mask_token=a , add_prefix_space=a , trim_offsets=a , **a , ) lowerCAmelCase__ : Optional[int] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , a ) != add_prefix_space: lowerCAmelCase__ : Dict = getattr(a , pre_tok_state.pop('type' ) ) lowerCAmelCase__ : Union[str, Any] = add_prefix_space lowerCAmelCase__ : Tuple = pre_tok_class(**a ) lowerCAmelCase__ : Optional[Any] = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowerCAmelCase__ : List[str] = 'post_processor' lowerCAmelCase__ : List[str] = getattr(self.backend_tokenizer , a , a ) if tokenizer_component_instance: lowerCAmelCase__ : int = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCAmelCase__ : Union[str, Any] = tuple(state['sep'] ) if "cls" in state: lowerCAmelCase__ : Union[str, Any] = tuple(state['cls'] ) lowerCAmelCase__ : Any = False if state.get('add_prefix_space' , a ) != add_prefix_space: lowerCAmelCase__ : Optional[int] = add_prefix_space lowerCAmelCase__ : Optional[Any] = True if state.get('trim_offsets' , a ) != trim_offsets: lowerCAmelCase__ : str = trim_offsets lowerCAmelCase__ : Union[str, Any] = True if changes_to_apply: lowerCAmelCase__ : Dict = getattr(a , state.pop('type' ) ) lowerCAmelCase__ : int = component_class(**a ) setattr(self.backend_tokenizer , a , a ) @property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def _lowerCamelCase ( self : str , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else value lowerCAmelCase__ : Optional[Any] = value def _lowerCamelCase ( self : Any , *a : Optional[int] , **a : Dict ): '''simple docstring''' lowerCAmelCase__ : str = kwargs.get('is_split_into_words' , a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' 'to use it with pretokenized inputs.' ) return super()._batch_encode_plus(*a , **a ) def _lowerCamelCase ( self : Dict , *a : str , **a : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = kwargs.get('is_split_into_words' , a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' 'to use it with pretokenized inputs.' ) return super()._encode_plus(*a , **a ) def _lowerCamelCase ( self : str , a : str , a : Optional[str] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = self._tokenizer.model.save(a , name=a ) return tuple(a ) def _lowerCamelCase ( self : Any , a : Optional[int] , a : int=None ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : str = [self.sep_token_id] lowerCAmelCase__ : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
717
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
'''simple docstring''' import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: return EnvironmentCommand() class A__ ( __magic_name__ ): @staticmethod def _lowerCamelCase ( a : ArgumentParser ): '''simple docstring''' lowerCAmelCase__ : List[str] = parser.add_parser('env' ) download_parser.set_defaults(func=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : int = huggingface_hub.__version__ lowerCAmelCase__ : List[Any] = 'not installed' lowerCAmelCase__ : Optional[int] = 'NA' if is_torch_available(): import torch lowerCAmelCase__ : Tuple = torch.__version__ lowerCAmelCase__ : int = torch.cuda.is_available() lowerCAmelCase__ : Union[str, Any] = 'not installed' if is_transformers_available(): import transformers lowerCAmelCase__ : str = transformers.__version__ lowerCAmelCase__ : List[Any] = 'not installed' if is_accelerate_available(): import accelerate lowerCAmelCase__ : Optional[int] = accelerate.__version__ lowerCAmelCase__ : str = 'not installed' if is_xformers_available(): import xformers lowerCAmelCase__ : str = xformers.__version__ lowerCAmelCase__ : Tuple = { '`diffusers` version': version, 'Platform': platform.platform(), 'Python version': platform.python_version(), 'PyTorch version (GPU?)': f'''{pt_version} ({pt_cuda_available})''', 'Huggingface_hub version': hub_version, 'Transformers version': transformers_version, 'Accelerate version': accelerate_version, 'xFormers version': xformers_version, 'Using GPU in script?': '<fill in>', 'Using distributed or parallel set-up in script?': '<fill in>', } print('\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n' ) print(self.format_dict(a ) ) return info @staticmethod def _lowerCamelCase ( a : int ): '''simple docstring''' return "\n".join([f'''- {prop}: {val}''' for prop, val in d.items()] ) + "\n"
718
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
import numpy class A__ : def __init__( self : Tuple , a : numpy.ndarray , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : int = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Dict = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : List[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowerCAmelCase__ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : List[Any] = numpy.zeros(output_array.shape ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowerCAmelCase__ : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowerCAmelCase__ : int = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowerCamelCase ( self : Optional[int] , a : numpy.ndarray , a : int , a : bool ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): lowerCAmelCase__ : Any = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Tuple = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def _lowerCamelCase ( self : Optional[Any] , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : Dict = input_arr lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return (value) * (1 - (value)) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : List[str] = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ , output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ , iterations=10 , give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
719
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable lowerCamelCase__ = { """configuration_gpt_neox_japanese""": ["""GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoXJapaneseConfig"""], """tokenization_gpt_neox_japanese""": ["""GPTNeoXJapaneseTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTNeoXJapaneseForCausalLM""", """GPTNeoXJapaneseLayer""", """GPTNeoXJapaneseModel""", """GPTNeoXJapanesePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
720
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
0
import math def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0 , SCREAMING_SNAKE_CASE_ = 0 ) -> list: lowerCAmelCase__ : str = end or len(SCREAMING_SNAKE_CASE_ ) for i in range(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Optional[Any] = i lowerCAmelCase__ : int = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: lowerCAmelCase__ : str = array[temp_index - 1] temp_index -= 1 lowerCAmelCase__ : Tuple = temp_index_value return array def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> None: # Max Heap lowerCAmelCase__ : Any = index lowerCAmelCase__ : int = 2 * index + 1 # Left Node lowerCAmelCase__ : Optional[int] = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: lowerCAmelCase__ : List[str] = left_index if right_index < heap_size and array[largest] < array[right_index]: lowerCAmelCase__ : Union[str, Any] = right_index if largest != index: lowerCAmelCase__ : Dict = array[largest], array[index] heapify(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: lowerCAmelCase__ : int = len(SCREAMING_SNAKE_CASE_ ) for i in range(n // 2 , -1 , -1 ): heapify(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for i in range(n - 1 , 0 , -1 ): lowerCAmelCase__ : Tuple = array[0], array[i] heapify(SCREAMING_SNAKE_CASE_ , 0 , SCREAMING_SNAKE_CASE_ ) return array def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : List[Any] = low lowerCAmelCase__ : Dict = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i lowerCAmelCase__ : Optional[Any] = array[j], array[i] i += 1 def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: if len(SCREAMING_SNAKE_CASE_ ) == 0: return array lowerCAmelCase__ : int = 2 * math.ceil(math.loga(len(SCREAMING_SNAKE_CASE_ ) ) ) lowerCAmelCase__ : Union[str, Any] = 16 return intro_sort(SCREAMING_SNAKE_CASE_ , 0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> list: while end - start > size_threshold: if max_depth == 0: return heap_sort(SCREAMING_SNAKE_CASE_ ) max_depth -= 1 lowerCAmelCase__ : Optional[Any] = median_of_a(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , start + ((end - start) // 2) + 1 , end - 1 ) lowerCAmelCase__ : str = partition(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) intro_sort(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = p return insertion_sort(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase__ = input("""Enter numbers separated by a comma : """).strip() lowerCamelCase__ = [float(item) for item in user_input.split(""",""")] print(sort(unsorted))
721
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
69
0
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class A__ : def __init__( self : int , a : str = "cpu" , a : str = "openai/clip-vit-large-patch14" ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = device lowerCAmelCase__ : Any = CLIPTokenizerFast.from_pretrained(a ) lowerCAmelCase__ : Tuple = [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] lowerCAmelCase__ : Optional[Any] = [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] lowerCAmelCase__ : List[str] = torchvision.transforms.Normalize(self.image_mean , self.image_std ) lowerCAmelCase__ : str = torchvision.transforms.Resize(224 ) lowerCAmelCase__ : Union[str, Any] = torchvision.transforms.CenterCrop(224 ) def _lowerCamelCase ( self : int , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.resize(a ) lowerCAmelCase__ : Tuple = self.center_crop(a ) lowerCAmelCase__ : Any = self.normalize(a ) return images def __call__( self : Union[str, Any] , a : Optional[int]=None , a : Any=None , **a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.tokenizer(text=a , **a ) lowerCAmelCase__ : Optional[int] = self.preprocess_img(a ) lowerCAmelCase__ : Any = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class A__ ( nn.Module ): def __init__( self : Union[str, Any] , a : List[Any]=10 , a : Optional[int]=0.0_1 , a : List[Any]=None , a : Optional[int]=None , a : str=None , a : Optional[Any]=None , a : Union[str, Any]=None , a : List[Any]=None , a : Dict=False , a : Optional[int]=True , a : Optional[int]="image" , a : Tuple=True , a : int=False , a : Optional[int]=False , a : int=False , ): '''simple docstring''' super().__init__() lowerCAmelCase__ : str = None lowerCAmelCase__ : Optional[int] = device if device else get_device() if vqgan: lowerCAmelCase__ : Any = vqgan else: lowerCAmelCase__ : Dict = load_vqgan(self.device , conf_path=a , ckpt_path=a ) self.vqgan.eval() if clip: lowerCAmelCase__ : Dict = clip else: lowerCAmelCase__ : Optional[Any] = CLIPModel.from_pretrained('openai/clip-vit-base-patch32' ) self.clip.to(self.device ) lowerCAmelCase__ : Any = ProcessorGradientFlow(device=self.device ) lowerCAmelCase__ : Any = iterations lowerCAmelCase__ : Union[str, Any] = lr lowerCAmelCase__ : Tuple = log lowerCAmelCase__ : int = make_grid lowerCAmelCase__ : Optional[Any] = return_val lowerCAmelCase__ : List[Any] = quantize lowerCAmelCase__ : Optional[Any] = self.vqgan.decoder.z_shape def _lowerCamelCase ( self : str , a : List[Any]=None , a : List[Any]=None , a : int=5 , a : Dict=True ): '''simple docstring''' lowerCAmelCase__ : Any = [] if output_path is None: lowerCAmelCase__ : Optional[int] = './animation.gif' if input_path is None: lowerCAmelCase__ : List[str] = self.save_path lowerCAmelCase__ : Tuple = sorted(glob(input_path + '/*' ) ) if not len(a ): raise ValueError( 'No images found in save path, aborting (did you pass save_intermediate=True to the generate' ' function?)' ) if len(a ) == 1: print('Only one image found in save path, (did you pass save_intermediate=True to the generate function?)' ) lowerCAmelCase__ : List[Any] = total_duration / len(a ) lowerCAmelCase__ : Union[str, Any] = [frame_duration] * len(a ) if extend_frames: lowerCAmelCase__ : Union[str, Any] = 1.5 lowerCAmelCase__ : Union[str, Any] = 3 for file_name in paths: if file_name.endswith('.png' ): images.append(imageio.imread(a ) ) imageio.mimsave(a , a , duration=a ) print(f'''gif saved to {output_path}''' ) def _lowerCamelCase ( self : str , a : Dict=None , a : List[str]=None ): '''simple docstring''' if not (path or img): raise ValueError('Input either path or tensor' ) if img is not None: raise NotImplementedError lowerCAmelCase__ : Any = preprocess(Image.open(a ) , target_image_size=256 ).to(self.device ) lowerCAmelCase__ : str = preprocess_vqgan(a ) lowerCAmelCase__ : List[Any] = self.vqgan.encode(a ) return z def _lowerCamelCase ( self : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.latent.detach().requires_grad_() lowerCAmelCase__ : Optional[int] = base_latent + transform_vector if self.quantize: lowerCAmelCase__ : Dict = self.vqgan.quantize(a ) else: lowerCAmelCase__ : int = trans_latent return self.vqgan.decode(a ) def _lowerCamelCase ( self : Dict , a : Any , a : Any , a : List[Any]=None ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.clip_preprocessor(text=a , images=a , return_tensors='pt' , padding=a ) lowerCAmelCase__ : int = self.clip(**a ) lowerCAmelCase__ : Optional[int] = clip_outputs.logits_per_image if weights is not None: lowerCAmelCase__ : List[str] = similarity_logits * weights return similarity_logits.sum() def _lowerCamelCase ( self : List[Any] , a : Optional[Any] , a : Any , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self._get_clip_similarity(pos_prompts['prompts'] , a , weights=(1 / pos_prompts['weights']) ) if neg_prompts: lowerCAmelCase__ : Optional[Any] = self._get_clip_similarity(neg_prompts['prompts'] , a , weights=neg_prompts['weights'] ) else: lowerCAmelCase__ : str = torch.tensor([1] , device=self.device ) lowerCAmelCase__ : Dict = -torch.log(a ) + torch.log(a ) return loss def _lowerCamelCase ( self : Union[str, Any] , a : str , a : Dict , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = torch.randn_like(self.latent , requires_grad=a , device=self.device ) lowerCAmelCase__ : Tuple = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() lowerCAmelCase__ : Optional[Any] = self._add_vector(a ) lowerCAmelCase__ : List[Any] = loop_post_process(a ) lowerCAmelCase__ : Union[str, Any] = self._get_CLIP_loss(a , a , a ) print('CLIP loss' , a ) if self.log: wandb.log({'CLIP Loss': clip_loss} ) clip_loss.backward(retain_graph=a ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def _lowerCamelCase ( self : Optional[int] , a : Dict , a : List[Any] , a : Union[str, Any] ): '''simple docstring''' wandb.init(reinit=a , project='face-editor' ) wandb.config.update({'Positive Prompts': positive_prompts} ) wandb.config.update({'Negative Prompts': negative_prompts} ) wandb.config.update({'lr': self.lr, 'iterations': self.iterations} ) if image_path: lowerCAmelCase__ : Optional[Any] = Image.open(a ) lowerCAmelCase__ : Dict = image.resize((256, 256) ) wandb.log('Original Image' , wandb.Image(a ) ) def _lowerCamelCase ( self : Union[str, Any] , a : List[Any] ): '''simple docstring''' if not prompts: return [] lowerCAmelCase__ : Tuple = [] lowerCAmelCase__ : int = [] if isinstance(a , a ): lowerCAmelCase__ : Tuple = [prompt.strip() for prompt in prompts.split('|' )] for prompt in prompts: if isinstance(a , (tuple, list) ): lowerCAmelCase__ : Any = prompt[0] lowerCAmelCase__ : List[str] = float(prompt[1] ) elif ":" in prompt: lowerCAmelCase__ : Union[str, Any] = prompt.split(':' ) lowerCAmelCase__ : Optional[Any] = float(a ) else: lowerCAmelCase__ : Optional[int] = prompt lowerCAmelCase__ : Optional[Any] = 1.0 processed_prompts.append(a ) weights.append(a ) return { "prompts": processed_prompts, "weights": torch.tensor(a , device=self.device ), } def _lowerCamelCase ( self : Tuple , a : Tuple , a : Any=None , a : List[Any]=None , a : Union[str, Any]=True , a : int=False , a : Optional[Any]=True , a : Optional[Any]=True , a : Any=None , ): '''simple docstring''' if image_path: lowerCAmelCase__ : Dict = self._get_latent(a ) else: lowerCAmelCase__ : Tuple = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(a , a , a ) assert pos_prompts, "You must provide at least one positive prompt." lowerCAmelCase__ : Union[str, Any] = self.process_prompts(a ) lowerCAmelCase__ : Optional[Any] = self.process_prompts(a ) if save_final and save_path is None: lowerCAmelCase__ : Optional[int] = os.path.join('./outputs/' , '_'.join(pos_prompts['prompts'] ) ) if not os.path.exists(a ): os.makedirs(a ) else: lowerCAmelCase__ : Optional[int] = save_path + '_' + get_timestamp() os.makedirs(a ) lowerCAmelCase__ : Tuple = save_path lowerCAmelCase__ : str = self.vqgan.decode(self.latent )[0] if show_intermediate: print('Original Image' ) show_pil(custom_to_pil(a ) ) lowerCAmelCase__ : Tuple = loop_post_process(a ) for iter, transformed_img in enumerate(self._optimize_CLIP(a , a , a ) ): if show_intermediate: show_pil(a ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , f'''iter_{iter:03d}.png''' ) ) if self.log: wandb.log({'Image': wandb.Image(a )} ) if show_final: show_pil(a ) if save_final: transformed_img.save(os.path.join(self.save_path , f'''iter_{iter:03d}_final.png''' ) )
700
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __magic_name__ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'LayoutLMv3ImageProcessor' lowercase = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self : Optional[int] , a : Union[str, Any]=None , a : Optional[Any]=None , **a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , a , ) lowerCAmelCase__ : int = kwargs.pop('feature_extractor' ) lowerCAmelCase__ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(a , a ) def __call__( self : List[Any] , a : List[Any] , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a : Union[List[List[int]], List[List[List[int]]]] = None , a : Optional[Union[List[int], List[List[int]]]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : str , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor lowerCAmelCase__ : List[str] = self.image_processor(images=a , return_tensors=a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a , a ): lowerCAmelCase__ : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ : List[str] = features['words'] lowerCAmelCase__ : List[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_token_type_ids=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_length=a , verbose=a , return_tensors=a , **a , ) # add pixel values lowerCAmelCase__ : Tuple = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase__ : List[str] = self.get_overflowing_images(a , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase__ : List[str] = images return encoded_inputs def _lowerCamelCase ( self : Any , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a ) != len(a ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' f''' {len(a )} and {len(a )}''' ) return images_with_overflow def _lowerCamelCase ( self : Union[str, Any] , *a : Optional[Any] , **a : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Tuple , *a : List[str] , **a : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*a , **a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , a , ) return self.image_processor_class @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , a , ) return self.image_processor
69
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionAttendAndExcitePipeline, UNetaDConditionModel, ) from diffusers.utils import load_numpy, skip_mps, slow from diffusers.utils.testing_utils import require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin lowerCamelCase__ = False @skip_mps class A__ ( __magic_name__ , __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = StableDiffusionAttendAndExcitePipeline lowercase = False lowercase = TEXT_TO_IMAGE_PARAMS lowercase = TEXT_TO_IMAGE_BATCH_PARAMS.union({'token_indices'} ) lowercase = TEXT_TO_IMAGE_IMAGE_PARAMS lowercase = TEXT_TO_IMAGE_IMAGE_PARAMS @classmethod def _lowerCamelCase ( cls : List[Any] ): '''simple docstring''' super().setUpClass() torch.use_deterministic_algorithms(a ) @classmethod def _lowerCamelCase ( cls : List[str] ): '''simple docstring''' super().tearDownClass() torch.use_deterministic_algorithms(a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=a , ) lowerCAmelCase__ : List[Any] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='scaled_linear' , clip_sample=a , set_alpha_to_one=a , ) torch.manual_seed(0 ) lowerCAmelCase__ : Any = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) lowerCAmelCase__ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='gelu' , projection_dim=512 , ) lowerCAmelCase__ : Dict = CLIPTextModel(a ) lowerCAmelCase__ : Optional[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) lowerCAmelCase__ : str = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def _lowerCamelCase ( self : Optional[Any] , a : int , a : Tuple=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : Any = torch.manual_seed(a ) else: lowerCAmelCase__ : Optional[int] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Dict = { 'prompt': 'a cat and a frog', 'token_indices': [2, 5], 'generator': generator, 'num_inference_steps': 1, 'guidance_scale': 6.0, 'output_type': 'numpy', 'max_iter_to_alter': 2, 'thresholds': {0: 0.7}, } return inputs def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = 'cpu' lowerCAmelCase__ : str = self.get_dummy_components() lowerCAmelCase__ : Any = self.pipeline_class(**a ) pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : List[Any] = pipe(**a ).images lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 64, 64, 3) ) lowerCAmelCase__ : Dict = np.array( [0.6_3_9_0_5_3_6_4, 0.6_2_8_9_7_3_0_7, 0.4_8_5_9_9_0_1_7, 0.5_1_3_3_6_2_4, 0.5_5_5_0_0_4_8, 0.4_5_7_6_9_5_1_6, 0.5_0_3_2_6_9_7_3, 0.5_0_2_3_1_3_9, 0.4_5_3_8_4_4_9_6] ) lowerCAmelCase__ : int = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(a , 1E-3 ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' super().test_cpu_offload_forward_pass(expected_max_diff=5E-4 ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self._test_inference_batch_single_identical(batch_size=2 , expected_max_diff=7E-4 ) def _lowerCamelCase ( self : str ): '''simple docstring''' super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' super().test_pt_np_pil_outputs_equivalent(expected_max_diff=5E-4 ) def _lowerCamelCase ( self : int ): '''simple docstring''' super().test_save_load_local(expected_max_difference=5E-4 ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().test_save_load_optional_components(expected_max_difference=4E-4 ) @require_torch_gpu @slow class A__ ( unittest.TestCase ): @classmethod def _lowerCamelCase ( cls : Any ): '''simple docstring''' super().setUpClass() torch.use_deterministic_algorithms(a ) @classmethod def _lowerCamelCase ( cls : Union[str, Any] ): '''simple docstring''' super().tearDownClass() torch.use_deterministic_algorithms(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = torch.manual_seed(51 ) lowerCAmelCase__ : Dict = StableDiffusionAttendAndExcitePipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , safety_checker=a , torch_dtype=torch.floataa ) pipe.to('cuda' ) lowerCAmelCase__ : Tuple = 'a painting of an elephant with glasses' lowerCAmelCase__ : Union[str, Any] = [5, 7] lowerCAmelCase__ : List[Any] = pipe( prompt=a , token_indices=a , guidance_scale=7.5 , generator=a , num_inference_steps=5 , max_iter_to_alter=5 , output_type='numpy' , ).images[0] lowerCAmelCase__ : Optional[int] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/attend-and-excite/elephant_glasses.npy' ) assert np.abs((expected_image - image).max() ) < 5E-1
701
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
69
0
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class A__ : @staticmethod def _lowerCamelCase ( *a : int , **a : List[str] ): '''simple docstring''' pass @is_pipeline_test @require_vision class A__ ( unittest.TestCase ): @require_torch def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = pipeline( model='hf-internal-testing/tiny-random-clip-zero-shot-image-classification' , ) lowerCAmelCase__ : Union[str, Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Optional[int] = image_classifier(a , candidate_labels=['a', 'b', 'c'] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(a ) , [ [{'score': 0.3_3_3, 'label': 'a'}, {'score': 0.3_3_3, 'label': 'b'}, {'score': 0.3_3_3, 'label': 'c'}], [{'score': 0.3_3_3, 'label': 'a'}, {'score': 0.3_3_3, 'label': 'c'}, {'score': 0.3_3_3, 'label': 'b'}], ] , ) lowerCAmelCase__ : str = image_classifier([image] * 5 , candidate_labels=['A', 'B', 'C'] , batch_size=2 ) self.assertEqual( nested_simplify(a ) , [ [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], ] , ) @require_tf def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = pipeline( model='hf-internal-testing/tiny-random-clip-zero-shot-image-classification' , framework='tf' ) lowerCAmelCase__ : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : List[Any] = image_classifier(a , candidate_labels=['a', 'b', 'c'] ) self.assertEqual( nested_simplify(a ) , [{'score': 0.3_3_3, 'label': 'a'}, {'score': 0.3_3_3, 'label': 'b'}, {'score': 0.3_3_3, 'label': 'c'}] , ) lowerCAmelCase__ : Optional[int] = image_classifier([image] * 5 , candidate_labels=['A', 'B', 'C'] , batch_size=2 ) self.assertEqual( nested_simplify(a ) , [ [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], [ {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, {'score': 0.3_3_3, 'label': ANY(a )}, ], ] , ) @slow @require_torch def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = pipeline( task='zero-shot-image-classification' , model='openai/clip-vit-base-patch32' , ) # This is an image of 2 cats with remotes and no planes lowerCAmelCase__ : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Optional[Any] = image_classifier(a , candidate_labels=['cat', 'plane', 'remote'] ) self.assertEqual( nested_simplify(a ) , [ {'score': 0.5_1_1, 'label': 'remote'}, {'score': 0.4_8_5, 'label': 'cat'}, {'score': 0.0_0_4, 'label': 'plane'}, ] , ) lowerCAmelCase__ : Dict = image_classifier([image] * 5 , candidate_labels=['cat', 'plane', 'remote'] , batch_size=2 ) self.assertEqual( nested_simplify(a ) , [ [ {'score': 0.5_1_1, 'label': 'remote'}, {'score': 0.4_8_5, 'label': 'cat'}, {'score': 0.0_0_4, 'label': 'plane'}, ], ] * 5 , ) @slow @require_tf def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = pipeline( task='zero-shot-image-classification' , model='openai/clip-vit-base-patch32' , framework='tf' ) # This is an image of 2 cats with remotes and no planes lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Optional[Any] = image_classifier(a , candidate_labels=['cat', 'plane', 'remote'] ) self.assertEqual( nested_simplify(a ) , [ {'score': 0.5_1_1, 'label': 'remote'}, {'score': 0.4_8_5, 'label': 'cat'}, {'score': 0.0_0_4, 'label': 'plane'}, ] , ) lowerCAmelCase__ : Any = image_classifier([image] * 5 , candidate_labels=['cat', 'plane', 'remote'] , batch_size=2 ) self.assertEqual( nested_simplify(a ) , [ [ {'score': 0.5_1_1, 'label': 'remote'}, {'score': 0.4_8_5, 'label': 'cat'}, {'score': 0.0_0_4, 'label': 'plane'}, ], ] * 5 , )
702
lowerCamelCase__ = """Alexander Joslin""" import operator as op from .stack import Stack def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} lowerCAmelCase__ : Stack[int] = Stack() lowerCAmelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE_ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE_ ) elif i == ")": # RULE 4 lowerCAmelCase__ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : List[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : Tuple = operators[opr](SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) operand_stack.push(SCREAMING_SNAKE_CASE_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
69
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase__ = { """configuration_graphormer""": ["""GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GraphormerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """GraphormerForGraphClassification""", """GraphormerModel""", """GraphormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
703
import numpy class A__ : def __init__( self : Tuple , a : numpy.ndarray , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : int = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Dict = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : List[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowerCAmelCase__ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : List[Any] = numpy.zeros(output_array.shape ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowerCAmelCase__ : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowerCAmelCase__ : int = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowerCamelCase ( self : Optional[int] , a : numpy.ndarray , a : int , a : bool ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): lowerCAmelCase__ : Any = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Tuple = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def _lowerCamelCase ( self : Optional[Any] , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : Dict = input_arr lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return (value) * (1 - (value)) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : List[str] = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ , output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ , iterations=10 , give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
69
0
import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : str = nn.functional.normalize(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[int] = nn.functional.normalize(SCREAMING_SNAKE_CASE_ ) return torch.mm(SCREAMING_SNAKE_CASE_ , normalized_text_embeds.t() ) class A__ ( __magic_name__ ): lowercase = CLIPConfig lowercase = ['CLIPEncoderLayer'] def __init__( self : Any , a : CLIPConfig ): '''simple docstring''' super().__init__(a ) lowerCAmelCase__ : Optional[int] = CLIPVisionModel(config.vision_config ) lowerCAmelCase__ : Tuple = nn.Linear(config.vision_config.hidden_size , config.projection_dim , bias=a ) lowerCAmelCase__ : List[str] = nn.Parameter(torch.ones(17 , config.projection_dim ) , requires_grad=a ) lowerCAmelCase__ : List[str] = nn.Parameter(torch.ones(3 , config.projection_dim ) , requires_grad=a ) lowerCAmelCase__ : Dict = nn.Parameter(torch.ones(17 ) , requires_grad=a ) lowerCAmelCase__ : List[str] = nn.Parameter(torch.ones(3 ) , requires_grad=a ) @torch.no_grad() def _lowerCamelCase ( self : int , a : int , a : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.vision_model(a )[1] # pooled_output lowerCAmelCase__ : Optional[int] = self.visual_projection(a ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowerCAmelCase__ : Any = cosine_distance(a , self.special_care_embeds ).cpu().float().numpy() lowerCAmelCase__ : Any = cosine_distance(a , self.concept_embeds ).cpu().float().numpy() lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : Optional[int] = image_embeds.shape[0] for i in range(a ): lowerCAmelCase__ : List[str] = {'special_scores': {}, 'special_care': [], 'concept_scores': {}, 'bad_concepts': []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images lowerCAmelCase__ : str = 0.0 for concept_idx in range(len(special_cos_dist[0] ) ): lowerCAmelCase__ : int = special_cos_dist[i][concept_idx] lowerCAmelCase__ : str = self.special_care_embeds_weights[concept_idx].item() lowerCAmelCase__ : Optional[Any] = round(concept_cos - concept_threshold + adjustment , 3 ) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img['special_scores'][concept_idx]} ) lowerCAmelCase__ : str = 0.0_1 for concept_idx in range(len(cos_dist[0] ) ): lowerCAmelCase__ : Optional[int] = cos_dist[i][concept_idx] lowerCAmelCase__ : List[Any] = self.concept_embeds_weights[concept_idx].item() lowerCAmelCase__ : List[str] = round(concept_cos - concept_threshold + adjustment , 3 ) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(a ) result.append(a ) lowerCAmelCase__ : Optional[int] = [len(res['bad_concepts'] ) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def _lowerCamelCase ( self : Optional[Any] , a : torch.FloatTensor , a : torch.FloatTensor ): '''simple docstring''' lowerCAmelCase__ : Dict = self.vision_model(a )[1] # pooled_output lowerCAmelCase__ : Optional[Any] = self.visual_projection(a ) lowerCAmelCase__ : Tuple = cosine_distance(a , self.special_care_embeds ) lowerCAmelCase__ : Optional[Any] = cosine_distance(a , self.concept_embeds ) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images lowerCAmelCase__ : Union[str, Any] = 0.0 lowerCAmelCase__ : Optional[Any] = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) lowerCAmelCase__ : str = torch.any(special_scores > 0 , dim=1 ) lowerCAmelCase__ : List[str] = special_care * 0.0_1 lowerCAmelCase__ : Tuple = special_adjustment.unsqueeze(1 ).expand(-1 , cos_dist.shape[1] ) lowerCAmelCase__ : int = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) lowerCAmelCase__ : Tuple = torch.any(concept_scores > 0 , dim=1 ) return images, has_nsfw_concepts
704
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A__ : def __init__( self : int , a : str , a : Union[str, Any]=13 , a : int=32 , a : Optional[Any]=2 , a : Tuple=3 , a : List[Any]=16 , a : List[str]=[1, 2, 1] , a : int=[2, 2, 4] , a : int=2 , a : Optional[Any]=2.0 , a : Optional[int]=True , a : Dict=0.0 , a : Any=0.0 , a : int=0.1 , a : List[str]="gelu" , a : Optional[Any]=False , a : str=True , a : Dict=0.0_2 , a : Any=1E-5 , a : Optional[int]=True , a : str=None , a : str=True , a : int=10 , a : str=8 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : Optional[Any] = patch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = embed_dim lowerCAmelCase__ : Tuple = depths lowerCAmelCase__ : List[str] = num_heads lowerCAmelCase__ : List[Any] = window_size lowerCAmelCase__ : Any = mlp_ratio lowerCAmelCase__ : Optional[Any] = qkv_bias lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : int = drop_path_rate lowerCAmelCase__ : Optional[Any] = hidden_act lowerCAmelCase__ : int = use_absolute_embeddings lowerCAmelCase__ : List[str] = patch_norm lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Dict = use_labels lowerCAmelCase__ : List[Any] = type_sequence_label_size lowerCAmelCase__ : Optional[Any] = encoder_stride def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = None if self.use_labels: lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : int = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : List[str] , a : Any , a : str , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = SwinvaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model(a ) lowerCAmelCase__ : str = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase__ : Tuple = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SwinvaForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : Dict = SwinvaForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Union[str, Any] , a : int , a : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.type_sequence_label_size lowerCAmelCase__ : List[Any] = SwinvaForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = config_and_inputs lowerCAmelCase__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = SwinvaModelTester(self ) lowerCAmelCase__ : int = ConfigTester(self , config_class=a , embed_dim=37 ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[int] = model_class(a ) lowerCAmelCase__ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Tuple = [*signature.parameters.keys()] lowerCAmelCase__ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = True for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = True lowerCAmelCase__ : str = False lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Dict = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : int = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Dict = outputs.attentions lowerCAmelCase__ : Dict = len(self.model_tester.depths ) self.assertEqual(len(a ) , a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Optional[int] = config.window_size**2 lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowerCAmelCase__ : Tuple = len(a ) # Check attention is always last and order is fine lowerCAmelCase__ : str = True lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(**self._prepare_for_class(a , a ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): lowerCAmelCase__ : Optional[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase__ : Any = 2 self.assertEqual(out_len + added_hidden_states , len(a ) ) lowerCAmelCase__ : Dict = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _lowerCamelCase ( self : int , a : Optional[int] , a : int , a : Optional[Any] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.hidden_states lowerCAmelCase__ : str = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a ) , a ) # Swinv2 has a different seq_length lowerCAmelCase__ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowerCAmelCase__ : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(a ) , a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = reshaped_hidden_states[0].shape lowerCAmelCase__ : List[str] = ( reshaped_hidden_states[0].view(a , a , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : str = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : List[str] = True self.check_hidden_states_output(a , a , a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Any = 3 lowerCAmelCase__ : int = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase__ : str = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : str = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase__ : Any = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase__ : str = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[str] = SwinvaModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = _config_zero_init(a ) for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(config=a ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( a ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Any = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**a ) # verify the logits lowerCAmelCase__ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Optional[Any] = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) )
69
0
import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Optional[int] , a : Union[str, Any]=13 , a : List[Any]=7 , a : Optional[int]=True , a : Dict=True , a : int=True , a : Tuple=True , a : Any=99 , a : Tuple=32 , a : List[str]=5 , a : Dict=4 , a : Dict=37 , a : Union[str, Any]="gelu" , a : Union[str, Any]=0.1 , a : Tuple=0.1 , a : Optional[Any]=512 , a : Optional[Any]=16 , a : Tuple=2 , a : List[Any]=0.0_2 , a : Optional[int]=False , a : int=True , a : Union[str, Any]="None" , a : List[Any]=3 , a : int=4 , a : List[str]=None , ): '''simple docstring''' lowerCAmelCase__ : Tuple = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : int = seq_length lowerCAmelCase__ : str = is_training lowerCAmelCase__ : Union[str, Any] = use_input_mask lowerCAmelCase__ : int = use_token_type_ids lowerCAmelCase__ : Optional[int] = use_labels lowerCAmelCase__ : List[str] = vocab_size lowerCAmelCase__ : Dict = hidden_size lowerCAmelCase__ : Optional[int] = num_hidden_layers lowerCAmelCase__ : Optional[int] = num_attention_heads lowerCAmelCase__ : Optional[Any] = intermediate_size lowerCAmelCase__ : List[str] = hidden_act lowerCAmelCase__ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : Any = max_position_embeddings lowerCAmelCase__ : int = type_vocab_size lowerCAmelCase__ : Optional[int] = type_sequence_label_size lowerCAmelCase__ : int = initializer_range lowerCAmelCase__ : Any = num_labels lowerCAmelCase__ : Optional[Any] = num_choices lowerCAmelCase__ : str = relative_attention lowerCAmelCase__ : int = position_biased_input lowerCAmelCase__ : List[str] = pos_att_type lowerCAmelCase__ : Tuple = scope def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : List[Any] = None if self.use_input_mask: lowerCAmelCase__ : int = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) lowerCAmelCase__ : Union[str, Any] = None if self.use_token_type_ids: lowerCAmelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Dict = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : List[Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = self.get_config() lowerCAmelCase__ : List[str] = 300 return config def _lowerCamelCase ( self : str , a : Tuple ): '''simple docstring''' self.parent.assertListEqual(list(result.loss.size() ) , [] ) def _lowerCamelCase ( self : List[str] , a : Optional[Any] , a : str , a : int , a : Tuple , a : Any , a : Tuple , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = DebertaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a , attention_mask=a , token_type_ids=a )[0] lowerCAmelCase__ : Optional[Any] = model(a , token_type_ids=a )[0] lowerCAmelCase__ : Union[str, Any] = model(a )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def _lowerCamelCase ( self : str , a : Dict , a : int , a : List[str] , a : Tuple , a : str , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = DebertaForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , attention_mask=a , token_type_ids=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : List[Any] , a : Optional[Any] , a : List[str] , a : Tuple , a : List[str] , a : Any , a : Any , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Optional[Any] = DebertaForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , attention_mask=a , token_type_ids=a , labels=a ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(a ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Any , a : List[Any] , a : Any , a : Union[str, Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.num_labels lowerCAmelCase__ : str = DebertaForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , token_type_ids=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : List[str] , a : List[Any] , a : List[str] , a : Dict , a : Tuple , a : Optional[int] , a : List[Any] , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = DebertaForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Tuple = model( a , attention_mask=a , token_type_ids=a , start_positions=a , end_positions=a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : str = self.prepare_config_and_inputs() ( lowerCAmelCase__ ) : List[str] = config_and_inputs lowerCAmelCase__ : str = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) lowercase = ( { 'feature-extraction': DebertaModel, 'fill-mask': DebertaForMaskedLM, 'question-answering': DebertaForQuestionAnswering, 'text-classification': DebertaForSequenceClassification, 'token-classification': DebertaForTokenClassification, 'zero-shot': DebertaForSequenceClassification, } if is_torch_available() else {} ) lowercase = True lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : str = DebertaModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , hidden_size=37 ) def _lowerCamelCase ( self : str ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*a ) @slow def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Tuple = DebertaModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_torch @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase ): @unittest.skip(reason='Model not available yet' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' pass @slow def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = DebertaModel.from_pretrained('microsoft/deberta-base' ) lowerCAmelCase__ : Dict = torch.tensor([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) lowerCAmelCase__ : Union[str, Any] = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): lowerCAmelCase__ : Tuple = model(a , attention_mask=a )[0] # compare the actual values for a slice. lowerCAmelCase__ : int = torch.tensor( [[[-0.5_9_8_6, -0.8_0_5_5, -0.8_4_6_2], [1.4_4_8_4, -0.9_3_4_8, -0.8_0_5_9], [0.3_1_2_3, 0.0_0_3_2, -1.4_1_3_1]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , a , atol=1E-4 ) , f'''{output[:, 1:4, 1:4]}''' )
705
from itertools import permutations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase__ : str = [7, 11, 13, 17] for i, test in enumerate(SCREAMING_SNAKE_CASE_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 10 ) -> int: return sum( int(''.join(map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) ) for num in permutations(range(SCREAMING_SNAKE_CASE_ ) ) if is_substring_divisible(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: lowerCAmelCase__ : Any = [0] * len(SCREAMING_SNAKE_CASE_ ) for i in range(1 , len(SCREAMING_SNAKE_CASE_ ) ): # use last results for better performance - dynamic programming lowerCAmelCase__ : Optional[int] = prefix_result[i - 1] while j > 0 and input_string[i] != input_string[j]: lowerCAmelCase__ : Union[str, Any] = prefix_result[j - 1] if input_string[i] == input_string[j]: j += 1 lowerCAmelCase__ : Any = j return prefix_result def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: return max(prefix_function(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": import doctest doctest.testmod()
706
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
69
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input lowerCamelCase__ = """Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine""" def lowerCAmelCase__ ( ) -> Optional[int]: lowerCAmelCase__ : Dict = _ask_options( 'In which compute environment are you running?' , ['This machine', 'AWS (Amazon SageMaker)'] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: lowerCAmelCase__ : Dict = get_sagemaker_input() else: lowerCAmelCase__ : Tuple = get_cluster_input() return config def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_=None ) -> Optional[int]: if subparsers is not None: lowerCAmelCase__ : Any = subparsers.add_parser('config' , description=SCREAMING_SNAKE_CASE_ ) else: lowerCAmelCase__ : Optional[Any] = argparse.ArgumentParser('Accelerate config command' , description=SCREAMING_SNAKE_CASE_ ) parser.add_argument( '--config_file' , default=SCREAMING_SNAKE_CASE_ , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=SCREAMING_SNAKE_CASE_ ) return parser def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : List[Any] = get_user_input() if args.config_file is not None: lowerCAmelCase__ : Union[str, Any] = args.config_file else: if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): os.makedirs(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = default_yaml_config_file if config_file.endswith('.json' ): config.to_json_file(SCREAMING_SNAKE_CASE_ ) else: config.to_yaml_file(SCREAMING_SNAKE_CASE_ ) print(F'''accelerate configuration saved at {config_file}''' ) def lowerCAmelCase__ ( ) -> Union[str, Any]: lowerCAmelCase__ : Optional[Any] = config_command_parser() lowerCAmelCase__ : Tuple = parser.parse_args() config_command(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": main()
707
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ , lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
69
0
from __future__ import annotations import string from itertools import cycle, product from pathlib import Path lowerCamelCase__ = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) lowerCamelCase__ = [ord(letter) for letter in string.ascii_lowercase] lowerCamelCase__ = {ord(char) for char in VALID_CHARS} lowerCamelCase__ = ["""the""", """be""", """to""", """of""", """and""", """in""", """that""", """have"""] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str | None: lowerCAmelCase__ : str = "" lowerCAmelCase__ : int lowerCAmelCase__ : int lowerCAmelCase__ : int for keychar, cipherchar in zip(cycle(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Dict = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(SCREAMING_SNAKE_CASE_ ) return decoded def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list[str]: lowerCAmelCase__ : list[str] = [] for key in product(SCREAMING_SNAKE_CASE_ , repeat=3 ): lowerCAmelCase__ : Tuple = try_key(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if encoded is not None: possibles.append(SCREAMING_SNAKE_CASE_ ) return possibles def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = "p059_cipher.txt" ) -> int: lowerCAmelCase__ : list[int] lowerCAmelCase__ : list[str] lowerCAmelCase__ : str lowerCAmelCase__ : str lowerCAmelCase__ : str = Path(SCREAMING_SNAKE_CASE_ ).parent.joinpath(SCREAMING_SNAKE_CASE_ ).read_text(encoding='utf-8' ) lowerCAmelCase__ : int = [int(SCREAMING_SNAKE_CASE_ ) for number in data.strip().split(',' )] lowerCAmelCase__ : str = filter_valid_chars(SCREAMING_SNAKE_CASE_ ) for common_word in COMMON_WORDS: lowerCAmelCase__ : Union[str, Any] = filter_common_word(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) == 1: break lowerCAmelCase__ : List[str] = possibles[0] return sum(ord(SCREAMING_SNAKE_CASE_ ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
708
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
0
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } lowerCamelCase__ = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } lowerCamelCase__ = { """facebook/blenderbot_small-90M""": 512, } class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = BlenderbotSmallTokenizer def __init__( self : Union[str, Any] , a : Tuple=None , a : Optional[Any]=None , a : List[str]="<|endoftext|>" , a : Optional[int]="<|endoftext|>" , a : List[str]="<|endoftext|>" , a : List[str]=False , a : Union[str, Any]=True , **a : Any , ): '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=a , merges=a , add_prefix_space=a , trim_offsets=a , ) , bos_token=a , eos_token=a , unk_token=a , **a , ) lowerCAmelCase__ : Any = add_prefix_space def _lowerCamelCase ( self : Any , a : Dict , a : Union[str, Any]=None ): '''simple docstring''' lowerCAmelCase__ : List[Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowerCamelCase ( self : Any , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : List[Any] = [self.sep_token_id] lowerCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
709
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """google/rembert""": 256, } lowerCamelCase__ = """▁""" class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = RemBertTokenizer def __init__( self : Optional[Any] , a : str=None , a : Any=None , a : List[Any]=True , a : str=True , a : Dict=False , a : Dict="[CLS]" , a : int="[SEP]" , a : Tuple="<unk>" , a : Optional[Any]="[SEP]" , a : Tuple="<pad>" , a : Dict="[CLS]" , a : Optional[Any]="[MASK]" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( a , tokenizer_file=a , do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : int = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : List[Any] = keep_accents lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
69
0
import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 lowerCamelCase__ = get_tests_dir("""fixtures/dummy_feature_extractor_config.json""") lowerCamelCase__ = get_tests_dir("""fixtures/vocab.json""") lowerCamelCase__ = get_tests_dir("""fixtures""") class A__ ( unittest.TestCase ): lowercase = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = 0 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) self.assertIsInstance(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ : Optional[Any] = WavaVecaConfig() lowerCAmelCase__ : List[str] = AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) # save in new folder model_config.save_pretrained(a ) processor.save_pretrained(a ) lowerCAmelCase__ : Tuple = AutoProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(a , os.path.join(a , a ) ) copyfile(a , os.path.join(a , 'vocab.json' ) ) lowerCAmelCase__ : Tuple = AutoProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ : Dict = WavaVecaFeatureExtractor() lowerCAmelCase__ : List[Any] = AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) lowerCAmelCase__ : Dict = WavaVecaProcessor(a , a ) # save in new folder processor.save_pretrained(a ) # drop `processor_class` in tokenizer with open(os.path.join(a , a ) , 'r' ) as f: lowerCAmelCase__ : List[str] = json.load(a ) config_dict.pop('processor_class' ) with open(os.path.join(a , a ) , 'w' ) as f: f.write(json.dumps(a ) ) lowerCAmelCase__ : int = AutoProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _lowerCamelCase ( self : str ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ : Optional[Any] = WavaVecaFeatureExtractor() lowerCAmelCase__ : Dict = AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) lowerCAmelCase__ : List[str] = WavaVecaProcessor(a , a ) # save in new folder processor.save_pretrained(a ) # drop `processor_class` in feature extractor with open(os.path.join(a , a ) , 'r' ) as f: lowerCAmelCase__ : Optional[Any] = json.load(a ) config_dict.pop('processor_class' ) with open(os.path.join(a , a ) , 'w' ) as f: f.write(json.dumps(a ) ) lowerCAmelCase__ : Optional[Any] = AutoProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase__ : Any = WavaVecaConfig(processor_class='Wav2Vec2Processor' ) model_config.save_pretrained(a ) # copy relevant files copyfile(a , os.path.join(a , 'vocab.json' ) ) # create emtpy sample processor with open(os.path.join(a , a ) , 'w' ) as f: f.write('{}' ) lowerCAmelCase__ : Tuple = AutoProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' with self.assertRaises(a ): lowerCAmelCase__ : Dict = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowerCAmelCase__ : Dict = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=a ) lowerCAmelCase__ : Any = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' , trust_remote_code=a ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) lowerCAmelCase__ : Dict = processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) lowerCAmelCase__ : Optional[int] = processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) # Test we can also load the slow version lowerCAmelCase__ : Any = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=a , use_fast=a ) lowerCAmelCase__ : List[str] = new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , 'NewTokenizer' ) else: self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' try: AutoConfig.register('custom' , a ) AutoFeatureExtractor.register(a , a ) AutoTokenizer.register(a , slow_tokenizer_class=a ) AutoProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoProcessor.register(a , a ) # Now that the config is registered, it can be used as any other config with the auto-API lowerCAmelCase__ : Optional[Any] = CustomFeatureExtractor.from_pretrained(a ) with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : List[Any] = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : Tuple = CustomTokenizer(a ) lowerCAmelCase__ : Dict = CustomProcessor(a , a ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(a ) lowerCAmelCase__ : List[str] = AutoProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' class A__ ( __magic_name__ ): lowercase = False class A__ ( __magic_name__ ): lowercase = False class A__ ( __magic_name__ ): lowercase = 'AutoFeatureExtractor' lowercase = 'AutoTokenizer' lowercase = False try: AutoConfig.register('custom' , a ) AutoFeatureExtractor.register(a , a ) AutoTokenizer.register(a , slow_tokenizer_class=a ) AutoProcessor.register(a , a ) # If remote code is not set, the default is to use local classes. lowerCAmelCase__ : Dict = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. lowerCAmelCase__ : Dict = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=a ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. lowerCAmelCase__ : Dict = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=a ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(processor.__class__.__name__ , 'BertTokenizerFast' ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-convnext' ) self.assertEqual(processor.__class__.__name__ , 'ConvNextImageProcessor' ) @is_staging_test class A__ ( unittest.TestCase ): lowercase = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] @classmethod def _lowerCamelCase ( cls : Tuple ): '''simple docstring''' lowerCAmelCase__ : List[str] = TOKEN HfFolder.save_token(a ) @classmethod def _lowerCamelCase ( cls : str ): '''simple docstring''' try: delete_repo(token=cls._token , repo_id='test-processor' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-processor-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-processor' ) except HTTPError: pass def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = WavaVecaProcessor.from_pretrained(a ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(a , 'test-processor' ) , push_to_hub=a , use_auth_token=self._token ) lowerCAmelCase__ : int = WavaVecaProcessor.from_pretrained(f'''{USER}/test-processor''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(a , getattr(new_processor.feature_extractor , a ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = WavaVecaProcessor.from_pretrained(a ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(a , 'test-processor-org' ) , push_to_hub=a , use_auth_token=self._token , organization='valid_org' , ) lowerCAmelCase__ : List[str] = WavaVecaProcessor.from_pretrained('valid_org/test-processor-org' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(a , getattr(new_processor.feature_extractor , a ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() lowerCAmelCase__ : Union[str, Any] = CustomFeatureExtractor.from_pretrained(a ) with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase__ : Optional[int] = os.path.join(a , 'vocab.txt' ) with open(a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) lowerCAmelCase__ : Any = CustomTokenizer(a ) lowerCAmelCase__ : Tuple = CustomProcessor(a , a ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(f'''{USER}/test-dynamic-processor''' , token=self._token ) lowerCAmelCase__ : str = Repository(a , clone_from=f'''{USER}/test-dynamic-processor''' , token=self._token ) processor.save_pretrained(a ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { 'AutoFeatureExtractor': 'custom_feature_extraction.CustomFeatureExtractor', 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(a , 'tokenizer_config.json' ) ) as f: lowerCAmelCase__ : List[str] = json.load(a ) self.assertDictEqual( tokenizer_config['auto_map'] , { 'AutoTokenizer': ['custom_tokenization.CustomTokenizer', None], 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(a , 'custom_feature_extraction.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(a , 'custom_tokenization.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(a , 'custom_processing.py' ) ) ) repo.push_to_hub() lowerCAmelCase__ : int = AutoProcessor.from_pretrained(f'''{USER}/test-dynamic-processor''' , trust_remote_code=a ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , 'CustomProcessor' )
710
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : int = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : List[Any] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : int = None lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Dict = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Union[str, Any] = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
69
0
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class A__ : lowercase = 42 # [batch_size x 3] lowercase = 42 # [batch_size x 3] lowercase = 42 # [batch_size x 3] lowercase = 42 # [batch_size x 3] lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 def _lowerCamelCase ( self : str ): '''simple docstring''' assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = torch.arange(self.height * self.width ) lowerCAmelCase__ : List[Any] = torch.stack( [ pixel_indices % self.width, torch.div(a , self.width , rounding_mode='trunc' ), ] , axis=1 , ) return coords @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.shape lowerCAmelCase__ : Dict = int(np.prod(a ) ) lowerCAmelCase__ : Union[str, Any] = self.get_image_coords() lowerCAmelCase__ : Union[str, Any] = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) lowerCAmelCase__ : Optional[Any] = self.get_camera_rays(a ) lowerCAmelCase__ : Any = rays.view(a , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def _lowerCamelCase ( self : Optional[int] , a : torch.Tensor ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] lowerCAmelCase__ : Union[str, Any] = coords.view(a , -1 , 2 ) lowerCAmelCase__ : Optional[int] = self.resolution() lowerCAmelCase__ : List[str] = self.fov() lowerCAmelCase__ : int = (flat.float() / (res - 1)) * 2 - 1 lowerCAmelCase__ : str = fracs * torch.tan(fov / 2 ) lowerCAmelCase__ : Tuple = fracs.view(a , -1 , 2 ) lowerCAmelCase__ : Optional[int] = ( self.z.view(a , 1 , 3 ) + self.x.view(a , 1 , 3 ) * fracs[:, :, :1] + self.y.view(a , 1 , 3 ) * fracs[:, :, 1:] ) lowerCAmelCase__ : List[Any] = directions / directions.norm(dim=-1 , keepdim=a ) lowerCAmelCase__ : Any = torch.stack( [ torch.broadcast_to(self.origin.view(a , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(a , *a , 2 , 3 ) def _lowerCamelCase ( self : Dict , a : int , a : int ): '''simple docstring''' assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=a , height=a , x_fov=self.x_fov , y_fov=self.y_fov , ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> DifferentiableProjectiveCamera: lowerCAmelCase__ : str = [] lowerCAmelCase__ : List[Any] = [] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Optional[int] = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): lowerCAmelCase__ : Optional[Any] = np.array([np.sin(SCREAMING_SNAKE_CASE_ ), np.cos(SCREAMING_SNAKE_CASE_ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) lowerCAmelCase__ : List[Any] = -z * 4 lowerCAmelCase__ : Tuple = np.array([np.cos(SCREAMING_SNAKE_CASE_ ), -np.sin(SCREAMING_SNAKE_CASE_ ), 0.0] ) lowerCAmelCase__ : Optional[Any] = np.cross(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) origins.append(SCREAMING_SNAKE_CASE_ ) xs.append(SCREAMING_SNAKE_CASE_ ) ys.append(SCREAMING_SNAKE_CASE_ ) zs.append(SCREAMING_SNAKE_CASE_ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(SCREAMING_SNAKE_CASE_ , axis=0 ) ).float() , x=torch.from_numpy(np.stack(SCREAMING_SNAKE_CASE_ , axis=0 ) ).float() , y=torch.from_numpy(np.stack(SCREAMING_SNAKE_CASE_ , axis=0 ) ).float() , z=torch.from_numpy(np.stack(SCREAMING_SNAKE_CASE_ , axis=0 ) ).float() , width=SCREAMING_SNAKE_CASE_ , height=SCREAMING_SNAKE_CASE_ , x_fov=0.7 , y_fov=0.7 , shape=(1, len(SCREAMING_SNAKE_CASE_ )) , )
711
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
0
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_153.1_833 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 258.9_606 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 202.0_296 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
712
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
0
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 3 ) -> qiskit.result.counts.Counts: if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise TypeError('number of qubits must be a integer.' ) if number_of_qubits <= 0: raise ValueError('number of qubits must be > 0.' ) if math.floor(SCREAMING_SNAKE_CASE_ ) != number_of_qubits: raise ValueError('number of qubits must be exact integer.' ) if number_of_qubits > 10: raise ValueError('number of qubits too large to simulate(>10).' ) lowerCAmelCase__ : Any = QuantumRegister(SCREAMING_SNAKE_CASE_ , 'qr' ) lowerCAmelCase__ : List[Any] = ClassicalRegister(SCREAMING_SNAKE_CASE_ , 'cr' ) lowerCAmelCase__ : Optional[int] = QuantumCircuit(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = number_of_qubits for i in range(SCREAMING_SNAKE_CASE_ ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(SCREAMING_SNAKE_CASE_ ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(SCREAMING_SNAKE_CASE_ , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # simulate with 10000 shots lowerCAmelCase__ : str = Aer.get_backend('qasm_simulator' ) lowerCAmelCase__ : List[Any] = execute(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , shots=10_000 ) return job.result().get_counts(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": print( F"""Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}""" )
713
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
0
lowerCamelCase__ = {"""a""": ["""c""", """b"""], """b""": ["""d""", """e"""], """c""": [], """d""": [], """e""": []} lowerCamelCase__ = ["""a""", """b""", """c""", """d""", """e"""] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: lowerCAmelCase__ : Any = start # add current to visited visited.append(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: lowerCAmelCase__ : str = topological_sort(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # if all neighbors visited add current to sort sort.append(SCREAMING_SNAKE_CASE_ ) # if all vertices haven't been visited select a new one to visit if len(SCREAMING_SNAKE_CASE_ ) != len(SCREAMING_SNAKE_CASE_ ): for vertice in vertices: if vertice not in visited: lowerCAmelCase__ : Tuple = topological_sort(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # return sort return sort if __name__ == "__main__": lowerCamelCase__ = topological_sort("""a""", [], []) print(sort)
714
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
0
from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker lowerCamelCase__ = """CompVis/stable-diffusion-v1-1""" lowerCamelCase__ = """CompVis/stable-diffusion-v1-2""" lowerCamelCase__ = """CompVis/stable-diffusion-v1-3""" lowerCamelCase__ = """CompVis/stable-diffusion-v1-4""" class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : AutoencoderKL , a : CLIPTextModel , a : CLIPTokenizer , a : UNetaDConditionModel , a : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , a : StableDiffusionSafetyChecker , a : CLIPImageProcessor , a : bool = True , ): '''simple docstring''' super()._init_() lowerCAmelCase__ : List[Any] = StableDiffusionPipeline.from_pretrained(a ) lowerCAmelCase__ : Dict = StableDiffusionPipeline.from_pretrained(a ) lowerCAmelCase__ : int = StableDiffusionPipeline.from_pretrained(a ) lowerCAmelCase__ : List[Any] = StableDiffusionPipeline( vae=a , text_encoder=a , tokenizer=a , unet=a , scheduler=a , safety_checker=a , feature_extractor=a , requires_safety_checker=a , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea ) @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return {k: getattr(self , a ) for k in self.config.keys() if not k.startswith('_' )} def _lowerCamelCase ( self : List[Any] , a : Optional[Union[str, int]] = "auto" ): '''simple docstring''' if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowerCAmelCase__ : List[str] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' self.enable_attention_slicing(a ) @torch.no_grad() def _lowerCamelCase ( self : List[Any] , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : Tuple , ): '''simple docstring''' return self.pipea( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) @torch.no_grad() def _lowerCamelCase ( self : Tuple , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : int , ): '''simple docstring''' return self.pipea( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) @torch.no_grad() def _lowerCamelCase ( self : List[Any] , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : List[str] , ): '''simple docstring''' return self.pipea( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) @torch.no_grad() def _lowerCamelCase ( self : List[Any] , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : List[str] , ): '''simple docstring''' return self.pipea( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) @torch.no_grad() def _lowerCamelCase ( self : Tuple , a : Union[str, List[str]] , a : int = 512 , a : int = 512 , a : int = 50 , a : float = 7.5 , a : Optional[Union[str, List[str]]] = None , a : Optional[int] = 1 , a : float = 0.0 , a : Optional[torch.Generator] = None , a : Optional[torch.FloatTensor] = None , a : Optional[str] = "pil" , a : bool = True , a : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a : int = 1 , **a : List[Any] , ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cuda' if torch.cuda.is_available() else 'cpu' self.to(a ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` must be divisible by 8 but are {height} and {width}.''' ) # Get first result from Stable Diffusion Checkpoint v1.1 lowerCAmelCase__ : Optional[Any] = self.textaimg_sda_a( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) # Get first result from Stable Diffusion Checkpoint v1.2 lowerCAmelCase__ : Dict = self.textaimg_sda_a( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) # Get first result from Stable Diffusion Checkpoint v1.3 lowerCAmelCase__ : str = self.textaimg_sda_a( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) # Get first result from Stable Diffusion Checkpoint v1.4 lowerCAmelCase__ : List[str] = self.textaimg_sda_a( prompt=a , height=a , width=a , num_inference_steps=a , guidance_scale=a , negative_prompt=a , num_images_per_prompt=a , eta=a , generator=a , latents=a , output_type=a , return_dict=a , callback=a , callback_steps=a , **a , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
715
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
0
'''simple docstring''' import argparse import collections import torch from flax import traverse_util from tax import checkpoints from transformers import TaConfig, TaEncoderModel, TaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_="attention" ) -> List[str]: lowerCAmelCase__ : Optional[Any] = params[F'''{prefix}/layers_{i}/{layer_name}/key/kernel'''] lowerCAmelCase__ : str = params[F'''{prefix}/layers_{i}/{layer_name}/out/kernel'''] lowerCAmelCase__ : List[str] = params[F'''{prefix}/layers_{i}/{layer_name}/query/kernel'''] lowerCAmelCase__ : List[Any] = params[F'''{prefix}/layers_{i}/{layer_name}/value/kernel'''] return k, o, q, v def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ) -> str: if split_mlp_wi: lowerCAmelCase__ : Optional[int] = params[F'''{prefix}/layers_{i}/mlp/wi_0/kernel'''] lowerCAmelCase__ : Tuple = params[F'''{prefix}/layers_{i}/mlp/wi_1/kernel'''] lowerCAmelCase__ : str = (wi_a, wi_a) else: lowerCAmelCase__ : Tuple = params[F'''{prefix}/layers_{i}/mlp/wi/kernel'''] lowerCAmelCase__ : Union[str, Any] = params[F'''{prefix}/layers_{i}/mlp/wo/kernel'''] return wi, wo def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: return params[F'''{prefix}/layers_{i}/{layer_name}/scale'''] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , *, SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Tuple = traverse_util.flatten_dict(variables['target'] ) lowerCAmelCase__ : int = {'/'.join(SCREAMING_SNAKE_CASE_ ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi lowerCAmelCase__ : Dict = 'encoder/layers_0/mlp/wi_0/kernel' in old print('Split MLP:' , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = collections.OrderedDict() # Shared embeddings. lowerCAmelCase__ : Optional[Any] = old['token_embedder/embedding'] # Encoder. for i in range(SCREAMING_SNAKE_CASE_ ): # Block i, layer 0 (Self Attention). lowerCAmelCase__ : Optional[Any] = tax_layer_norm_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'encoder' , 'pre_attention_layer_norm' ) lowerCAmelCase__ : Union[str, Any] = tax_attention_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'encoder' , 'attention' ) lowerCAmelCase__ : Tuple = layer_norm lowerCAmelCase__ : Tuple = k.T lowerCAmelCase__ : int = o.T lowerCAmelCase__ : Union[str, Any] = q.T lowerCAmelCase__ : Optional[Any] = v.T # Block i, layer 1 (MLP). lowerCAmelCase__ : Dict = tax_layer_norm_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'encoder' , 'pre_mlp_layer_norm' ) lowerCAmelCase__ : Optional[Any] = tax_mlp_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'encoder' , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Union[str, Any] = layer_norm if split_mlp_wi: lowerCAmelCase__ : Dict = wi[0].T lowerCAmelCase__ : Union[str, Any] = wi[1].T else: lowerCAmelCase__ : int = wi.T lowerCAmelCase__ : int = wo.T lowerCAmelCase__ : Tuple = old[ 'encoder/relpos_bias/rel_embedding' ].T lowerCAmelCase__ : Union[str, Any] = old['encoder/encoder_norm/scale'] if not is_encoder_only: # Decoder. for i in range(SCREAMING_SNAKE_CASE_ ): # Block i, layer 0 (Self Attention). lowerCAmelCase__ : str = tax_layer_norm_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'decoder' , 'pre_self_attention_layer_norm' ) lowerCAmelCase__ : int = tax_attention_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'decoder' , 'self_attention' ) lowerCAmelCase__ : List[str] = layer_norm lowerCAmelCase__ : Tuple = k.T lowerCAmelCase__ : Union[str, Any] = o.T lowerCAmelCase__ : Optional[int] = q.T lowerCAmelCase__ : Tuple = v.T # Block i, layer 1 (Cross Attention). lowerCAmelCase__ : str = tax_layer_norm_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'decoder' , 'pre_cross_attention_layer_norm' ) lowerCAmelCase__ : Tuple = tax_attention_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'decoder' , 'encoder_decoder_attention' ) lowerCAmelCase__ : Tuple = layer_norm lowerCAmelCase__ : List[Any] = k.T lowerCAmelCase__ : Optional[int] = o.T lowerCAmelCase__ : int = q.T lowerCAmelCase__ : Optional[Any] = v.T # Block i, layer 2 (MLP). lowerCAmelCase__ : Optional[int] = tax_layer_norm_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'decoder' , 'pre_mlp_layer_norm' ) lowerCAmelCase__ : Tuple = tax_mlp_lookup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 'decoder' , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = layer_norm if split_mlp_wi: lowerCAmelCase__ : Dict = wi[0].T lowerCAmelCase__ : str = wi[1].T else: lowerCAmelCase__ : Any = wi.T lowerCAmelCase__ : str = wo.T lowerCAmelCase__ : Dict = old['decoder/decoder_norm/scale'] lowerCAmelCase__ : List[Any] = old[ 'decoder/relpos_bias/rel_embedding' ].T # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: lowerCAmelCase__ : Dict = old['decoder/logits_dense/kernel'].T return new def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Tuple = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: lowerCAmelCase__ : Tuple = state_dict['shared.weight'] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: lowerCAmelCase__ : int = state_dict['shared.weight'] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('Using shared word embeddings as lm_head.' ) lowerCAmelCase__ : Dict = state_dict['shared.weight'] return state_dict def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: lowerCAmelCase__ : str = checkpoints.load_tax_checkpoint(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = convert_tax_to_pytorch(SCREAMING_SNAKE_CASE_ , num_layers=config.num_layers , is_encoder_only=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Union[str, Any] = make_state_dict(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) model.load_state_dict(SCREAMING_SNAKE_CASE_ , strict=SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = False ) -> List[str]: lowerCAmelCase__ : Optional[Any] = TaConfig.from_json_file(SCREAMING_SNAKE_CASE_ ) print(F'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: lowerCAmelCase__ : Any = TaEncoderModel(SCREAMING_SNAKE_CASE_ ) else: lowerCAmelCase__ : Optional[int] = TaForConditionalGeneration(SCREAMING_SNAKE_CASE_ ) # Load weights from tf checkpoint load_tax_weights_in_ta(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) # Verify that we can load the checkpoint. model.from_pretrained(SCREAMING_SNAKE_CASE_ ) print('Done' ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser(description="""Converts a native T5X checkpoint into a PyTorch checkpoint.""") # Required parameters parser.add_argument( """--t5x_checkpoint_path""", default=None, type=str, required=True, help="""Path to the T5X checkpoint.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--is_encoder_only""", action="""store_true""", help="""Check if the model is encoder-decoder model""", default=False ) lowerCamelCase__ = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only )
716
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
0
from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time lowerCamelCase__ = Lock() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(SCREAMING_SNAKE_CASE_ ) process_lock.release() # receive your right neighbor's value process_lock.acquire() lowerCAmelCase__ : List[Any] = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left lowerCAmelCase__ : List[Any] = min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(SCREAMING_SNAKE_CASE_ ) process_lock.release() # receive your left neighbor's value process_lock.acquire() lowerCAmelCase__ : str = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right lowerCAmelCase__ : Optional[int] = max(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # after all swaps are performed, send the values back to main result_pipe[1].send(SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Tuple = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop lowerCAmelCase__ : Tuple = Pipe() lowerCAmelCase__ : Union[str, Any] = Pipe() process_array_.append( Process( target=SCREAMING_SNAKE_CASE_ , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) lowerCAmelCase__ : Tuple = temp_rs lowerCAmelCase__ : List[Any] = temp_rr for i in range(1 , len(SCREAMING_SNAKE_CASE_ ) - 1 ): lowerCAmelCase__ : Optional[int] = Pipe() lowerCAmelCase__ : str = Pipe() process_array_.append( Process( target=SCREAMING_SNAKE_CASE_ , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) lowerCAmelCase__ : str = temp_rs lowerCAmelCase__ : List[str] = temp_rr process_array_.append( Process( target=SCREAMING_SNAKE_CASE_ , args=( len(SCREAMING_SNAKE_CASE_ ) - 1, arr[len(SCREAMING_SNAKE_CASE_ ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(SCREAMING_SNAKE_CASE_ ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ : Tuple = result_pipe[p][0].recv() process_array_[p].join() return arr def lowerCAmelCase__ ( ) -> List[str]: lowerCAmelCase__ : int = list(range(10 , 0 , -1 ) ) print('Initial List' ) print(*SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = odd_even_transposition(SCREAMING_SNAKE_CASE_ ) print('Sorted List\n' ) print(*SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": main()
717
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
'''simple docstring''' from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['input_features', 'attention_mask'] def __init__( self : Dict , a : List[Any]=80 , a : Tuple=16_000 , a : List[str]=80 , a : int=0.0 , a : str=True , a : List[Any]=True , a : Any=True , **a : int , ): '''simple docstring''' super().__init__(feature_size=a , sampling_rate=a , padding_value=a , **a ) lowerCAmelCase__ : Optional[int] = num_mel_bins lowerCAmelCase__ : Union[str, Any] = do_ceptral_normalize lowerCAmelCase__ : int = normalize_means lowerCAmelCase__ : Tuple = normalize_vars lowerCAmelCase__ : Tuple = True def _lowerCamelCase ( self : List[str] , a : np.ndarray , ): '''simple docstring''' lowerCAmelCase__ : str = waveform * (2**15) # Kaldi compliance: 16-bit signed integers lowerCAmelCase__ : Union[str, Any] = torch.from_numpy(a ).unsqueeze(0 ) lowerCAmelCase__ : Optional[int] = ta_kaldi.fbank(a , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def _lowerCamelCase ( a : np.ndarray , a : int , a : Optional[bool] = True , a : Optional[bool] = True , a : float = 0.0 , ): '''simple docstring''' if normalize_means: lowerCAmelCase__ : int = x[:input_length].mean(axis=0 ) lowerCAmelCase__ : Dict = np.subtract(a , a ) if normalize_vars: lowerCAmelCase__ : Dict = x[:input_length].std(axis=0 ) lowerCAmelCase__ : List[Any] = np.divide(a , a ) if input_length < x.shape[0]: lowerCAmelCase__ : Optional[Any] = padding_value # make sure array is in float32 lowerCAmelCase__ : Any = x.astype(np.floataa ) return x def _lowerCamelCase ( self : Dict , a : List[np.ndarray] , a : Optional[np.ndarray] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(a , a , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(a , a ) ] def __call__( self : Any , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Union[bool, str, PaddingStrategy] = False , a : Optional[int] = None , a : bool = False , a : Optional[int] = None , a : Optional[Union[str, TensorType]] = None , a : Optional[int] = None , a : Optional[bool] = None , **a : Any , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : Optional[int] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : Optional[int] = [np.asarray(a , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Dict = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : Tuple = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Any = [raw_speech] # extract fbank features lowerCAmelCase__ : Tuple = [self._extract_fbank_features(a ) for waveform in raw_speech] # convert into correct format for padding lowerCAmelCase__ : Tuple = BatchFeature({'input_features': features} ) lowerCAmelCase__ : Tuple = self.pad( a , padding=a , max_length=a , truncation=a , pad_to_multiple_of=a , return_attention_mask=a , **a , ) # make sure list is in array format lowerCAmelCase__ : Optional[int] = padded_inputs.get('input_features' ) if isinstance(input_features[0] , a ): lowerCAmelCase__ : int = [np.asarray(a , dtype=np.floataa ) for feature in input_features] lowerCAmelCase__ : int = padded_inputs.get('attention_mask' ) if attention_mask is not None: lowerCAmelCase__ : str = [np.asarray(a , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: lowerCAmelCase__ : Tuple = ( np.array(a , dtype=np.intaa ) if self._get_padding_strategies(a , max_length=a ) is not PaddingStrategy.DO_NOT_PAD else None ) lowerCAmelCase__ : int = self.normalize( padded_inputs['input_features'] , attention_mask=a ) if return_tensors is not None: lowerCAmelCase__ : int = padded_inputs.convert_to_tensors(a ) return padded_inputs
718
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
from typing import Dict from .base import GenericTensor, Pipeline class A__ ( __magic_name__ ): def _lowerCamelCase ( self : Dict , a : Optional[int]=None , a : Optional[int]=None , a : Dict=None , **a : Optional[int] ): '''simple docstring''' if tokenize_kwargs is None: lowerCAmelCase__ : List[str] = {} if truncation is not None: if "truncation" in tokenize_kwargs: raise ValueError( 'truncation parameter defined twice (given as keyword argument as well as in tokenize_kwargs)' ) lowerCAmelCase__ : str = truncation lowerCAmelCase__ : Optional[Any] = tokenize_kwargs lowerCAmelCase__ : Tuple = {} if return_tensors is not None: lowerCAmelCase__ : Union[str, Any] = return_tensors return preprocess_params, {}, postprocess_params def _lowerCamelCase ( self : Dict , a : List[Any] , **a : Tuple ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.framework lowerCAmelCase__ : Dict = self.tokenizer(a , return_tensors=a , **a ) return model_inputs def _lowerCamelCase ( self : Optional[int] , a : str ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model(**a ) return model_outputs def _lowerCamelCase ( self : int , a : Union[str, Any] , a : Union[str, Any]=False ): '''simple docstring''' if return_tensors: return model_outputs[0] if self.framework == "pt": return model_outputs[0].tolist() elif self.framework == "tf": return model_outputs[0].numpy().tolist() def __call__( self : int , *a : Optional[Any] , **a : Any ): '''simple docstring''' return super().__call__(*a , **a )
719
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
from collections.abc import Callable def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> float: lowerCAmelCase__ : float = a lowerCAmelCase__ : float = b if function(SCREAMING_SNAKE_CASE_ ) == 0: # one of the a or b is a root for the function return a elif function(SCREAMING_SNAKE_CASE_ ) == 0: return b elif ( function(SCREAMING_SNAKE_CASE_ ) * function(SCREAMING_SNAKE_CASE_ ) > 0 ): # if none of these are root and they are both positive or negative, # then this algorithm can't find the root raise ValueError('could not find root in given interval.' ) else: lowerCAmelCase__ : float = start + (end - start) / 2.0 while abs(start - mid ) > 10**-7: # until precisely equals to 10^-7 if function(SCREAMING_SNAKE_CASE_ ) == 0: return mid elif function(SCREAMING_SNAKE_CASE_ ) * function(SCREAMING_SNAKE_CASE_ ) < 0: lowerCAmelCase__ : Any = mid else: lowerCAmelCase__ : Dict = mid lowerCAmelCase__ : List[Any] = start + (end - start) / 2.0 return mid def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> float: return x**3 - 2 * x - 5 if __name__ == "__main__": print(bisection(f, 1, 1000)) import doctest doctest.testmod()
720
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
0
from __future__ import annotations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: print(F'''Vertex\tShortest Distance from vertex {src}''' ) for i, d in enumerate(SCREAMING_SNAKE_CASE_ ): print(F'''{i}\t\t{d}''' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: for j in range(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Dict = (graph[j][k] for k in ['src', 'dst', 'weight']) if distance[u] != float('inf' ) and distance[u] + w < distance[v]: return True return False def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> list[float]: lowerCAmelCase__ : Union[str, Any] = [float('inf' )] * vertex_count lowerCAmelCase__ : str = 0.0 for _ in range(vertex_count - 1 ): for j in range(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : int = (graph[j][k] for k in ['src', 'dst', 'weight']) if distance[u] != float('inf' ) and distance[u] + w < distance[v]: lowerCAmelCase__ : str = distance[u] + w lowerCAmelCase__ : Optional[Any] = check_negative_cycle(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if negative_cycle_exists: raise Exception('Negative cycle found' ) return distance if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase__ = int(input("""Enter number of vertices: """).strip()) lowerCamelCase__ = int(input("""Enter number of edges: """).strip()) lowerCamelCase__ = [{} for _ in range(E)] for i in range(E): print("""Edge """, i + 1) lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ = ( int(x) for x in input("""Enter source, destination, weight: """).strip().split(""" """) ) lowerCamelCase__ = {"""src""": src, """dst""": dest, """weight""": weight} lowerCamelCase__ = int(input("""\nEnter shortest path source:""").strip()) lowerCamelCase__ = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
721
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
69
0
from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """allegro/herbert-base-cased""": """https://huggingface.co/allegro/herbert-base-cased/resolve/main/vocab.json""" }, """merges_file""": { """allegro/herbert-base-cased""": """https://huggingface.co/allegro/herbert-base-cased/resolve/main/merges.txt""" }, } lowerCamelCase__ = {"""allegro/herbert-base-cased""": 514} lowerCamelCase__ = {} class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_INIT_CONFIGURATION lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = HerbertTokenizer def __init__( self : int , a : Dict=None , a : str=None , a : str=None , a : List[Any]="<s>" , a : Union[str, Any]="<unk>" , a : Union[str, Any]="<pad>" , a : Union[str, Any]="<mask>" , a : List[str]="</s>" , **a : List[Any] , ): '''simple docstring''' super().__init__( a , a , tokenizer_file=a , cls_token=a , unk_token=a , pad_token=a , mask_token=a , sep_token=a , **a , ) def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.cls_token_id] lowerCAmelCase__ : Any = [self.sep_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : Optional[Any] , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[str] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = [self.sep_token_id] lowerCAmelCase__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Optional[Any] , a : str , a : Optional[str] = None ): '''simple docstring''' lowerCAmelCase__ : int = self._tokenizer.model.save(a , name=a ) return tuple(a )
700
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __magic_name__ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'LayoutLMv3ImageProcessor' lowercase = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self : Optional[int] , a : Union[str, Any]=None , a : Optional[Any]=None , **a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , a , ) lowerCAmelCase__ : int = kwargs.pop('feature_extractor' ) lowerCAmelCase__ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(a , a ) def __call__( self : List[Any] , a : List[Any] , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a : Union[List[List[int]], List[List[List[int]]]] = None , a : Optional[Union[List[int], List[List[int]]]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : str , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor lowerCAmelCase__ : List[str] = self.image_processor(images=a , return_tensors=a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a , a ): lowerCAmelCase__ : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ : List[str] = features['words'] lowerCAmelCase__ : List[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_token_type_ids=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_length=a , verbose=a , return_tensors=a , **a , ) # add pixel values lowerCAmelCase__ : Tuple = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase__ : List[str] = self.get_overflowing_images(a , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase__ : List[str] = images return encoded_inputs def _lowerCamelCase ( self : Any , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a ) != len(a ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' f''' {len(a )} and {len(a )}''' ) return images_with_overflow def _lowerCamelCase ( self : Union[str, Any] , *a : Optional[Any] , **a : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Tuple , *a : List[str] , **a : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*a , **a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , a , ) return self.image_processor_class @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , a , ) return self.image_processor
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: assert isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ), F'''The input value of [n={number}] is not an integer''' if number == 1: return 2 elif number < 1: lowerCAmelCase__ : Dict = F'''The input value of [n={number}] has to be > 0''' raise ValueError(SCREAMING_SNAKE_CASE_ ) else: lowerCAmelCase__ : str = sylvester(number - 1 ) lowerCAmelCase__ : Dict = num - 1 lowerCAmelCase__ : Dict = num return lower * upper + 1 if __name__ == "__main__": print(F"""The 8th number in Sylvester's sequence: {sylvester(8)}""")
701
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
69
0
from __future__ import annotations class A__ : def __init__( self : Any , a : int = 0 ): '''simple docstring''' lowerCAmelCase__ : Any = key def _lowerCamelCase ( self : Any , a : str , a : int ): '''simple docstring''' assert isinstance(a , a ) and isinstance(a , a ) lowerCAmelCase__ : str = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(a ) ^ key ) for ch in content] def _lowerCamelCase ( self : Union[str, Any] , a : str , a : int ): '''simple docstring''' assert isinstance(a , a ) and isinstance(a , a ) lowerCAmelCase__ : Optional[Any] = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(a ) ^ key ) for ch in content] def _lowerCamelCase ( self : Union[str, Any] , a : str , a : int = 0 ): '''simple docstring''' assert isinstance(a , a ) and isinstance(a , a ) lowerCAmelCase__ : List[str] = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned lowerCAmelCase__ : Union[str, Any] = '' for ch in content: ans += chr(ord(a ) ^ key ) return ans def _lowerCamelCase ( self : Dict , a : str , a : int = 0 ): '''simple docstring''' assert isinstance(a , a ) and isinstance(a , a ) lowerCAmelCase__ : List[str] = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned lowerCAmelCase__ : Any = '' for ch in content: ans += chr(ord(a ) ^ key ) return ans def _lowerCamelCase ( self : List[str] , a : str , a : int = 0 ): '''simple docstring''' assert isinstance(a , a ) and isinstance(a , a ) try: with open(a ) as fin, open('encrypt.out' , 'w+' ) as fout: # actual encrypt-process for line in fin: fout.write(self.encrypt_string(a , a ) ) except OSError: return False return True def _lowerCamelCase ( self : List[Any] , a : str , a : int ): '''simple docstring''' assert isinstance(a , a ) and isinstance(a , a ) try: with open(a ) as fin, open('decrypt.out' , 'w+' ) as fout: # actual encrypt-process for line in fin: fout.write(self.decrypt_string(a , a ) ) except OSError: return False return True # Tests # crypt = XORCipher() # key = 67 # # test encrypt # print(crypt.encrypt("hallo welt",key)) # # test decrypt # print(crypt.decrypt(crypt.encrypt("hallo welt",key), key)) # # test encrypt_string # print(crypt.encrypt_string("hallo welt",key)) # # test decrypt_string # print(crypt.decrypt_string(crypt.encrypt_string("hallo welt",key),key)) # if (crypt.encrypt_file("test.txt",key)): # print("encrypt successful") # else: # print("encrypt unsuccessful") # if (crypt.decrypt_file("encrypt.out",key)): # print("decrypt successful") # else: # print("decrypt unsuccessful")
702
lowerCamelCase__ = """Alexander Joslin""" import operator as op from .stack import Stack def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} lowerCAmelCase__ : Stack[int] = Stack() lowerCAmelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE_ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE_ ) elif i == ")": # RULE 4 lowerCAmelCase__ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : List[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : Tuple = operators[opr](SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) operand_stack.push(SCREAMING_SNAKE_CASE_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
69
0
import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """b0""": efficientnet.EfficientNetBa, """b1""": efficientnet.EfficientNetBa, """b2""": efficientnet.EfficientNetBa, """b3""": efficientnet.EfficientNetBa, """b4""": efficientnet.EfficientNetBa, """b5""": efficientnet.EfficientNetBa, """b6""": efficientnet.EfficientNetBa, """b7""": efficientnet.EfficientNetBa, } lowerCamelCase__ = { """b0""": { """hidden_dim""": 1280, """width_coef""": 1.0, """depth_coef""": 1.0, """image_size""": 224, """dropout_rate""": 0.2, """dw_padding""": [], }, """b1""": { """hidden_dim""": 1280, """width_coef""": 1.0, """depth_coef""": 1.1, """image_size""": 240, """dropout_rate""": 0.2, """dw_padding""": [16], }, """b2""": { """hidden_dim""": 1408, """width_coef""": 1.1, """depth_coef""": 1.2, """image_size""": 260, """dropout_rate""": 0.3, """dw_padding""": [5, 8, 16], }, """b3""": { """hidden_dim""": 1536, """width_coef""": 1.2, """depth_coef""": 1.4, """image_size""": 300, """dropout_rate""": 0.3, """dw_padding""": [5, 18], }, """b4""": { """hidden_dim""": 1792, """width_coef""": 1.4, """depth_coef""": 1.8, """image_size""": 380, """dropout_rate""": 0.4, """dw_padding""": [6], }, """b5""": { """hidden_dim""": 2048, """width_coef""": 1.6, """depth_coef""": 2.2, """image_size""": 456, """dropout_rate""": 0.4, """dw_padding""": [13, 27], }, """b6""": { """hidden_dim""": 2304, """width_coef""": 1.8, """depth_coef""": 2.6, """image_size""": 528, """dropout_rate""": 0.5, """dw_padding""": [31], }, """b7""": { """hidden_dim""": 2560, """width_coef""": 2.0, """depth_coef""": 3.1, """image_size""": 600, """dropout_rate""": 0.5, """dw_padding""": [18], }, } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[Any]: lowerCAmelCase__ : int = EfficientNetConfig() lowerCAmelCase__ : Union[str, Any] = CONFIG_MAP[model_name]['hidden_dim'] lowerCAmelCase__ : List[str] = CONFIG_MAP[model_name]['width_coef'] lowerCAmelCase__ : Union[str, Any] = CONFIG_MAP[model_name]['depth_coef'] lowerCAmelCase__ : Tuple = CONFIG_MAP[model_name]['image_size'] lowerCAmelCase__ : str = CONFIG_MAP[model_name]['dropout_rate'] lowerCAmelCase__ : Any = CONFIG_MAP[model_name]['dw_padding'] lowerCAmelCase__ : int = 'huggingface/label-files' lowerCAmelCase__ : List[str] = 'imagenet-1k-id2label.json' lowerCAmelCase__ : Tuple = 1_000 lowerCAmelCase__ : List[str] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type='dataset' ) , 'r' ) ) lowerCAmelCase__ : Union[str, Any] = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} lowerCAmelCase__ : Union[str, Any] = idalabel lowerCAmelCase__ : int = {v: k for k, v in idalabel.items()} return config def lowerCAmelCase__ ( ) -> Any: lowerCAmelCase__ : Any = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCAmelCase__ : Union[str, Any] = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) return im def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : Any = CONFIG_MAP[model_name]['image_size'] lowerCAmelCase__ : List[str] = EfficientNetImageProcessor( size={'height': size, 'width': size} , image_mean=[0.485, 0.456, 0.406] , image_std=[0.47853944, 0.4732864, 0.47434163] , do_center_crop=SCREAMING_SNAKE_CASE_ , ) return preprocessor def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : Tuple = [v.split('_' )[0].split('block' )[1] for v in original_param_names if v.startswith('block' )] lowerCAmelCase__ : int = sorted(set(SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase__ : Dict = len(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = {b: str(SCREAMING_SNAKE_CASE_ ) for b, i in zip(SCREAMING_SNAKE_CASE_ , range(SCREAMING_SNAKE_CASE_ ) )} lowerCAmelCase__ : int = [] rename_keys.append(('stem_conv/kernel:0', 'embeddings.convolution.weight') ) rename_keys.append(('stem_bn/gamma:0', 'embeddings.batchnorm.weight') ) rename_keys.append(('stem_bn/beta:0', 'embeddings.batchnorm.bias') ) rename_keys.append(('stem_bn/moving_mean:0', 'embeddings.batchnorm.running_mean') ) rename_keys.append(('stem_bn/moving_variance:0', 'embeddings.batchnorm.running_var') ) for b in block_names: lowerCAmelCase__ : List[str] = block_name_mapping[b] rename_keys.append((F'''block{b}_expand_conv/kernel:0''', F'''encoder.blocks.{hf_b}.expansion.expand_conv.weight''') ) rename_keys.append((F'''block{b}_expand_bn/gamma:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.weight''') ) rename_keys.append((F'''block{b}_expand_bn/beta:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.bias''') ) rename_keys.append( (F'''block{b}_expand_bn/moving_mean:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.running_mean''') ) rename_keys.append( (F'''block{b}_expand_bn/moving_variance:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.running_var''') ) rename_keys.append( (F'''block{b}_dwconv/depthwise_kernel:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight''') ) rename_keys.append((F'''block{b}_bn/gamma:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight''') ) rename_keys.append((F'''block{b}_bn/beta:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias''') ) rename_keys.append( (F'''block{b}_bn/moving_mean:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean''') ) rename_keys.append( (F'''block{b}_bn/moving_variance:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var''') ) rename_keys.append((F'''block{b}_se_reduce/kernel:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.reduce.weight''') ) rename_keys.append((F'''block{b}_se_reduce/bias:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.reduce.bias''') ) rename_keys.append((F'''block{b}_se_expand/kernel:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.expand.weight''') ) rename_keys.append((F'''block{b}_se_expand/bias:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.expand.bias''') ) rename_keys.append( (F'''block{b}_project_conv/kernel:0''', F'''encoder.blocks.{hf_b}.projection.project_conv.weight''') ) rename_keys.append((F'''block{b}_project_bn/gamma:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.weight''') ) rename_keys.append((F'''block{b}_project_bn/beta:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.bias''') ) rename_keys.append( (F'''block{b}_project_bn/moving_mean:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.running_mean''') ) rename_keys.append( (F'''block{b}_project_bn/moving_variance:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.running_var''') ) rename_keys.append(('top_conv/kernel:0', 'encoder.top_conv.weight') ) rename_keys.append(('top_bn/gamma:0', 'encoder.top_bn.weight') ) rename_keys.append(('top_bn/beta:0', 'encoder.top_bn.bias') ) rename_keys.append(('top_bn/moving_mean:0', 'encoder.top_bn.running_mean') ) rename_keys.append(('top_bn/moving_variance:0', 'encoder.top_bn.running_var') ) lowerCAmelCase__ : Any = {} for item in rename_keys: if item[0] in original_param_names: lowerCAmelCase__ : Tuple = 'efficientnet.' + item[1] lowerCAmelCase__ : Tuple = 'classifier.weight' lowerCAmelCase__ : Union[str, Any] = 'classifier.bias' return key_mapping def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Dict: for key, value in tf_params.items(): if "normalization" in key: continue lowerCAmelCase__ : str = key_mapping[key] if "_conv" in key and "kernel" in key: lowerCAmelCase__ : int = torch.from_numpy(SCREAMING_SNAKE_CASE_ ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: lowerCAmelCase__ : List[str] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: lowerCAmelCase__ : Any = torch.from_numpy(np.transpose(SCREAMING_SNAKE_CASE_ ) ) else: lowerCAmelCase__ : Optional[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(SCREAMING_SNAKE_CASE_ ) @torch.no_grad() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: lowerCAmelCase__ : List[Any] = model_classes[model_name]( include_top=SCREAMING_SNAKE_CASE_ , weights='imagenet' , input_tensor=SCREAMING_SNAKE_CASE_ , input_shape=SCREAMING_SNAKE_CASE_ , pooling=SCREAMING_SNAKE_CASE_ , classes=1_000 , classifier_activation='softmax' , ) lowerCAmelCase__ : Union[str, Any] = original_model.trainable_variables lowerCAmelCase__ : Dict = original_model.non_trainable_variables lowerCAmelCase__ : str = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: lowerCAmelCase__ : List[str] = param.numpy() lowerCAmelCase__ : Optional[Any] = list(tf_params.keys() ) # Load HuggingFace model lowerCAmelCase__ : str = get_efficientnet_config(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = EfficientNetForImageClassification(SCREAMING_SNAKE_CASE_ ).eval() lowerCAmelCase__ : Any = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print('Converting parameters...' ) lowerCAmelCase__ : Optional[Any] = rename_keys(SCREAMING_SNAKE_CASE_ ) replace_params(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Initialize preprocessor and preprocess input image lowerCAmelCase__ : Any = convert_image_processor(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = preprocessor(images=prepare_img() , return_tensors='pt' ) # HF model inference hf_model.eval() with torch.no_grad(): lowerCAmelCase__ : List[Any] = hf_model(**SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = outputs.logits.detach().numpy() # Original model inference lowerCAmelCase__ : Optional[Any] = False lowerCAmelCase__ : Union[str, Any] = CONFIG_MAP[model_name]['image_size'] lowerCAmelCase__ : Optional[int] = prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) lowerCAmelCase__ : List[str] = image.img_to_array(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Tuple = np.expand_dims(SCREAMING_SNAKE_CASE_ , axis=0 ) lowerCAmelCase__ : Union[str, Any] = original_model.predict(SCREAMING_SNAKE_CASE_ ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1e-3 ), "The predicted logits are not the same." print('Model outputs match!' ) if save_model: # Create folder to save model if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): os.mkdir(SCREAMING_SNAKE_CASE_ ) # Save converted model and image processor hf_model.save_pretrained(SCREAMING_SNAKE_CASE_ ) preprocessor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: # Push model and image processor to hub print(F'''Pushing converted {model_name} to the hub...''' ) lowerCAmelCase__ : List[str] = F'''efficientnet-{model_name}''' preprocessor.push_to_hub(SCREAMING_SNAKE_CASE_ ) hf_model.push_to_hub(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""b0""", type=str, help="""Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""hf_model""", type=str, help="""Path to the output PyTorch model directory.""", ) parser.add_argument("""--save_model""", action="""store_true""", help="""Save model to local""") parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Push model and image processor to the hub""") lowerCamelCase__ = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
703
import numpy class A__ : def __init__( self : Tuple , a : numpy.ndarray , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : int = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Dict = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : List[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowerCAmelCase__ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : List[Any] = numpy.zeros(output_array.shape ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowerCAmelCase__ : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowerCAmelCase__ : int = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowerCamelCase ( self : Optional[int] , a : numpy.ndarray , a : int , a : bool ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): lowerCAmelCase__ : Any = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Tuple = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def _lowerCamelCase ( self : Optional[Any] , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : Dict = input_arr lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return (value) * (1 - (value)) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : List[str] = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ , output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ , iterations=10 , give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
69
0
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
704
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A__ : def __init__( self : int , a : str , a : Union[str, Any]=13 , a : int=32 , a : Optional[Any]=2 , a : Tuple=3 , a : List[Any]=16 , a : List[str]=[1, 2, 1] , a : int=[2, 2, 4] , a : int=2 , a : Optional[Any]=2.0 , a : Optional[int]=True , a : Dict=0.0 , a : Any=0.0 , a : int=0.1 , a : List[str]="gelu" , a : Optional[Any]=False , a : str=True , a : Dict=0.0_2 , a : Any=1E-5 , a : Optional[int]=True , a : str=None , a : str=True , a : int=10 , a : str=8 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : Optional[Any] = patch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = embed_dim lowerCAmelCase__ : Tuple = depths lowerCAmelCase__ : List[str] = num_heads lowerCAmelCase__ : List[Any] = window_size lowerCAmelCase__ : Any = mlp_ratio lowerCAmelCase__ : Optional[Any] = qkv_bias lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : int = drop_path_rate lowerCAmelCase__ : Optional[Any] = hidden_act lowerCAmelCase__ : int = use_absolute_embeddings lowerCAmelCase__ : List[str] = patch_norm lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Dict = use_labels lowerCAmelCase__ : List[Any] = type_sequence_label_size lowerCAmelCase__ : Optional[Any] = encoder_stride def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = None if self.use_labels: lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : int = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : List[str] , a : Any , a : str , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = SwinvaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model(a ) lowerCAmelCase__ : str = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase__ : Tuple = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SwinvaForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : Dict = SwinvaForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Union[str, Any] , a : int , a : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.type_sequence_label_size lowerCAmelCase__ : List[Any] = SwinvaForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = config_and_inputs lowerCAmelCase__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = SwinvaModelTester(self ) lowerCAmelCase__ : int = ConfigTester(self , config_class=a , embed_dim=37 ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[int] = model_class(a ) lowerCAmelCase__ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Tuple = [*signature.parameters.keys()] lowerCAmelCase__ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = True for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = True lowerCAmelCase__ : str = False lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Dict = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : int = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Dict = outputs.attentions lowerCAmelCase__ : Dict = len(self.model_tester.depths ) self.assertEqual(len(a ) , a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Optional[int] = config.window_size**2 lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowerCAmelCase__ : Tuple = len(a ) # Check attention is always last and order is fine lowerCAmelCase__ : str = True lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(**self._prepare_for_class(a , a ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): lowerCAmelCase__ : Optional[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase__ : Any = 2 self.assertEqual(out_len + added_hidden_states , len(a ) ) lowerCAmelCase__ : Dict = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _lowerCamelCase ( self : int , a : Optional[int] , a : int , a : Optional[Any] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.hidden_states lowerCAmelCase__ : str = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a ) , a ) # Swinv2 has a different seq_length lowerCAmelCase__ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowerCAmelCase__ : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(a ) , a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = reshaped_hidden_states[0].shape lowerCAmelCase__ : List[str] = ( reshaped_hidden_states[0].view(a , a , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : str = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : List[str] = True self.check_hidden_states_output(a , a , a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Any = 3 lowerCAmelCase__ : int = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase__ : str = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : str = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase__ : Any = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase__ : str = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[str] = SwinvaModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = _config_zero_init(a ) for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(config=a ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( a ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Any = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**a ) # verify the logits lowerCAmelCase__ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Optional[Any] = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) )
69
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_table_transformer""": [ """TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TableTransformerConfig""", """TableTransformerOnnxConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TableTransformerForObjectDetection""", """TableTransformerModel""", """TableTransformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
705
from itertools import permutations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase__ : str = [7, 11, 13, 17] for i, test in enumerate(SCREAMING_SNAKE_CASE_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 10 ) -> int: return sum( int(''.join(map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) ) for num in permutations(range(SCREAMING_SNAKE_CASE_ ) ) if is_substring_divisible(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
69
0
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=80.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=80.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=80.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=80.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=80.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
706
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
69
0
import gc import importlib.metadata import tempfile import unittest from packaging import version from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoTokenizer, BitsAndBytesConfig, pipeline, ) from transformers.testing_utils import ( is_torch_available, require_accelerate, require_bitsandbytes, require_torch, require_torch_gpu, require_torch_multi_gpu, slow, ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: if model.config.model_type == "gpt2": return model.transformer.h[0].mlp.c_fc return model.transformer.h[0].mlp.dense_ah_to_h if is_torch_available(): import torch import torch.nn as nn class A__ ( nn.Module ): '''simple docstring''' def __init__( self : Optional[int] , a : nn.Module , a : int ): '''simple docstring''' super().__init__() lowerCAmelCase__ : Optional[Any] = module lowerCAmelCase__ : Dict = nn.Sequential( nn.Linear(module.in_features , a , bias=a ) , nn.Linear(a , module.out_features , bias=a ) , ) lowerCAmelCase__ : List[Any] = (2.0 / (5 * min(module.in_features , module.out_features ))) ** 0.5 nn.init.normal_(self.adapter[0].weight , std=a ) nn.init.zeros_(self.adapter[1].weight ) self.adapter.to(module.weight.device ) def _lowerCamelCase ( self : List[str] , a : Any , *a : int , **a : Tuple ): '''simple docstring''' return self.module(a , *a , **a ) + self.adapter(a ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class A__ ( unittest.TestCase ): '''simple docstring''' lowercase = 'bigscience/bloom-1b7' # Constant values lowercase = 2.1_09_65_95_52_69_25_74 lowercase = 'Hello my name is' lowercase = set() EXPECTED_OUTPUTS.add('Hello my name is John and I am a professional photographer. I' ) EXPECTED_OUTPUTS.add('Hello my name is John.\nI am a friend of your father.\n' ) EXPECTED_OUTPUTS.add('Hello my name is John Doe, I am a student at the University' ) lowercase = 10 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = AutoTokenizer.from_pretrained(self.model_name ) class A__ ( __magic_name__ ): '''simple docstring''' def _lowerCamelCase ( self : List[str] ): '''simple docstring''' super().setUp() # Models and tokenizer lowerCAmelCase__ : int = AutoModelForCausalLM.from_pretrained( self.model_name , torch_dtype=torch.floataa , device_map='auto' ) lowerCAmelCase__ : Union[str, Any] = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=a , device_map='auto' ) def _lowerCamelCase ( self : str ): '''simple docstring''' del self.model_fpaa del self.model_abit gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_abit.config self.assertTrue(hasattr(a , 'quantization_config' ) ) lowerCAmelCase__ : List[Any] = config.to_dict() lowerCAmelCase__ : List[str] = config.to_diff_dict() lowerCAmelCase__ : Optional[int] = config.to_json_string() def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' from bitsandbytes.nn import Paramsabit lowerCAmelCase__ : Union[str, Any] = self.model_fpaa.get_memory_footprint() lowerCAmelCase__ : Optional[int] = self.model_abit.get_memory_footprint() self.assertAlmostEqual(mem_fpaa / mem_abit , self.EXPECTED_RELATIVE_DIFFERENCE ) lowerCAmelCase__ : Union[str, Any] = get_some_linear_layer(self.model_abit ) self.assertTrue(linear.weight.__class__ == Paramsabit ) def _lowerCamelCase ( self : str ): '''simple docstring''' from transformers import TaPreTrainedModel self.model_fpaa.get_memory_footprint() self.model_abit.get_memory_footprint() for name, module in self.model_abit.named_modules(): if isinstance(a , torch.nn.Linear ): if name not in ["lm_head"] + TaPreTrainedModel._keep_in_fpaa_modules: # 4-bit parameters are packed in uint8 variables self.assertTrue(module.weight.dtype == torch.uinta ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.tokenizer(self.input_text , return_tensors='pt' ) lowerCAmelCase__ : List[str] = self.model_abit.generate(input_ids=encoded_input['input_ids'].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=a ) , self.EXPECTED_OUTPUTS ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = BitsAndBytesConfig() lowerCAmelCase__ : Optional[Any] = True lowerCAmelCase__ : Optional[Any] = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=a , device_map='auto' ) lowerCAmelCase__ : int = self.tokenizer(self.input_text , return_tensors='pt' ) lowerCAmelCase__ : int = model_abit_from_config.generate( input_ids=encoded_input['input_ids'].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=a ) , self.EXPECTED_OUTPUTS ) def _lowerCamelCase ( self : str ): '''simple docstring''' with self.assertRaises(a ), tempfile.TemporaryDirectory() as tmpdirname: self.model_abit.save_pretrained(a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = BitsAndBytesConfig() with self.assertRaises(a ): lowerCAmelCase__ : Union[str, Any] = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=a , load_in_abit=a , device_map='auto' , bnb_abit_quant_type='nf4' , ) def _lowerCamelCase ( self : str ): '''simple docstring''' with self.assertRaises(a ): # Tries with `str` self.model_abit.to('cpu' ) with self.assertRaises(a ): # Tries with a `dtype`` self.model_abit.to(torch.floataa ) with self.assertRaises(a ): # Tries with a `device` self.model_abit.to(torch.device('cuda:0' ) ) with self.assertRaises(a ): # Tries with a `device` self.model_abit.float() with self.assertRaises(a ): # Tries with a `device` self.model_abit.half() # Test if we did not break anything lowerCAmelCase__ : str = self.tokenizer(self.input_text , return_tensors='pt' ) lowerCAmelCase__ : List[str] = self.model_fpaa.to(torch.floataa ) lowerCAmelCase__ : Tuple = self.model_fpaa.generate(input_ids=encoded_input['input_ids'].to(0 ) , max_new_tokens=10 ) # Check this does not throw an error lowerCAmelCase__ : List[Any] = self.model_fpaa.to('cpu' ) # Check this does not throw an error lowerCAmelCase__ : Union[str, Any] = self.model_fpaa.half() # Check this does not throw an error lowerCAmelCase__ : Optional[Any] = self.model_fpaa.float() def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : List[Any] = AutoModelForSeqaSeqLM.from_pretrained('t5-small' , load_in_abit=a , device_map='auto' ) self.assertTrue(model.decoder.block[0].layer[2].DenseReluDense.wo.weight.dtype == torch.floataa ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class A__ ( unittest.TestCase ): '''simple docstring''' @classmethod def _lowerCamelCase ( cls : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = 't5-small' lowerCAmelCase__ : Optional[int] = 'google/flan-t5-small' # flan-t5 uses dense-act instead of dense-relu-dense lowerCAmelCase__ : Union[str, Any] = AutoTokenizer.from_pretrained(cls.model_name ) lowerCAmelCase__ : Optional[int] = 'Translate in German: Hello, my dog is cute' def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Any ): '''simple docstring''' from transformers import TaForConditionalGeneration lowerCAmelCase__ : Any = TaForConditionalGeneration._keep_in_fpaa_modules lowerCAmelCase__ : Tuple = None # test with `t5-small` lowerCAmelCase__ : Union[str, Any] = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=a , device_map='auto' ) lowerCAmelCase__ : Optional[Any] = self.tokenizer(self.input_text , return_tensors='pt' ).to(0 ) lowerCAmelCase__ : Any = model.generate(**a ) # test with `flan-t5-small` lowerCAmelCase__ : List[Any] = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=a , device_map='auto' ) lowerCAmelCase__ : Optional[int] = self.tokenizer(self.input_text , return_tensors='pt' ).to(0 ) lowerCAmelCase__ : Optional[int] = model.generate(**a ) lowerCAmelCase__ : Any = modules def _lowerCamelCase ( self : List[str] ): '''simple docstring''' import bitsandbytes as bnb from transformers import TaForConditionalGeneration # test with `t5-small` lowerCAmelCase__ : Tuple = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=a , device_map='auto' ) # there was a bug with decoders - this test checks that it is fixed self.assertTrue(isinstance(model.decoder.block[0].layer[0].SelfAttention.q , bnb.nn.Linearabit ) ) lowerCAmelCase__ : List[Any] = self.tokenizer(self.input_text , return_tensors='pt' ).to(0 ) lowerCAmelCase__ : Optional[Any] = model.generate(**a ) # test with `flan-t5-small` lowerCAmelCase__ : Any = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=a , device_map='auto' ) lowerCAmelCase__ : Tuple = self.tokenizer(self.input_text , return_tensors='pt' ).to(0 ) lowerCAmelCase__ : Optional[int] = model.generate(**a ) class A__ ( __magic_name__ ): '''simple docstring''' def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().setUp() # model_name lowerCAmelCase__ : Dict = 'bigscience/bloom-560m' lowerCAmelCase__ : int = 't5-small' # Different types of model lowerCAmelCase__ : str = AutoModel.from_pretrained(self.model_name , load_in_abit=a , device_map='auto' ) # Sequence classification model lowerCAmelCase__ : Optional[int] = AutoModelForSequenceClassification.from_pretrained( self.model_name , load_in_abit=a , device_map='auto' ) # CausalLM model lowerCAmelCase__ : Optional[int] = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=a , device_map='auto' ) # Seq2seq model lowerCAmelCase__ : str = AutoModelForSeqaSeqLM.from_pretrained( self.seq_to_seq_name , load_in_abit=a , device_map='auto' ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' del self.base_model del self.sequence_model del self.model_abit del self.seq_to_seq_model gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' from bitsandbytes.nn import Paramsabit self.assertTrue(self.base_model.h[-1].mlp.dense_ah_to_h.weight.__class__ == Paramsabit ) # Other heads should be nn.Parameter self.assertTrue(self.model_abit.lm_head.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.sequence_model.score.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.seq_to_seq_model.lm_head.weight.__class__ == torch.nn.Parameter ) class A__ ( __magic_name__ ): '''simple docstring''' def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' del self.pipe gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = pipeline( 'text-generation' , model=self.model_name , model_kwargs={'device_map': 'auto', 'load_in_4bit': True, 'torch_dtype': torch.floataa} , max_new_tokens=self.MAX_NEW_TOKENS , ) # Real second forward pass lowerCAmelCase__ : Optional[int] = self.pipe(self.input_text ) self.assertIn(pipeline_output[0]['generated_text'] , self.EXPECTED_OUTPUTS ) @require_torch_multi_gpu class A__ ( __magic_name__ ): '''simple docstring''' def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().setUp() def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = AutoModelForCausalLM.from_pretrained( self.model_name , load_in_abit=a , device_map='balanced' ) # Check correct device map self.assertEqual(set(model_parallel.hf_device_map.values() ) , {0, 1} ) # Check that inference pass works on the model lowerCAmelCase__ : Optional[int] = self.tokenizer(self.input_text , return_tensors='pt' ) # Second real batch lowerCAmelCase__ : Any = model_parallel.generate(input_ids=encoded_input['input_ids'].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_parallel[0] , skip_special_tokens=a ) , self.EXPECTED_OUTPUTS ) class A__ ( __magic_name__ ): '''simple docstring''' def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'facebook/opt-350m' super().setUp() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if version.parse(importlib.metadata.version('bitsandbytes' ) ) < version.parse('0.37.0' ): return # Step 1: freeze all parameters lowerCAmelCase__ : List[Any] = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=a ) self.assertEqual(set(model.hf_device_map.values() ) , {torch.cuda.current_device()} ) for param in model.parameters(): lowerCAmelCase__ : Any = False # freeze the model - train adapters later if param.ndim == 1: # cast the small parameters (e.g. layernorm) to fp32 for stability lowerCAmelCase__ : List[str] = param.data.to(torch.floataa ) # Step 2: add adapters for _, module in model.named_modules(): if "OPTAttention" in repr(type(a ) ): lowerCAmelCase__ : Optional[int] = LoRALayer(module.q_proj , rank=16 ) lowerCAmelCase__ : Optional[int] = LoRALayer(module.k_proj , rank=16 ) lowerCAmelCase__ : Any = LoRALayer(module.v_proj , rank=16 ) # Step 3: dummy batch lowerCAmelCase__ : List[Any] = self.tokenizer('Test batch ' , return_tensors='pt' ).to(0 ) # Step 4: Check if the gradient is not None with torch.cuda.amp.autocast(): lowerCAmelCase__ : Tuple = model.forward(**a ) out.logits.norm().backward() for module in model.modules(): if isinstance(a , a ): self.assertTrue(module.adapter[1].weight.grad is not None ) self.assertTrue(module.adapter[1].weight.grad.norm().item() > 0 ) elif isinstance(a , nn.Embedding ): self.assertTrue(module.weight.grad is None ) class A__ ( __magic_name__ ): '''simple docstring''' lowercase = 'gpt2-xl' lowercase = 3.31_91_85_48_54_15_21_87
707
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ , lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
69
0
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (KDPMaDiscreteScheduler,) lowercase = 10 def _lowerCamelCase ( self : Tuple , **a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = { 'num_train_timesteps': 1_100, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', } config.update(**a ) return config def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : str ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_0_1, 0.0_0_0_1, 0.0_0_1] , [0.0_0_0_2, 0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : Any = scheduler_class(**a ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : Tuple = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase__ : List[Any] = sample.to(a ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase__ : Tuple = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : int = model(a , a ) lowerCAmelCase__ : Union[str, Any] = scheduler.step(a , a , a ) lowerCAmelCase__ : Optional[int] = output.prev_sample lowerCAmelCase__ : Any = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : int = torch.mean(torch.abs(a ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6934E-07 ) < 1E-2 assert abs(result_mean.item() - 6.1112E-10 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.693428650170972E-07 ) < 1E-2 assert abs(result_mean.item() - 0.0_0_0_2 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' if torch_device == "mps": return lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : str = scheduler_class(**a ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase__ : List[Any] = self.dummy_model() lowerCAmelCase__ : List[Any] = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase__ : List[Any] = sample.to(a ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase__ : List[Any] = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : Optional[int] = model(a , a ) lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a ) lowerCAmelCase__ : Optional[int] = output.prev_sample lowerCAmelCase__ : Tuple = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : str = torch.mean(torch.abs(a ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3 def _lowerCamelCase ( self : Tuple ): '''simple docstring''' if torch_device == "mps": return lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Tuple = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) scheduler.set_timesteps(self.num_inference_steps , device=a ) lowerCAmelCase__ : List[Any] = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter.to(a ) * scheduler.init_noise_sigma for t in scheduler.timesteps: lowerCAmelCase__ : List[Any] = scheduler.scale_model_input(a , a ) lowerCAmelCase__ : str = model(a , a ) lowerCAmelCase__ : Dict = scheduler.step(a , a , a ) lowerCAmelCase__ : str = output.prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[int] = torch.mean(torch.abs(a ) ) if str(a ).startswith('cpu' ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3
708
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise TypeError('Input value must be an \'int\' type' ) lowerCAmelCase__ : List[Any] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
709
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """google/rembert""": 256, } lowerCamelCase__ = """▁""" class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = RemBertTokenizer def __init__( self : Optional[Any] , a : str=None , a : Any=None , a : List[Any]=True , a : str=True , a : Dict=False , a : Dict="[CLS]" , a : int="[SEP]" , a : Tuple="<unk>" , a : Optional[Any]="[SEP]" , a : Tuple="<pad>" , a : Dict="[CLS]" , a : Optional[Any]="[MASK]" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( a , tokenizer_file=a , do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : int = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : List[Any] = keep_accents lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> float: lowerCAmelCase__ : Union[str, Any] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError('All input parameters must be positive' ) if any(p > 1 for p in parameters[1:4] ): raise ValueError('Relative densities cannot be greater than one' ) else: lowerCAmelCase__ : List[Any] = 1 - (matter_density + radiation_density + dark_energy) lowerCAmelCase__ : List[str] = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) lowerCAmelCase__ : Optional[Any] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation lowerCamelCase__ = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1E-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
710
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : int = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : List[Any] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : int = None lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Dict = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Union[str, Any] = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
69
0
from __future__ import annotations from PIL import Image # Define glider example lowerCamelCase__ = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example lowerCamelCase__ = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list[list[int]]: lowerCAmelCase__ : List[Any] = [] for i in range(len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ : Optional[int] = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours lowerCAmelCase__ : Optional[int] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(SCREAMING_SNAKE_CASE_ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(SCREAMING_SNAKE_CASE_ ) - 1: neighbour_count += cells[i + 1][j] if i < len(SCREAMING_SNAKE_CASE_ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. lowerCAmelCase__ : Dict = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(SCREAMING_SNAKE_CASE_ ) return next_generation def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> list[Image.Image]: lowerCAmelCase__ : Optional[int] = [] for _ in range(SCREAMING_SNAKE_CASE_ ): # Create output image lowerCAmelCase__ : Any = Image.new('RGB' , (len(cells[0] ), len(SCREAMING_SNAKE_CASE_ )) ) lowerCAmelCase__ : str = img.load() # Save cells to image for x in range(len(SCREAMING_SNAKE_CASE_ ) ): for y in range(len(cells[0] ) ): lowerCAmelCase__ : Optional[int] = 255 - cells[y][x] * 255 lowerCAmelCase__ : Optional[int] = (colour, colour, colour) # Save image images.append(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[int] = new_generation(SCREAMING_SNAKE_CASE_ ) return images if __name__ == "__main__": lowerCamelCase__ = generate_images(GLIDER, 16) images[0].save("""out.gif""", save_all=True, append_images=images[1:])
711
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> float: lowerCAmelCase__ : Union[str, Any] = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def lowerCAmelCase__ ( ) -> Union[str, Any]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
712
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
0
from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class A__ ( __magic_name__ ): lowercase = 42 lowercase = 42 if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
713
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
0
from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class A__ ( __magic_name__ ): def __init__( self : int , a : NestedDataStructureLike[PathLike] , a : Optional[NamedSplit] = None , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( a , split=a , features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : Dict = path_or_paths if isinstance(a , a ) else {self.split: path_or_paths} lowerCAmelCase__ : Optional[Any] = Text( cache_dir=a , data_files=a , features=a , **a , ) def _lowerCamelCase ( self : int ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : int = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Dict = None lowerCAmelCase__ : Optional[Any] = None lowerCAmelCase__ : str = None lowerCAmelCase__ : int = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Dict = self.builder.as_dataset( split=self.split , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
714
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
0
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = KandinskyVaaPriorPipeline lowercase = ['prompt'] lowercase = ['prompt', 'negative_prompt'] lowercase = [ 'num_images_per_prompt', 'generator', 'num_inference_steps', 'latents', 'negative_prompt', 'guidance_scale', 'output_type', 'return_dict', ] lowercase = False @property def _lowerCamelCase ( self : int ): '''simple docstring''' return 32 @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return 32 @property def _lowerCamelCase ( self : str ): '''simple docstring''' return self.time_input_dim @property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return self.time_input_dim * 4 @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return 100 @property def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Dict = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) return CLIPTextModelWithProjection(a ) @property def _lowerCamelCase ( self : Tuple ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : List[str] = { 'num_attention_heads': 2, 'attention_head_dim': 12, 'embedding_dim': self.text_embedder_hidden_size, 'num_layers': 1, } lowerCAmelCase__ : Union[str, Any] = PriorTransformer(**a ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 lowerCAmelCase__ : Optional[int] = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : str = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) lowerCAmelCase__ : str = CLIPVisionModelWithProjection(a ) return model @property def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Dict = CLIPImageProcessor( crop_size=224 , do_center_crop=a , do_normalize=a , do_resize=a , image_mean=[0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] , image_std=[0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] , resample=3 , size=224 , ) return image_processor def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.dummy_prior lowerCAmelCase__ : List[str] = self.dummy_image_encoder lowerCAmelCase__ : List[Any] = self.dummy_text_encoder lowerCAmelCase__ : Tuple = self.dummy_tokenizer lowerCAmelCase__ : Optional[Any] = self.dummy_image_processor lowerCAmelCase__ : Any = UnCLIPScheduler( variance_type='fixed_small_log' , prediction_type='sample' , num_train_timesteps=1_000 , clip_sample=a , clip_sample_range=10.0 , ) lowerCAmelCase__ : int = { 'prior': prior, 'image_encoder': image_encoder, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'scheduler': scheduler, 'image_processor': image_processor, } return components def _lowerCamelCase ( self : List[str] , a : Dict , a : List[str]=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : int = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Dict = { 'prompt': 'horse', 'generator': generator, 'guidance_scale': 4.0, 'num_inference_steps': 2, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Any = self.pipeline_class(**a ) lowerCAmelCase__ : Union[str, Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = pipe(**self.get_dummy_inputs(a ) ) lowerCAmelCase__ : List[Any] = output.image_embeds lowerCAmelCase__ : Optional[Any] = pipe( **self.get_dummy_inputs(a ) , return_dict=a , )[0] lowerCAmelCase__ : Any = image[0, -10:] lowerCAmelCase__ : Dict = image_from_tuple[0, -10:] assert image.shape == (1, 32) lowerCAmelCase__ : str = np.array( [-0.0_5_3_2, 1.7_1_2_0, 0.3_6_5_6, -1.0_8_5_2, -0.8_9_4_6, -1.1_7_5_6, 0.4_3_4_8, 0.2_4_8_2, 0.5_1_4_6, -0.1_1_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = torch_device == 'cpu' lowerCAmelCase__ : Optional[Any] = True lowerCAmelCase__ : Any = False self._test_inference_batch_single_identical( test_max_difference=a , relax_max_difference=a , test_mean_pixel_difference=a , ) @skip_mps def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = torch_device == 'cpu' lowerCAmelCase__ : int = False self._test_attention_slicing_forward_pass( test_max_difference=a , test_mean_pixel_difference=a , )
715
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
0
'''simple docstring''' import random def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = False ) -> dict: lowerCAmelCase__ : dict = {i: [] for i in range(SCREAMING_SNAKE_CASE_ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(SCREAMING_SNAKE_CASE_ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(SCREAMING_SNAKE_CASE_ ): for j in range(i + 1 , SCREAMING_SNAKE_CASE_ ): if random.random() < probability: graph[i].append(SCREAMING_SNAKE_CASE_ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(SCREAMING_SNAKE_CASE_ ) return graph def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: return { i: [j for j in range(SCREAMING_SNAKE_CASE_ ) if i != j] for i in range(SCREAMING_SNAKE_CASE_ ) } if __name__ == "__main__": import doctest doctest.testmod()
716
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> list: lowerCAmelCase__ : Tuple = len(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Union[str, Any] = [] for i in range(len(SCREAMING_SNAKE_CASE_ ) - pat_len + 1 ): lowerCAmelCase__ : List[Any] = True for j in range(SCREAMING_SNAKE_CASE_ ): if s[i + j] != pattern[j]: lowerCAmelCase__ : Optional[Any] = False break if match_found: position.append(SCREAMING_SNAKE_CASE_ ) return position if __name__ == "__main__": assert naive_pattern_search("""ABCDEFG""", """DE""") == [3] print(naive_pattern_search("""ABAAABCDBBABCDDEBCABC""", """ABC"""))
717
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
'''simple docstring''' import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class A__ ( unittest.TestCase , __magic_name__ ): def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[str] = load_tool('text-to-speech' ) self.tool.setup() def _lowerCamelCase ( self : Tuple ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : str = self.tool('hey' ) lowerCAmelCase__ : Dict = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_0_0_5_9_6_6_6_6_8_8_3_2_1_1_5_8_2_9, -0.0_0_0_3_6_5_7_6_4_0_1_9_0_7_9_5_0_6_4, -0.0_0_0_1_3_4_3_9_5_0_2_7_9_9_8_8_3_4_8_5] ) , ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : str = self.tool('hey' ) lowerCAmelCase__ : List[str] = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_0_0_5_9_6_6_6_6_8_8_3_2_1_1_5_8_2_9, -0.0_0_0_3_6_5_7_6_4_0_1_9_0_7_9_5_0_6_4, -0.0_0_0_1_3_4_3_9_5_0_2_7_9_9_8_8_3_4_8_5] ) , ) )
718
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: if len(SCREAMING_SNAKE_CASE_ ) != len(SCREAMING_SNAKE_CASE_ ): raise ValueError('String lengths must match!' ) lowerCAmelCase__ : Tuple = 0 for chara, chara in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
719
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError('The given input must be positive' ) # get the generated string sequence lowerCAmelCase__ : Any = gray_code_sequence_string(SCREAMING_SNAKE_CASE_ ) # # convert them to integers for i in range(len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ : Any = int(sequence[i] , 2 ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] lowerCAmelCase__ : Optional[Any] = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits lowerCAmelCase__ : Optional[Any] = gray_code_sequence_string(bit_count - 1 ) lowerCAmelCase__ : Union[str, Any] = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): lowerCAmelCase__ : Optional[Any] = '0' + smaller_sequence[i] sequence.append(SCREAMING_SNAKE_CASE_ ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): lowerCAmelCase__ : int = '1' + smaller_sequence[i] sequence.append(SCREAMING_SNAKE_CASE_ ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
720
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
0
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['input_features', 'attention_mask'] def __init__( self : Optional[Any] , a : Union[str, Any]=80 , a : List[Any]=16_000 , a : Union[str, Any]=0.0 , a : int=10 , a : Any=25 , a : str="hamming_window" , a : str=32_768.0 , a : Any=0.9_7 , a : Union[str, Any]=1.0 , a : Optional[int]=True , a : Optional[int]=True , a : int=False , **a : Dict , ): '''simple docstring''' super().__init__(feature_size=a , sampling_rate=a , padding_value=a , **a ) lowerCAmelCase__ : Optional[Any] = feature_size lowerCAmelCase__ : Any = sampling_rate lowerCAmelCase__ : int = padding_value lowerCAmelCase__ : Optional[Any] = hop_length lowerCAmelCase__ : Any = win_length lowerCAmelCase__ : List[str] = frame_signal_scale lowerCAmelCase__ : Union[str, Any] = preemphasis_coeff lowerCAmelCase__ : Tuple = mel_floor lowerCAmelCase__ : List[str] = normalize_means lowerCAmelCase__ : int = normalize_vars lowerCAmelCase__ : List[Any] = win_function lowerCAmelCase__ : int = return_attention_mask lowerCAmelCase__ : int = win_length * sampling_rate // 1_000 lowerCAmelCase__ : Tuple = hop_length * sampling_rate // 1_000 lowerCAmelCase__ : Tuple = optimal_fft_length(self.sample_size ) lowerCAmelCase__ : Any = (self.n_fft // 2) + 1 def _lowerCamelCase ( self : Optional[Any] , a : np.array ): '''simple docstring''' if self.win_function == "hamming_window": lowerCAmelCase__ : Optional[int] = window_function(window_length=self.sample_size , name=self.win_function , periodic=a ) else: lowerCAmelCase__ : Tuple = window_function(window_length=self.sample_size , name=self.win_function ) lowerCAmelCase__ : Tuple = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) lowerCAmelCase__ : Any = spectrogram( one_waveform * self.frame_signal_scale , window=a , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=a , preemphasis=self.preemphasis_coeff , mel_filters=a , mel_floor=self.mel_floor , log_mel='log' , ) return msfc_features.T def _lowerCamelCase ( self : Optional[Any] , a : int , a : Union[str, Any] , a : List[str] ): '''simple docstring''' if self.normalize_means: lowerCAmelCase__ : Dict = x[:input_length].mean(axis=0 ) lowerCAmelCase__ : List[Any] = np.subtract(a , a ) if self.normalize_vars: lowerCAmelCase__ : Union[str, Any] = x[:input_length].std(axis=0 ) lowerCAmelCase__ : Any = np.divide(a , a ) if input_length < x.shape[0]: lowerCAmelCase__ : List[str] = padding_value # make sure array is in float32 lowerCAmelCase__ : Optional[int] = x.astype(np.floataa ) return x def _lowerCamelCase ( self : Union[str, Any] , a : List[np.ndarray] , a : Optional[np.ndarray] = None ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(a , a , self.padding_value ) for x, n in zip(a , a )] def __call__( self : Union[str, Any] , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Union[bool, str, PaddingStrategy] = False , a : Optional[int] = None , a : bool = False , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[Union[str, TensorType]] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the ``sampling_rate`` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : int = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : int = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : str = [np.asarray(a , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Dict = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : Optional[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : str = [raw_speech] # extract fbank features lowerCAmelCase__ : Dict = [self._extract_mfsc_features(a ) for one_waveform in raw_speech] # convert into correct format for padding lowerCAmelCase__ : List[str] = BatchFeature({'input_features': features} ) lowerCAmelCase__ : Optional[Any] = self.pad( a , padding=a , max_length=a , truncation=a , pad_to_multiple_of=a , return_attention_mask=a , **a , ) # make sure list is in array format lowerCAmelCase__ : List[str] = padded_inputs.get('input_features' ) if isinstance(input_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in input_features] lowerCAmelCase__ : Optional[Any] = padded_inputs.get('attention_mask' ) if attention_mask is not None: lowerCAmelCase__ : List[Any] = [np.asarray(a , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCAmelCase__ : str = ( np.array(a , dtype=np.intaa ) if self._get_padding_strategies(a , max_length=a ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCAmelCase__ : Tuple = self.normalize( padded_inputs['input_features'] , attention_mask=a ) if return_tensors is not None: lowerCAmelCase__ : List[str] = padded_inputs.convert_to_tensors(a ) return padded_inputs
721
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
69
0
lowerCamelCase__ = """Alexander Joslin""" import operator as op from .stack import Stack def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : Union[str, Any] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} lowerCAmelCase__ : Stack[int] = Stack() lowerCAmelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE_ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE_ ) elif i == ")": # RULE 4 lowerCAmelCase__ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : List[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : Tuple = operators[opr](SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) operand_stack.push(SCREAMING_SNAKE_CASE_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
700
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A__ ( __magic_name__ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'LayoutLMv3ImageProcessor' lowercase = ('LayoutLMv3Tokenizer', 'LayoutLMv3TokenizerFast') def __init__( self : Optional[int] , a : Union[str, Any]=None , a : Optional[Any]=None , **a : str ): '''simple docstring''' lowerCAmelCase__ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , a , ) lowerCAmelCase__ : int = kwargs.pop('feature_extractor' ) lowerCAmelCase__ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(a , a ) def __call__( self : List[Any] , a : List[Any] , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a : Union[List[List[int]], List[List[List[int]]]] = None , a : Optional[Union[List[int], List[List[int]]]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : str , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor lowerCAmelCase__ : List[str] = self.image_processor(images=a , return_tensors=a ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a , a ): lowerCAmelCase__ : Optional[Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) lowerCAmelCase__ : List[str] = features['words'] lowerCAmelCase__ : List[Any] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_token_type_ids=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_length=a , verbose=a , return_tensors=a , **a , ) # add pixel values lowerCAmelCase__ : Tuple = features.pop('pixel_values' ) if return_overflowing_tokens is True: lowerCAmelCase__ : List[str] = self.get_overflowing_images(a , encoded_inputs['overflow_to_sample_mapping'] ) lowerCAmelCase__ : List[str] = images return encoded_inputs def _lowerCamelCase ( self : Any , a : List[str] , a : int ): '''simple docstring''' lowerCAmelCase__ : int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a ) != len(a ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' f''' {len(a )} and {len(a )}''' ) return images_with_overflow def _lowerCamelCase ( self : Union[str, Any] , *a : Optional[Any] , **a : List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*a , **a ) def _lowerCamelCase ( self : Tuple , *a : List[str] , **a : Optional[Any] ): '''simple docstring''' return self.tokenizer.decode(*a , **a ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , a , ) return self.image_processor_class @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , a , ) return self.image_processor
69
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCamelCase__ = {"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
701
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
69
0
import numpy as np def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> np.array: return 1 / (1 + np.exp(-vector )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> np.array: return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
702
lowerCamelCase__ = """Alexander Joslin""" import operator as op from .stack import Stack def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Union[str, Any] = {'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} lowerCAmelCase__ : Stack[int] = Stack() lowerCAmelCase__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(SCREAMING_SNAKE_CASE_ ) ) elif i in operators: # RULE 2 operator_stack.push(SCREAMING_SNAKE_CASE_ ) elif i == ")": # RULE 4 lowerCAmelCase__ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase__ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : List[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase__ : Tuple = operators[opr](SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) operand_stack.push(SCREAMING_SNAKE_CASE_ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
69
0
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): def __init__( self : List[Any] , *a : str , **a : Union[str, Any] ): '''simple docstring''' warnings.warn( 'The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use MobileViTImageProcessor instead.' , a , ) super().__init__(*a , **a )
703
import numpy class A__ : def __init__( self : Tuple , a : numpy.ndarray , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : int = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Dict = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : List[Any] = numpy.random.rand(3 , 1 ) # Real output values provided. lowerCAmelCase__ : str = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : List[Any] = numpy.zeros(output_array.shape ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) lowerCAmelCase__ : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) lowerCAmelCase__ : int = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowerCamelCase ( self : Optional[int] , a : numpy.ndarray , a : int , a : bool ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): lowerCAmelCase__ : Any = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Tuple = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def _lowerCamelCase ( self : Optional[Any] , a : numpy.ndarray ): '''simple docstring''' lowerCAmelCase__ : Dict = input_arr lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> numpy.ndarray: return (value) * (1 - (value)) def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : int = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : List[str] = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ , output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ , iterations=10 , give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
69
0
# Lint as: python3 import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union lowerCamelCase__ = re.compile(r"""^(?P<major>\d+)""" r"""\.(?P<minor>\d+)""" r"""\.(?P<patch>\d+)$""") @total_ordering @dataclass class A__ : lowercase = 42 lowercase = None lowercase = None lowercase = None lowercase = None def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = _str_to_version_tuple(self.version_str ) def __repr__( self : Tuple ): '''simple docstring''' return f'''{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}''' @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return self.major, self.minor, self.patch def _lowerCamelCase ( self : int , a : str ): '''simple docstring''' if isinstance(a , a ): return Version(a ) elif isinstance(a , a ): return other raise TypeError(f'''{other} (type {type(a )}) cannot be compared to version.''' ) def __eq__( self : Optional[Any] , a : Tuple ): '''simple docstring''' try: lowerCAmelCase__ : Optional[Any] = self._validate_operand(a ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self : Optional[int] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : str = self._validate_operand(a ) return self.tuple < other.tuple def __hash__( self : List[Any] ): '''simple docstring''' return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def _lowerCamelCase ( cls : Optional[int] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = {f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def _lowerCamelCase ( self : int ): '''simple docstring''' return self.version_str def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : Optional[int] = _VERSION_REG.match(SCREAMING_SNAKE_CASE_ ) if not res: raise ValueError(F'''Invalid version \'{version_str}\'. Format should be x.y.z with {{x,y,z}} being digits.''' ) return tuple(int(SCREAMING_SNAKE_CASE_ ) for v in [res.group('major' ), res.group('minor' ), res.group('patch' )] ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Optional[int]: return ".".join(str(SCREAMING_SNAKE_CASE_ ) for v in version_tuple )
704
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A__ : def __init__( self : int , a : str , a : Union[str, Any]=13 , a : int=32 , a : Optional[Any]=2 , a : Tuple=3 , a : List[Any]=16 , a : List[str]=[1, 2, 1] , a : int=[2, 2, 4] , a : int=2 , a : Optional[Any]=2.0 , a : Optional[int]=True , a : Dict=0.0 , a : Any=0.0 , a : int=0.1 , a : List[str]="gelu" , a : Optional[Any]=False , a : str=True , a : Dict=0.0_2 , a : Any=1E-5 , a : Optional[int]=True , a : str=None , a : str=True , a : int=10 , a : str=8 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : Optional[Any] = patch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = embed_dim lowerCAmelCase__ : Tuple = depths lowerCAmelCase__ : List[str] = num_heads lowerCAmelCase__ : List[Any] = window_size lowerCAmelCase__ : Any = mlp_ratio lowerCAmelCase__ : Optional[Any] = qkv_bias lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : int = drop_path_rate lowerCAmelCase__ : Optional[Any] = hidden_act lowerCAmelCase__ : int = use_absolute_embeddings lowerCAmelCase__ : List[str] = patch_norm lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : List[Any] = scope lowerCAmelCase__ : Dict = use_labels lowerCAmelCase__ : List[Any] = type_sequence_label_size lowerCAmelCase__ : Optional[Any] = encoder_stride def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[Any] = None if self.use_labels: lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : int = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : List[str] , a : Any , a : str , a : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = SwinvaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model(a ) lowerCAmelCase__ : str = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase__ : Tuple = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _lowerCamelCase ( self : Union[str, Any] , a : Optional[Any] , a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SwinvaForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model(a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : Dict = SwinvaForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : List[str] = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Union[str, Any] , a : int , a : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.type_sequence_label_size lowerCAmelCase__ : List[Any] = SwinvaForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = config_and_inputs lowerCAmelCase__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = SwinvaModelTester(self ) lowerCAmelCase__ : int = ConfigTester(self , config_class=a , embed_dim=37 ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[int] = model_class(a ) lowerCAmelCase__ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : Tuple = [*signature.parameters.keys()] lowerCAmelCase__ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = True for model_class in self.all_model_classes: lowerCAmelCase__ : Tuple = True lowerCAmelCase__ : str = False lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Dict = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : int = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Dict = outputs.attentions lowerCAmelCase__ : Dict = len(self.model_tester.depths ) self.assertEqual(len(a ) , a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Optional[int] = config.window_size**2 lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowerCAmelCase__ : Tuple = len(a ) # Check attention is always last and order is fine lowerCAmelCase__ : str = True lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(**self._prepare_for_class(a , a ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): lowerCAmelCase__ : Optional[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase__ : Any = 2 self.assertEqual(out_len + added_hidden_states , len(a ) ) lowerCAmelCase__ : Dict = outputs.attentions self.assertEqual(len(a ) , a ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _lowerCamelCase ( self : int , a : Optional[int] , a : int , a : Optional[Any] , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowerCAmelCase__ : Optional[Any] = model(**self._prepare_for_class(a , a ) ) lowerCAmelCase__ : Optional[Any] = outputs.hidden_states lowerCAmelCase__ : str = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a ) , a ) # Swinv2 has a different seq_length lowerCAmelCase__ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : List[str] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowerCAmelCase__ : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(a ) , a ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Any = reshaped_hidden_states[0].shape lowerCAmelCase__ : List[str] = ( reshaped_hidden_states[0].view(a , a , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : str = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : List[str] = True self.check_hidden_states_output(a , a , a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Any = 3 lowerCAmelCase__ : int = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase__ : str = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ : str = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase__ : Any = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase__ : str = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ : Any = True self.check_hidden_states_output(a , a , a , (padded_height, padded_width) ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : List[str] = SwinvaModel.from_pretrained(a ) self.assertIsNotNone(a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ : Optional[int] = _config_zero_init(a ) for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(config=a ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Tuple = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( a ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ : Any = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**a ) # verify the logits lowerCAmelCase__ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Optional[Any] = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) )
69
0
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowerCamelCase__ = get_tests_dir("""fixtures/spiece.model""") @require_sentencepiece @require_tokenizers class A__ ( __magic_name__ , unittest.TestCase ): lowercase = AlbertTokenizer lowercase = AlbertTokenizerFast lowercase = True lowercase = True lowercase = True def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase__ : Dict = AlbertTokenizer(a ) tokenizer.save_pretrained(self.tmpdirname ) def _lowerCamelCase ( self : Any , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = 'this is a test' lowerCAmelCase__ : List[str] = 'this is a test' return input_text, output_text def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : int = '<pad>' lowerCAmelCase__ : Tuple = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a ) , a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a ) , a ) def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<pad>' ) self.assertEqual(vocab_keys[1] , '<unk>' ) self.assertEqual(vocab_keys[-1] , '▁eloquent' ) self.assertEqual(len(a ) , 30_000 ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 30_000 ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' if not self.test_rust_tokenizer: return lowerCAmelCase__ : List[str] = self.get_tokenizer() lowerCAmelCase__ : List[str] = self.get_rust_tokenizer() lowerCAmelCase__ : Optional[int] = 'I was born in 92000, and this is falsé.' lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) lowerCAmelCase__ : Optional[Any] = rust_tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Any = tokenizer.encode(a , add_special_tokens=a ) lowerCAmelCase__ : List[str] = rust_tokenizer.encode(a , add_special_tokens=a ) self.assertListEqual(a , a ) lowerCAmelCase__ : List[str] = self.get_rust_tokenizer() lowerCAmelCase__ : Dict = tokenizer.encode(a ) lowerCAmelCase__ : List[Any] = rust_tokenizer.encode(a ) self.assertListEqual(a , a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[str] = AlbertTokenizer(a , keep_accents=a ) lowerCAmelCase__ : Optional[int] = tokenizer.tokenize('This is a test' ) self.assertListEqual(a , ['▁this', '▁is', '▁a', '▁test'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , [48, 25, 21, 1_289] ) lowerCAmelCase__ : Dict = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( a , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.'] ) lowerCAmelCase__ : Any = tokenizer.convert_tokens_to_ids(a ) self.assertListEqual(a , [31, 23, 386, 19, 561, 3_050, 15, 17, 48, 25, 8_256, 18, 1, 9] ) lowerCAmelCase__ : Union[str, Any] = tokenizer.convert_ids_to_tokens(a ) self.assertListEqual( a , ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.'] , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = AlbertTokenizer(a ) lowerCAmelCase__ : Tuple = tokenizer.encode('sequence builders' ) lowerCAmelCase__ : str = tokenizer.encode('multi-sequence build' ) lowerCAmelCase__ : Any = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : Dict = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = {'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'input_ids': [[2, 21_970, 13, 5, 6_092, 167, 28, 7_103, 2_153, 673, 8, 7_028, 12_051, 18, 17, 7_103, 2_153, 673, 8, 3_515, 18_684, 8, 4_461, 6, 1_927, 297, 8, 12_060, 2_607, 18, 13, 5, 4_461, 15, 10_538, 38, 8, 135, 15, 822, 58, 15, 993, 10_363, 15, 1_460, 8_005, 4_461, 15, 993, 255, 2_328, 9, 9, 9, 6, 26, 1_112, 816, 3_260, 13, 5, 103, 2_377, 6, 17, 1_112, 816, 2_782, 13, 5, 103, 10_641, 6, 29, 84, 2_512, 2_430, 782, 18_684, 2_761, 19, 808, 2_430, 2_556, 17, 855, 1_480, 9_477, 4_091, 128, 11_712, 15, 7_103, 2_153, 673, 17, 24_883, 9_990, 9, 3], [2, 11_502, 25, 1_006, 20, 782, 8, 11_809, 855, 1_732, 19_393, 18_667, 37, 367, 21_018, 69, 1_854, 34, 11_860, 19_124, 27, 156, 225, 17, 193, 4_141, 19, 65, 9_124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2_231, 886, 2_385, 17_659, 84, 14, 16_792, 1_952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a , model_name='albert-base-v2' , revision='6b6560eaf5ff2e250b00c50f380c5389a9c2d82e' , )
705
from itertools import permutations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> bool: if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase__ : str = [7, 11, 13, 17] for i, test in enumerate(SCREAMING_SNAKE_CASE_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = 10 ) -> int: return sum( int(''.join(map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) ) for num in permutations(range(SCREAMING_SNAKE_CASE_ ) ) if is_substring_divisible(SCREAMING_SNAKE_CASE_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
69
0
import os import jsonlines import numpy as np from tqdm import tqdm lowerCamelCase__ = 2048 lowerCamelCase__ = 4096 lowerCamelCase__ = 42 lowerCamelCase__ = os.environ.pop("""PROCESS_TRAIN""", """false""") lowerCamelCase__ = {"""null""": 0, """short""": 1, """long""": 2, """yes""": 3, """no""": 4} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: def choose_first(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): assert isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) == 1: lowerCAmelCase__ : Optional[Any] = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: lowerCAmelCase__ : Optional[int] = {k: [a[k]] for k in a} if len(a['start_token'] ) > 0: break return a lowerCAmelCase__ : Dict = {'id': example['id']} lowerCAmelCase__ : Tuple = example['annotations'] lowerCAmelCase__ : int = annotation['yes_no_answer'] if 0 in yes_no_answer or 1 in yes_no_answer: lowerCAmelCase__ : int = ['yes'] if 1 in yes_no_answer else ['no'] lowerCAmelCase__ : Dict = [] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Any = ['<cls>'] else: lowerCAmelCase__ : Any = ['short'] lowerCAmelCase__ : Dict = choose_first(annotation['short_answers'] ) if len(out['start_token'] ) == 0: # answer will be long if short is not available lowerCAmelCase__ : List[str] = ['long'] lowerCAmelCase__ : List[str] = choose_first(annotation['long_answer'] , is_long_answer=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = [] answer.update(SCREAMING_SNAKE_CASE_ ) # disregard some samples if len(answer['start_token'] ) > 1 or answer["start_token"] == answer["end_token"]: lowerCAmelCase__ : Any = True else: lowerCAmelCase__ : Tuple = False lowerCAmelCase__ : Tuple = ['start_token', 'end_token', 'start_byte', 'end_byte', 'text'] if not all(isinstance(answer[k] , SCREAMING_SNAKE_CASE_ ) for k in cols ): raise ValueError('Issue in ID' , example['id'] ) return answer def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ) -> Union[str, Any]: lowerCAmelCase__ : int = _get_single_answer(SCREAMING_SNAKE_CASE_ ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element lowerCAmelCase__ : List[Any] = example['document']['tokens'] lowerCAmelCase__ : Tuple = [] for i in range(len(doc['token'] ) ): if not doc["is_html"][i]: context.append(doc['token'][i] ) return { "context": " ".join(SCREAMING_SNAKE_CASE_ ), "answer": { "start_token": -100, # ignore index in cross-entropy "end_token": -100, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples lowerCAmelCase__ : int = ['start_token', 'end_token'] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 lowerCAmelCase__ : Tuple = example['document']['tokens'] lowerCAmelCase__ : int = answer['start_token'] lowerCAmelCase__ : List[Any] = answer['end_token'] lowerCAmelCase__ : Tuple = [] for i in range(len(doc['token'] ) ): if not doc["is_html"][i]: context.append(doc['token'][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 lowerCAmelCase__ : Tuple = ' '.join(context[start_token:end_token] ) # checking above code if assertion: lowerCAmelCase__ : Union[str, Any] = doc['is_html'][answer['start_token'] : answer['end_token']] lowerCAmelCase__ : List[Any] = doc['token'][answer['start_token'] : answer['end_token']] lowerCAmelCase__ : Any = ' '.join([old[i] for i in range(len(SCREAMING_SNAKE_CASE_ ) ) if not is_html[i]] ) if new != old: print('ID:' , example['id'] ) print('New:' , SCREAMING_SNAKE_CASE_ , end='\n' ) print('Old:' , SCREAMING_SNAKE_CASE_ , end='\n\n' ) return { "context": " ".join(SCREAMING_SNAKE_CASE_ ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=2_048 , SCREAMING_SNAKE_CASE_=4_096 , SCREAMING_SNAKE_CASE_=True ) -> Dict: # overlap will be of doc_stride - q_len lowerCAmelCase__ : Optional[int] = get_context_and_ans(SCREAMING_SNAKE_CASE_ , assertion=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = out['answer'] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } lowerCAmelCase__ : int = tokenizer(example['question']['text'] , out['context'] ).input_ids lowerCAmelCase__ : List[str] = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : List[str] = [] lowerCAmelCase__ : Dict = input_ids[:q_len] lowerCAmelCase__ : List[Any] = range(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) , max_length - doc_stride ) for i in doc_start_indices: lowerCAmelCase__ : Tuple = i + max_length - q_len lowerCAmelCase__ : Optional[Any] = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer['category'][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-100] * len(SCREAMING_SNAKE_CASE_ ), "end_token": [-100] * len(SCREAMING_SNAKE_CASE_ ), "category": category, }, } lowerCAmelCase__ : List[str] = out['context'].split() lowerCAmelCase__ : int = splitted_context[answer['end_token']] lowerCAmelCase__ : List[str] = len( tokenizer( ' '.join(splitted_context[: answer['start_token']] ) , add_special_tokens=SCREAMING_SNAKE_CASE_ , ).input_ids ) lowerCAmelCase__ : int = len( tokenizer(' '.join(splitted_context[: answer['end_token']] ) , add_special_tokens=SCREAMING_SNAKE_CASE_ ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token lowerCAmelCase__ : List[Any] = len(tokenizer(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 lowerCAmelCase__ : Optional[Any] = input_ids[answer['start_token'] : answer['end_token'] + 1] # right & left are inclusive lowerCAmelCase__ : Optional[int] = answer['start_token'] lowerCAmelCase__ : Dict = answer['end_token'] if assertion: lowerCAmelCase__ : Union[str, Any] = tokenizer.decode(SCREAMING_SNAKE_CASE_ ) if answer["span"] != new: print('ISSUE IN TOKENIZATION' ) print('OLD:' , answer['span'] ) print('NEW:' , SCREAMING_SNAKE_CASE_ , end='\n\n' ) if len(SCREAMING_SNAKE_CASE_ ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } lowerCAmelCase__ : Optional[Any] = input_ids[:q_len] lowerCAmelCase__ : Union[str, Any] = range(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) , max_length - doc_stride ) lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Any = [] lowerCAmelCase__ : Dict = [] lowerCAmelCase__ : int = [] # null, yes, no, long, short for i in doc_start_indices: lowerCAmelCase__ : Optional[int] = i + max_length - q_len lowerCAmelCase__ : int = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: lowerCAmelCase__ : Optional[int] = start_token - i + q_len lowerCAmelCase__ : List[Any] = end_token - i + q_len answers_category.append(answer['category'][0] ) # ["short"] -> "short" else: lowerCAmelCase__ : int = -100 lowerCAmelCase__ : Union[str, Any] = -100 answers_category.append('null' ) lowerCAmelCase__ : List[str] = inputs[-1][start_token : end_token + 1] answers_start_token.append(SCREAMING_SNAKE_CASE_ ) answers_end_token.append(SCREAMING_SNAKE_CASE_ ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print('ISSUE in strided for ID:' , example['id'] ) print('New:' , tokenizer.decode(SCREAMING_SNAKE_CASE_ ) ) print('Old:' , tokenizer.decode(SCREAMING_SNAKE_CASE_ ) , end='\n\n' ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=2_048 , SCREAMING_SNAKE_CASE_=4_096 , SCREAMING_SNAKE_CASE_=False ) -> Union[str, Any]: lowerCAmelCase__ : Dict = get_strided_contexts_and_ans( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , doc_stride=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , assertion=SCREAMING_SNAKE_CASE_ , ) return example def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: with jsonlines.open(SCREAMING_SNAKE_CASE_ , 'a' ) as writer: for example in tqdm(SCREAMING_SNAKE_CASE_ , total=len(SCREAMING_SNAKE_CASE_ ) , desc='Saving samples ... ' ): lowerCAmelCase__ : Union[str, Any] = example['labels'] for ids, start, end, cat in zip( example['input_ids'] , labels['start_token'] , labels['end_token'] , labels['category'] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { 'input_ids': ids, 'start_token': start, 'end_token': end, 'category': CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer lowerCamelCase__ = load_dataset("""natural_questions""") lowerCamelCase__ = BigBirdTokenizer.from_pretrained("""google/bigbird-roberta-base""") lowerCamelCase__ = data["""train""" if PROCESS_TRAIN == """true""" else """validation"""] lowerCamelCase__ = { """tokenizer""": tokenizer, """doc_stride""": DOC_STRIDE, """max_length""": MAX_LENGTH, """assertion""": False, } lowerCamelCase__ = data.map(prepare_inputs, fn_kwargs=fn_kwargs) lowerCamelCase__ = data.remove_columns(["""annotations""", """document""", """id""", """question"""]) print(data) np.random.seed(SEED) lowerCamelCase__ = """nq-training.jsonl""" if PROCESS_TRAIN == """true""" else """nq-validation.jsonl""" save_to_disk(data, file_name=cache_file_name)
706
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A__ ( __magic_name__ , unittest.TestCase ): lowercase = ConsistencyModelPipeline lowercase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Dict = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def _lowerCamelCase ( self : Optional[Any] , a : Union[str, Any]=False ): '''simple docstring''' if class_cond: lowerCAmelCase__ : Tuple = self.dummy_cond_unet else: lowerCAmelCase__ : Dict = self.dummy_uncond_unet # Default to CM multistep sampler lowerCAmelCase__ : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def _lowerCamelCase ( self : int , a : Optional[int] , a : Any=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : str = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : str = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[Any] = self.get_dummy_components() lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_dummy_inputs(a ) lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : Tuple = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Tuple = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : Union[str, Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Dict = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Tuple = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ : Optional[int] = self.get_dummy_components(class_cond=a ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(**a ) lowerCAmelCase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Tuple = self.get_dummy_inputs(a ) lowerCAmelCase__ : Dict = 1 lowerCAmelCase__ : Tuple = None lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : str = pipe(**a ).images assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ : Union[str, Any] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Dict = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Optional[Any] , a : Tuple=0 , a : Optional[Any]=False , a : Optional[Any]="cpu" , a : Union[str, Any]=torch.floataa , a : Dict=(1, 3, 64, 64) ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(a ) lowerCAmelCase__ : List[Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: lowerCAmelCase__ : Optional[int] = self.get_fixed_latents(seed=a , device=a , dtype=a , shape=a ) lowerCAmelCase__ : Tuple = latents return inputs def _lowerCamelCase ( self : str , a : Tuple=0 , a : Tuple="cpu" , a : Tuple=torch.floataa , a : str=(1, 3, 64, 64) ): '''simple docstring''' if type(a ) == str: lowerCAmelCase__ : str = torch.device(a ) lowerCAmelCase__ : List[str] = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Any = randn_tensor(a , generator=a , device=a , dtype=a ) return latents def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : List[Any] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Optional[Any] = self.get_inputs() lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : List[str] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Optional[int] = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : List[str] = self.get_inputs() lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase__ : Union[str, Any] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Tuple = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : str = self.get_inputs(get_fixed_latents=a , device=a ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : Dict = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : str = image[0, -3:, -3:, -1] lowerCAmelCase__ : str = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2' ) lowerCAmelCase__ : List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.0_0_2 , sigma_max=8_0.0 , ) lowerCAmelCase__ : Dict = ConsistencyModelPipeline(unet=a , scheduler=a ) pipe.to(torch_device=a , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Any = self.get_inputs(get_fixed_latents=a , device=a ) lowerCAmelCase__ : List[str] = 1 lowerCAmelCase__ : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a , enable_math=a , enable_mem_efficient=a ): lowerCAmelCase__ : List[str] = pipe(**a ).images assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase__ : Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
69
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class A__ ( __magic_name__ ): '''simple docstring''' lowercase = 'openai-gpt' lowercase = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : str , a : List[str]=40_478 , a : Union[str, Any]=512 , a : List[str]=768 , a : str=12 , a : Dict=12 , a : str="gelu" , a : Dict=0.1 , a : Any=0.1 , a : Optional[int]=0.1 , a : str=1E-5 , a : Dict=0.0_2 , a : Any="cls_index" , a : Tuple=True , a : Union[str, Any]=None , a : Union[str, Any]=True , a : Any=0.1 , **a : int , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = vocab_size lowerCAmelCase__ : str = n_positions lowerCAmelCase__ : str = n_embd lowerCAmelCase__ : Dict = n_layer lowerCAmelCase__ : Any = n_head lowerCAmelCase__ : int = afn lowerCAmelCase__ : Dict = resid_pdrop lowerCAmelCase__ : Union[str, Any] = embd_pdrop lowerCAmelCase__ : int = attn_pdrop lowerCAmelCase__ : List[Any] = layer_norm_epsilon lowerCAmelCase__ : Union[str, Any] = initializer_range lowerCAmelCase__ : Optional[Any] = summary_type lowerCAmelCase__ : int = summary_use_proj lowerCAmelCase__ : Tuple = summary_activation lowerCAmelCase__ : Union[str, Any] = summary_first_dropout lowerCAmelCase__ : str = summary_proj_to_labels super().__init__(**a )
707
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A__ ( __magic_name__ ): def __init__( self : int , a : List[str] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = params lowerCAmelCase__ : Union[str, Any] = np.array(a ) lowerCAmelCase__ : List[Any] = np.array([len(a ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a : List[str] ): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[int] ): '''simple docstring''' return len(self.lengths ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.params.max_model_input_size lowerCAmelCase__ : Optional[int] = self.lengths > max_len logger.info(f'''Splitting {sum(a )} too long sequences.''' ) def divide_chunks(a : List[str] , a : Tuple ): return [l[i : i + n] for i in range(0 , len(a ) , a )] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Union[str, Any] = [] if self.params.mlm: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: lowerCAmelCase__ , lowerCAmelCase__ : int = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: lowerCAmelCase__ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: lowerCAmelCase__ : Dict = np.insert(a , 0 , a ) if sub_s[-1] != sep_id: lowerCAmelCase__ : Dict = np.insert(a , len(a ) , a ) assert len(a ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a ) new_tok_ids.extend(a ) new_lengths.extend([len(a ) for l in sub_seqs] ) lowerCAmelCase__ : str = np.array(a ) lowerCAmelCase__ : Optional[Any] = np.array(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = len(self ) lowerCAmelCase__ : List[Any] = self.lengths > 11 lowerCAmelCase__ : Dict = self.token_ids[indices] lowerCAmelCase__ : Tuple = self.lengths[indices] lowerCAmelCase__ : Any = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: lowerCAmelCase__ : int = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : str = len(self ) lowerCAmelCase__ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) lowerCAmelCase__ : int = (unk_occs / self.lengths) < 0.5 lowerCAmelCase__ : List[str] = self.token_ids[indices] lowerCAmelCase__ : Optional[Any] = self.lengths[indices] lowerCAmelCase__ : Union[str, Any] = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : int , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = [t[0] for t in batch] lowerCAmelCase__ : List[str] = [t[1] for t in batch] assert len(a ) == len(a ) # Max for paddings lowerCAmelCase__ : List[str] = max(a ) # Pad token ids if self.params.mlm: lowerCAmelCase__ : str = self.params.special_tok_ids['pad_token'] else: lowerCAmelCase__ : Optional[int] = self.params.special_tok_ids['unk_token'] lowerCAmelCase__ : Tuple = [list(t.astype(a ) ) + [pad_idx] * (max_seq_len_ - len(a )) for t in token_ids] assert len(tk_ ) == len(a ) assert all(len(a ) == max_seq_len_ for t in tk_ ) lowerCAmelCase__ : Union[str, Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) lowerCAmelCase__ : List[str] = torch.tensor(a ) # (bs) return tk_t, lg_t
69
0
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { """huggingface/autoformer-tourism-monthly""": """https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json""", } class A__ ( __magic_name__ ): lowercase = 'autoformer' lowercase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', 'num_hidden_layers': 'encoder_layers', } def __init__( self : Optional[int] , a : Optional[int] = None , a : Optional[int] = None , a : str = "student_t" , a : str = "nll" , a : int = 1 , a : List[int] = [1, 2, 3, 4, 5, 6, 7] , a : bool = True , a : int = 0 , a : int = 0 , a : int = 0 , a : int = 0 , a : Optional[List[int]] = None , a : Optional[List[int]] = None , a : int = 64 , a : int = 2 , a : int = 2 , a : int = 2 , a : int = 2 , a : int = 32 , a : int = 32 , a : str = "gelu" , a : float = 0.1 , a : float = 0.1 , a : float = 0.1 , a : float = 0.1 , a : float = 0.1 , a : int = 100 , a : float = 0.0_2 , a : bool = True , a : Optional[Any]=True , a : int = 10 , a : int = 25 , a : int = 3 , **a : Optional[Any] , ): '''simple docstring''' lowerCAmelCase__ : List[str] = prediction_length lowerCAmelCase__ : Tuple = context_length if context_length is not None else prediction_length lowerCAmelCase__ : Optional[int] = distribution_output lowerCAmelCase__ : Union[str, Any] = loss lowerCAmelCase__ : Optional[Any] = input_size lowerCAmelCase__ : Tuple = num_time_features lowerCAmelCase__ : Optional[Any] = lags_sequence lowerCAmelCase__ : Union[str, Any] = scaling lowerCAmelCase__ : List[str] = num_dynamic_real_features lowerCAmelCase__ : Dict = num_static_real_features lowerCAmelCase__ : Dict = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(a ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) lowerCAmelCase__ : List[Any] = cardinality else: lowerCAmelCase__ : Tuple = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(a ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) lowerCAmelCase__ : str = embedding_dimension else: lowerCAmelCase__ : List[str] = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] lowerCAmelCase__ : Any = num_parallel_samples # Transformer architecture configuration lowerCAmelCase__ : Optional[Any] = input_size * len(self.lags_sequence ) + self._number_of_features lowerCAmelCase__ : List[str] = d_model lowerCAmelCase__ : Union[str, Any] = encoder_attention_heads lowerCAmelCase__ : Optional[Any] = decoder_attention_heads lowerCAmelCase__ : Union[str, Any] = encoder_ffn_dim lowerCAmelCase__ : Optional[Any] = decoder_ffn_dim lowerCAmelCase__ : Optional[Any] = encoder_layers lowerCAmelCase__ : Optional[Any] = decoder_layers lowerCAmelCase__ : Optional[int] = dropout lowerCAmelCase__ : Any = attention_dropout lowerCAmelCase__ : str = activation_dropout lowerCAmelCase__ : Optional[int] = encoder_layerdrop lowerCAmelCase__ : List[Any] = decoder_layerdrop lowerCAmelCase__ : str = activation_function lowerCAmelCase__ : Tuple = init_std lowerCAmelCase__ : int = use_cache # Autoformer lowerCAmelCase__ : int = label_length lowerCAmelCase__ : Any = moving_average lowerCAmelCase__ : Tuple = autocorrelation_factor super().__init__(is_encoder_decoder=a , **a ) @property def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
708
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class A__ ( unittest.TestCase ): def __init__( self : List[str] , a : str , a : List[Any]=7 , a : Dict=3 , a : Optional[Any]=18 , a : List[Any]=30 , a : Union[str, Any]=400 , a : Tuple=True , a : Optional[int]=32 , a : Dict=True , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = parent lowerCAmelCase__ : str = batch_size lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Dict = image_size lowerCAmelCase__ : List[str] = min_resolution lowerCAmelCase__ : int = max_resolution lowerCAmelCase__ : Any = do_resize lowerCAmelCase__ : Tuple = size_divisor lowerCAmelCase__ : Optional[Any] = do_rescale def _lowerCamelCase ( self : Dict ): '''simple docstring''' return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class A__ ( __magic_name__ , unittest.TestCase ): lowercase = GLPNImageProcessor if is_vision_available() else None def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = GLPNImageProcessingTester(self ) @property def _lowerCamelCase ( self : int ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a , 'do_resize' ) ) self.assertTrue(hasattr(a , 'size_divisor' ) ) self.assertTrue(hasattr(a , 'resample' ) ) self.assertTrue(hasattr(a , 'do_rescale' ) ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' pass def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=a ) for image in image_inputs: self.assertIsInstance(a , Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) lowerCAmelCase__ : Tuple = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase__ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=a , numpify=a ) for image in image_inputs: self.assertIsInstance(a , np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) lowerCAmelCase__ : Optional[int] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=a , torchify=a ) for image in image_inputs: self.assertIsInstance(a , torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) lowerCAmelCase__ : Tuple = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
709
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowerCamelCase__ = None lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } lowerCamelCase__ = { """google/rembert""": 256, } lowerCamelCase__ = """▁""" class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = RemBertTokenizer def __init__( self : Optional[Any] , a : str=None , a : Any=None , a : List[Any]=True , a : str=True , a : Dict=False , a : Dict="[CLS]" , a : int="[SEP]" , a : Tuple="<unk>" , a : Optional[Any]="[SEP]" , a : Tuple="<pad>" , a : Dict="[CLS]" , a : Optional[Any]="[MASK]" , **a : str , ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( a , tokenizer_file=a , do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : int = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : List[Any] = keep_accents lowerCAmelCase__ : Optional[Any] = vocab_file lowerCAmelCase__ : Union[str, Any] = False if not self.vocab_file else True def _lowerCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : List[Any] , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Tuple = [self.sep_token_id] lowerCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Tuple , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Any = abs(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : List[Any] = 0 while n > 0: res += n % 10 n //= 10 return res def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : List[str] = abs(SCREAMING_SNAKE_CASE_ ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> int: return sum(int(SCREAMING_SNAKE_CASE_ ) for c in str(abs(SCREAMING_SNAKE_CASE_ ) ) ) def lowerCAmelCase__ ( ) -> None: from collections.abc import Callable from timeit import timeit def benchmark_a_function(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> None: lowerCAmelCase__ : List[str] = F'''{func.__name__}({value})''' lowerCAmelCase__ : int = timeit(F'''__main__.{call}''' , setup='import __main__' ) print(F'''{call:56} = {func(SCREAMING_SNAKE_CASE_ )} -- {timing:.4f} seconds''' ) for value in (262_144, 1_125_899_906_842_624, 1_267_650_600_228_229_401_496_703_205_376): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
710
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class A__ ( __magic_name__ ): def __init__( self : List[Any] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , ): '''simple docstring''' super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowerCAmelCase__ : int = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' if self.streaming: lowerCAmelCase__ : List[Any] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase__ : Any = None lowerCAmelCase__ : int = None lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Dict = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowerCAmelCase__ : Union[str, Any] = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
69
0
import tensorflow as tf from ...tf_utils import shape_list class A__ ( tf.keras.layers.Layer ): def __init__( self : Any , a : List[str] , a : Optional[int] , a : Tuple , a : int , a : Optional[int]=1 , a : Dict=False , **a : Optional[Any] ): '''simple docstring''' super().__init__(**a ) lowerCAmelCase__ : int = vocab_size lowerCAmelCase__ : List[Any] = d_embed lowerCAmelCase__ : str = d_proj lowerCAmelCase__ : Optional[Any] = cutoffs + [vocab_size] lowerCAmelCase__ : Dict = [0] + self.cutoffs lowerCAmelCase__ : Optional[int] = div_val lowerCAmelCase__ : int = self.cutoffs[0] lowerCAmelCase__ : List[Any] = len(self.cutoffs ) - 1 lowerCAmelCase__ : str = self.shortlist_size + self.n_clusters lowerCAmelCase__ : Optional[Any] = keep_order lowerCAmelCase__ : int = [] lowerCAmelCase__ : Union[str, Any] = [] def _lowerCamelCase ( self : Union[str, Any] , a : List[str] ): '''simple docstring''' if self.n_clusters > 0: lowerCAmelCase__ : Optional[Any] = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer='zeros' , trainable=a , name='cluster_weight' ) lowerCAmelCase__ : Tuple = self.add_weight( shape=(self.n_clusters,) , initializer='zeros' , trainable=a , name='cluster_bias' ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: lowerCAmelCase__ : Any = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer='zeros' , trainable=a , name=f'''out_projs_._{i}''' , ) self.out_projs.append(a ) else: self.out_projs.append(a ) lowerCAmelCase__ : Optional[int] = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer='zeros' , trainable=a , name=f'''out_layers_._{i}_._weight''' , ) lowerCAmelCase__ : List[str] = self.add_weight( shape=(self.vocab_size,) , initializer='zeros' , trainable=a , name=f'''out_layers_._{i}_._bias''' , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): lowerCAmelCase__ : str = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowerCAmelCase__ : List[Any] = self.d_embed // (self.div_val**i) lowerCAmelCase__ : List[Any] = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer='zeros' , trainable=a , name=f'''out_projs_._{i}''' ) self.out_projs.append(a ) lowerCAmelCase__ : List[Any] = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer='zeros' , trainable=a , name=f'''out_layers_._{i}_._weight''' , ) lowerCAmelCase__ : List[Any] = self.add_weight( shape=(r_idx - l_idx,) , initializer='zeros' , trainable=a , name=f'''out_layers_._{i}_._bias''' , ) self.out_layers.append((weight, bias) ) super().build(a ) @staticmethod def _lowerCamelCase ( a : str , a : Tuple , a : Union[str, Any] , a : List[str]=None ): '''simple docstring''' lowerCAmelCase__ : str = x if proj is not None: lowerCAmelCase__ : Any = tf.einsum('ibd,ed->ibe' , a , a ) return tf.einsum('ibd,nd->ibn' , a , a ) + b @staticmethod def _lowerCamelCase ( a : Tuple , a : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = shape_list(a ) lowerCAmelCase__ : List[str] = tf.range(lp_size[0] , dtype=target.dtype ) lowerCAmelCase__ : Optional[Any] = tf.stack([r, target] , 1 ) return tf.gather_nd(a , a ) def _lowerCamelCase ( self : Any , a : Union[str, Any] , a : int , a : Optional[Any]=True , a : Dict=False ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 0 if self.n_clusters == 0: lowerCAmelCase__ : str = self._logit(a , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: lowerCAmelCase__ : Tuple = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=a , logits=a ) lowerCAmelCase__ : List[str] = tf.nn.log_softmax(a , axis=-1 ) else: lowerCAmelCase__ : Tuple = shape_list(a ) lowerCAmelCase__ : str = [] lowerCAmelCase__ : Optional[Any] = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): lowerCAmelCase__ : Union[str, Any] = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: lowerCAmelCase__ : List[Any] = (target >= l_idx) & (target < r_idx) lowerCAmelCase__ : Dict = tf.where(a ) lowerCAmelCase__ : int = tf.boolean_mask(a , a ) - l_idx if self.div_val == 1: lowerCAmelCase__ : Any = self.out_layers[0][0][l_idx:r_idx] lowerCAmelCase__ : Optional[int] = self.out_layers[0][1][l_idx:r_idx] else: lowerCAmelCase__ : Union[str, Any] = self.out_layers[i][0] lowerCAmelCase__ : List[str] = self.out_layers[i][1] if i == 0: lowerCAmelCase__ : Optional[Any] = tf.concat([cur_W, self.cluster_weight] , 0 ) lowerCAmelCase__ : Optional[int] = tf.concat([cur_b, self.cluster_bias] , 0 ) lowerCAmelCase__ : Optional[int] = self._logit(a , a , a , self.out_projs[0] ) lowerCAmelCase__ : Tuple = tf.nn.log_softmax(a ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: lowerCAmelCase__ : List[str] = tf.boolean_mask(a , a ) lowerCAmelCase__ : int = self._gather_logprob(a , a ) else: lowerCAmelCase__ : List[str] = self._logit(a , a , a , self.out_projs[i] ) lowerCAmelCase__ : Any = tf.nn.log_softmax(a ) lowerCAmelCase__ : Optional[int] = self.cutoffs[0] + i - 1 # No probability for the head cluster lowerCAmelCase__ : Optional[int] = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(a ) if target is not None: lowerCAmelCase__ : Optional[Any] = tf.boolean_mask(a , a ) lowerCAmelCase__ : Union[str, Any] = tf.boolean_mask(a , a ) lowerCAmelCase__ : str = self._gather_logprob(a , a ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(a , -cur_logprob , shape_list(a ) ) lowerCAmelCase__ : Dict = tf.concat(a , axis=-1 ) if target is not None: if return_mean: lowerCAmelCase__ : str = tf.reduce_mean(a ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(a ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(a , name=self.name , aggregation='mean' if return_mean else '' ) return out
711
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase__ = logging.get_logger(__name__) class A__ ( __magic_name__ ): lowercase = ['audio_values', 'audio_mask'] def __init__( self : Dict , a : Dict=2_048 , a : Optional[Any]=1 , a : List[Any]=[16, 16] , a : Dict=128 , a : List[str]=44_100 , a : Union[str, Any]=86 , a : Optional[Any]=2_048 , a : List[Any]=0.0 , **a : Tuple , ): '''simple docstring''' super().__init__( feature_size=a , sampling_rate=a , padding_value=a , **a , ) lowerCAmelCase__ : Optional[Any] = spectrogram_length lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Tuple = patch_size lowerCAmelCase__ : Optional[int] = feature_size // self.patch_size[1] lowerCAmelCase__ : Union[str, Any] = n_fft lowerCAmelCase__ : Union[str, Any] = sampling_rate // hop_length_to_sampling_rate lowerCAmelCase__ : int = sampling_rate lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=a , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=a , norm='slaney' , mel_scale='slaney' , ).T def _lowerCamelCase ( self : Optional[int] , a : np.array ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = spectrogram( a , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) lowerCAmelCase__ : Any = log_spec[:, :-1] lowerCAmelCase__ : Dict = log_spec - 2_0.0 lowerCAmelCase__ : Tuple = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : str , a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a : Optional[Union[str, TensorType]] = None , a : Optional[bool] = True , a : Optional[int] = None , a : bool = False , a : bool = False , **a : int , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' f''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ : Dict = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase__ : List[Any] = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ : int = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): lowerCAmelCase__ : Optional[Any] = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis lowerCAmelCase__ : int = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , a ): lowerCAmelCase__ : Optional[Any] = [np.asarray(a , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask lowerCAmelCase__ : Optional[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: lowerCAmelCase__ : Any = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] lowerCAmelCase__ : List[Any] = np.array(a ).astype(np.floataa ) # convert into correct format for padding lowerCAmelCase__ : int = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch lowerCAmelCase__ : Dict = np.ones([len(a ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) lowerCAmelCase__ : Optional[Any] = padded_audio_features * self.padding_value for i in range(len(a ) ): lowerCAmelCase__ : Tuple = audio_features[i] lowerCAmelCase__ : List[str] = feature # return as BatchFeature if return_attention_mask: lowerCAmelCase__ : Tuple = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: lowerCAmelCase__ : Any = {'audio_values': padded_audio_features} lowerCAmelCase__ : Any = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
69
0
import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {"""vocab_file""": """sentencepiece.model"""} lowerCamelCase__ = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, } lowerCamelCase__ = { """google/rembert""": 256, } class A__ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : List[Any] , a : Union[str, Any] , a : Dict=False , a : Optional[Any]=True , a : Optional[Any]=True , a : Optional[int]="[CLS]" , a : Optional[Any]="[SEP]" , a : Dict="[UNK]" , a : Optional[int]="[SEP]" , a : Union[str, Any]="[PAD]" , a : List[Any]="[CLS]" , a : List[str]="[MASK]" , **a : Dict , ): '''simple docstring''' super().__init__( do_lower_case=a , remove_space=a , keep_accents=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , **a , ) lowerCAmelCase__ : Dict = do_lower_case lowerCAmelCase__ : int = remove_space lowerCAmelCase__ : Any = keep_accents lowerCAmelCase__ : Optional[int] = vocab_file lowerCAmelCase__ : Optional[int] = spm.SentencePieceProcessor() self.sp_model.Load(a ) @property def _lowerCamelCase ( self : Tuple ): '''simple docstring''' return len(self.sp_model ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = {self.convert_ids_to_tokens(a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : int ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.__dict__.copy() lowerCAmelCase__ : Any = None return state def __setstate__( self : List[str] , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = d lowerCAmelCase__ : int = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def _lowerCamelCase ( self : str , a : List[str] , a : int=False ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.sp_model.EncodeAsPieces(a ) return pieces def _lowerCamelCase ( self : int , a : List[str] ): '''simple docstring''' return self.sp_model.PieceToId(a ) def _lowerCamelCase ( self : Union[str, Any] , a : Any ): '''simple docstring''' return self.sp_model.IdToPiece(a ) def _lowerCamelCase ( self : Optional[int] , a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.sp_model.decode_pieces(a ) return out_string def _lowerCamelCase ( self : Tuple , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = [self.sep_token_id] lowerCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self : Any , a : List[int] , a : Optional[List[int]] = None , a : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a )) + [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1] def _lowerCamelCase ( self : str , a : List[int] , a : Optional[List[int]] = None ): '''simple docstring''' lowerCAmelCase__ : Dict = [self.sep_token_id] lowerCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self : Any , a : str , a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(a ): logger.error('Vocabulary path ({}) should be a directory'.format(a ) ) return lowerCAmelCase__ : int = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ): copyfile(self.vocab_file , a ) return (out_vocab_file,)
712
import unittest from transformers import DonutProcessor lowerCamelCase__ = """naver-clova-ix/donut-base""" class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : List[str] = DonutProcessor.from_pretrained(a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Dict = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } lowerCAmelCase__ : Union[str, Any] = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) lowerCAmelCase__ : Optional[Any] = self.processor.tokenajson(a ) self.assertDictEqual(a , a )
69
0
from __future__ import annotations import math def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list[int]: if num <= 0: lowerCAmelCase__ : Dict = F'''{num}: Invalid input, please enter a positive integer.''' raise ValueError(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[int] = [True] * (num + 1) lowerCAmelCase__ : str = [] lowerCAmelCase__ : Optional[Any] = 2 lowerCAmelCase__ : List[str] = int(math.sqrt(SCREAMING_SNAKE_CASE_ ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE_ ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE_ ): if sieve[i] is True: lowerCAmelCase__ : Dict = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE_ ) return prime if __name__ == "__main__": print(prime_sieve(int(input("""Enter a positive integer: """).strip())))
713
from numpy import exp, pi, sqrt def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = 1.0 ) -> int: return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
69
0
import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __magic_name__ , unittest.TestCase ): lowercase = GPTaTokenizer lowercase = GPTaTokenizerFast lowercase = True lowercase = {'add_prefix_space': True} lowercase = False def _lowerCamelCase ( self : Tuple ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : Optional[Any] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', '<|endoftext|>', ] lowerCAmelCase__ : str = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Tuple = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] lowerCAmelCase__ : Optional[int] = {'unk_token': '<unk>'} lowerCAmelCase__ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(a ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : Any , **a : int ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **a ) def _lowerCamelCase ( self : int , **a : str ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **a ) def _lowerCamelCase ( self : str , a : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = 'lower newer' lowerCAmelCase__ : Union[str, Any] = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : List[Any] = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) lowerCAmelCase__ : Union[str, Any] = 'lower newer' lowerCAmelCase__ : str = ['\u0120low', 'er', '\u0120', 'n', 'e', 'w', 'er'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a , add_prefix_space=a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Union[str, Any] = tokens + [tokenizer.unk_token] lowerCAmelCase__ : List[str] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' if not self.test_rust_tokenizer: return lowerCAmelCase__ : Any = self.get_tokenizer() lowerCAmelCase__ : List[Any] = self.get_rust_tokenizer(add_prefix_space=a ) lowerCAmelCase__ : List[str] = 'lower newer' # Testing tokenization lowerCAmelCase__ : Union[str, Any] = tokenizer.tokenize(a , add_prefix_space=a ) lowerCAmelCase__ : Optional[Any] = rust_tokenizer.tokenize(a ) self.assertListEqual(a , a ) # Testing conversion to ids without special tokens lowerCAmelCase__ : Tuple = tokenizer.encode(a , add_special_tokens=a , add_prefix_space=a ) lowerCAmelCase__ : Union[str, Any] = rust_tokenizer.encode(a , add_special_tokens=a ) self.assertListEqual(a , a ) # Testing conversion to ids with special tokens lowerCAmelCase__ : List[Any] = self.get_rust_tokenizer(add_prefix_space=a ) lowerCAmelCase__ : Any = tokenizer.encode(a , add_prefix_space=a ) lowerCAmelCase__ : Union[str, Any] = rust_tokenizer.encode(a ) self.assertListEqual(a , a ) # Testing the unknown token lowerCAmelCase__ : Optional[Any] = tokens + [rust_tokenizer.unk_token] lowerCAmelCase__ : Any = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(a ) , a ) def _lowerCamelCase ( self : Any , *a : Optional[int] , **a : Tuple ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple , a : Optional[int]=15 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase__ : int = self.rust_tokenizer_class.from_pretrained(a , **a ) # Simple input lowerCAmelCase__ : str = 'This is a simple input' lowerCAmelCase__ : List[str] = ['This is a simple input 1', 'This is a simple input 2'] lowerCAmelCase__ : int = ('This is a simple input', 'This is a pair') lowerCAmelCase__ : Dict = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(a , tokenizer_r.encode , a , max_length=a , padding='max_length' ) # Simple input self.assertRaises(a , tokenizer_r.encode_plus , a , max_length=a , padding='max_length' ) # Simple input self.assertRaises( a , tokenizer_r.batch_encode_plus , a , max_length=a , padding='max_length' , ) # Pair input self.assertRaises(a , tokenizer_r.encode , a , max_length=a , padding='max_length' ) # Pair input self.assertRaises(a , tokenizer_r.encode_plus , a , max_length=a , padding='max_length' ) # Pair input self.assertRaises( a , tokenizer_r.batch_encode_plus , a , max_length=a , padding='max_length' , ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token='<pad>' ) # Simple input lowerCAmelCase__ : List[Any] = 'This is a simple input' lowerCAmelCase__ : int = ['This is a simple input looooooooong', 'This is a simple input'] lowerCAmelCase__ : str = ('This is a simple input', 'This is a pair') lowerCAmelCase__ : Any = [ ('This is a simple input loooooong', 'This is a simple input'), ('This is a simple pair loooooong', 'This is a simple pair'), ] lowerCAmelCase__ : List[str] = tokenizer.pad_token_id lowerCAmelCase__ : List[Any] = tokenizer(a , padding='max_length' , max_length=30 , return_tensors='np' ) lowerCAmelCase__ : List[str] = tokenizer(a , padding=a , truncate=a , return_tensors='np' ) lowerCAmelCase__ : Optional[int] = tokenizer(*a , padding='max_length' , max_length=60 , return_tensors='np' ) lowerCAmelCase__ : int = tokenizer(a , padding=a , truncate=a , return_tensors='np' ) # s # test single string max_length padding self.assertEqual(out_s['input_ids'].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['input_ids'] ) self.assertTrue(0 in out_s['attention_mask'] ) # s2 # test automatic padding self.assertEqual(out_sa['input_ids'].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['input_ids'][0] ) self.assertFalse(0 in out_sa['attention_mask'][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['input_ids'][1] ) self.assertTrue(0 in out_sa['attention_mask'][1] ) # p # test single pair max_length padding self.assertEqual(out_p['input_ids'].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['input_ids'] ) self.assertTrue(0 in out_p['attention_mask'] ) # p2 # test automatic padding pair self.assertEqual(out_pa['input_ids'].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['input_ids'][0] ) self.assertFalse(0 in out_pa['attention_mask'][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['input_ids'][1] ) self.assertTrue(0 in out_pa['attention_mask'][1] ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : List[str] = '$$$' lowerCAmelCase__ : int = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=a , add_bos_token=a ) lowerCAmelCase__ : Optional[int] = 'This is a simple input' lowerCAmelCase__ : List[Any] = ['This is a simple input 1', 'This is a simple input 2'] lowerCAmelCase__ : List[str] = tokenizer.bos_token_id lowerCAmelCase__ : Any = tokenizer(a ) lowerCAmelCase__ : Any = tokenizer(a ) self.assertEqual(out_s.input_ids[0] , a ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) lowerCAmelCase__ : int = tokenizer.decode(out_s.input_ids ) lowerCAmelCase__ : int = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , a ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' pass def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : List[Any] = [self.get_tokenizer(do_lower_case=a , add_bos_token=a )] for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): lowerCAmelCase__ : Dict = 'Encode this.' lowerCAmelCase__ : Optional[Any] = 'This one too please.' lowerCAmelCase__ : Union[str, Any] = tokenizer.encode(a , add_special_tokens=a ) encoded_sequence += tokenizer.encode(a , add_special_tokens=a ) lowerCAmelCase__ : Tuple = tokenizer.encode_plus( a , a , add_special_tokens=a , return_special_tokens_mask=a , ) lowerCAmelCase__ : List[Any] = encoded_sequence_dict['input_ids'] lowerCAmelCase__ : Union[str, Any] = encoded_sequence_dict['special_tokens_mask'] self.assertEqual(len(a ) , len(a ) ) lowerCAmelCase__ : Optional[Any] = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(a ) ] lowerCAmelCase__ : Union[str, Any] = [x for x in filtered_sequence if x is not None] self.assertEqual(a , a ) @require_tokenizers class A__ ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = AutoTokenizer.from_pretrained('facebook/opt-350m' , from_slow=a ) lowerCAmelCase__ : str = 'A photo of a cat' lowerCAmelCase__ : Dict = tokenizer.encode( a , ) self.assertEqual(a , [2, 250, 1_345, 9, 10, 4_758] ) tokenizer.save_pretrained('test_opt' ) lowerCAmelCase__ : Optional[Any] = AutoTokenizer.from_pretrained('./test_opt' ) lowerCAmelCase__ : str = tokenizer.encode( a , ) self.assertEqual(a , [2, 250, 1_345, 9, 10, 4_758] ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = AutoTokenizer.from_pretrained('facebook/opt-350m' , use_slow=a ) lowerCAmelCase__ : Optional[int] = 'A photo of a cat' lowerCAmelCase__ : Dict = tokenizer.encode( a , ) # Same as above self.assertEqual(a , [2, 250, 1_345, 9, 10, 4_758] ) @unittest.skip('This test is failing because of a bug in the fast tokenizer' ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = AutoTokenizer.from_pretrained('facebook/opt-350m' , from_slow=a ) lowerCAmelCase__ : Tuple = 'bos' lowerCAmelCase__ : Optional[int] = tokenizer.get_vocab()['bos'] lowerCAmelCase__ : List[Any] = 'A photo of a cat' lowerCAmelCase__ : Optional[Any] = tokenizer.encode( a , ) # We changed the bos token self.assertEqual(a , [31_957, 250, 1_345, 9, 10, 4_758] ) tokenizer.save_pretrained('./tok' ) lowerCAmelCase__ : Optional[int] = AutoTokenizer.from_pretrained('./tok' ) self.assertTrue(tokenizer.is_fast ) lowerCAmelCase__ : Optional[Any] = tokenizer.encode( a , ) self.assertEqual(a , [31_957, 250, 1_345, 9, 10, 4_758] )
714
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A__ ( __magic_name__ , unittest.TestCase ): lowercase = XLMTokenizer lowercase = False def _lowerCamelCase ( self : int ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : List[str] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowerCAmelCase__ : Any = dict(zip(a , range(len(a ) ) ) ) lowerCAmelCase__ : Optional[int] = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] lowerCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(a ) ) def _lowerCamelCase ( self : List[str] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : List[Any] = 'lower newer' lowerCAmelCase__ : Any = 'lower newer' return input_text, output_text def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Tuple = XLMTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase__ : Optional[int] = 'lower' lowerCAmelCase__ : Optional[Any] = ['low', 'er</w>'] lowerCAmelCase__ : Dict = tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowerCAmelCase__ : Tuple = tokens + ['<unk>'] lowerCAmelCase__ : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) @slow def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowerCAmelCase__ : Any = tokenizer.encode('sequence builders' , add_special_tokens=a ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=a ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(a ) lowerCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
0
from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class A__ : lowercase = XGLMConfig lowercase = {} lowercase = 'gelu' def __init__( self : Tuple , a : int , a : str=14 , a : Tuple=7 , a : Optional[int]=True , a : Optional[int]=True , a : List[str]=True , a : int=99 , a : Optional[int]=32 , a : Any=2 , a : Optional[Any]=4 , a : Optional[int]=37 , a : List[Any]="gelu" , a : Union[str, Any]=0.1 , a : Any=0.1 , a : Optional[int]=512 , a : Tuple=0.0_2 , ): '''simple docstring''' lowerCAmelCase__ : str = parent lowerCAmelCase__ : Tuple = batch_size lowerCAmelCase__ : Optional[int] = seq_length lowerCAmelCase__ : List[Any] = is_training lowerCAmelCase__ : str = use_input_mask lowerCAmelCase__ : Optional[Any] = use_labels lowerCAmelCase__ : List[str] = vocab_size lowerCAmelCase__ : Optional[int] = d_model lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : Any = num_attention_heads lowerCAmelCase__ : Dict = ffn_dim lowerCAmelCase__ : Any = activation_function lowerCAmelCase__ : Optional[Any] = activation_dropout lowerCAmelCase__ : Dict = attention_dropout lowerCAmelCase__ : int = max_position_embeddings lowerCAmelCase__ : Union[str, Any] = initializer_range lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : Optional[int] = 0 lowerCAmelCase__ : str = 2 lowerCAmelCase__ : List[str] = 1 def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return XGLMConfig.from_pretrained('facebook/xglm-564M' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[str] = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) lowerCAmelCase__ : Union[str, Any] = None if self.use_input_mask: lowerCAmelCase__ : Any = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : Optional[Any] = self.get_config() lowerCAmelCase__ : Tuple = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=a , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=a , ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.prepare_config_and_inputs() ( lowerCAmelCase__ ) : Tuple = config_and_inputs lowerCAmelCase__ : Tuple = { 'input_ids': input_ids, 'head_mask': head_mask, } return config, inputs_dict @require_tf class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () lowercase = (TFXGLMForCausalLM,) if is_tf_available() else () lowercase = ( {'feature-extraction': TFXGLMModel, 'text-generation': TFXGLMForCausalLM} if is_tf_available() else {} ) lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = TFXGLMModelTester(self ) lowerCAmelCase__ : Any = ConfigTester(self , config_class=a , n_embd=37 ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' self.config_tester.run_common_tests() @slow def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Tuple = TFXGLMModel.from_pretrained(a ) self.assertIsNotNone(a ) @unittest.skip(reason='Currently, model embeddings are going to undergo a major refactor.' ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' super().test_resize_token_embeddings() @require_tf class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Tuple , a : Optional[Any]=True ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) lowerCAmelCase__ : Dict = tf.convert_to_tensor([[2, 268, 9_865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off lowerCAmelCase__ : int = [2, 268, 9_865, 67, 11, 1_988, 57_252, 9_865, 5, 984, 67, 1_988, 213_838, 1_658, 53, 70_446, 33, 6_657, 278, 1_581] # fmt: on lowerCAmelCase__ : Dict = model.generate(a , do_sample=a , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) lowerCAmelCase__ : str = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) tf.random.set_seed(0 ) lowerCAmelCase__ : int = tokenizer('Today is a nice day and' , return_tensors='tf' ) lowerCAmelCase__ : List[Any] = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(':/CPU:0' ): lowerCAmelCase__ : Tuple = model.generate(a , do_sample=a , seed=[7, 0] ) lowerCAmelCase__ : Union[str, Any] = tokenizer.decode(output_ids[0] , skip_special_tokens=a ) lowerCAmelCase__ : int = ( 'Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due' ) self.assertEqual(a , a ) @slow def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) lowerCAmelCase__ : Union[str, Any] = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) lowerCAmelCase__ : Any = 'left' # use different length sentences to test batching lowerCAmelCase__ : List[Any] = [ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When', 'Hello, my dog is a little', ] lowerCAmelCase__ : Dict = tokenizer(a , return_tensors='tf' , padding=a ) lowerCAmelCase__ : int = inputs['input_ids'] lowerCAmelCase__ : Dict = model.generate(input_ids=a , attention_mask=inputs['attention_mask'] , max_new_tokens=12 ) lowerCAmelCase__ : Union[str, Any] = tokenizer(sentences[0] , return_tensors='tf' ).input_ids lowerCAmelCase__ : int = model.generate(input_ids=a , max_new_tokens=12 ) lowerCAmelCase__ : Tuple = tokenizer(sentences[1] , return_tensors='tf' ).input_ids lowerCAmelCase__ : Any = model.generate(input_ids=a , max_new_tokens=12 ) lowerCAmelCase__ : List[str] = tokenizer.batch_decode(a , skip_special_tokens=a ) lowerCAmelCase__ : Optional[int] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=a ) lowerCAmelCase__ : Optional[int] = tokenizer.decode(output_padded[0] , skip_special_tokens=a ) lowerCAmelCase__ : Any = [ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ' 'a single', 'Hello, my dog is a little bit of a shy one, but he is very friendly', ] self.assertListEqual(a , a ) self.assertListEqual(a , [non_padded_sentence, padded_sentence] )
715
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = image.size lowerCAmelCase__ , lowerCAmelCase__ : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase__ : Any = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) lowerCAmelCase__ : int = np.array(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) / 255.0 lowerCAmelCase__ : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) lowerCAmelCase__ : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ) return 2.0 * image - 1.0 class A__ ( __magic_name__ ): def __init__( self : List[str] , a : VQModel , a : UNetaDModel , a : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): '''simple docstring''' super().__init__() self.register_modules(vqvae=a , unet=a , scheduler=a ) @torch.no_grad() def __call__( self : int , a : Union[torch.Tensor, PIL.Image.Image] = None , a : Optional[int] = 1 , a : Optional[int] = 100 , a : Optional[float] = 0.0 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , ): '''simple docstring''' if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : str = 1 elif isinstance(a , torch.Tensor ): lowerCAmelCase__ : Union[str, Any] = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a )}''' ) if isinstance(a , PIL.Image.Image ): lowerCAmelCase__ : List[Any] = preprocess(a ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowerCAmelCase__ : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) lowerCAmelCase__ : Optional[Any] = next(self.unet.parameters() ).dtype lowerCAmelCase__ : List[str] = randn_tensor(a , generator=a , device=self.device , dtype=a ) lowerCAmelCase__ : Any = image.to(device=self.device , dtype=a ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a , device=self.device ) lowerCAmelCase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowerCAmelCase__ : Optional[Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowerCAmelCase__ : Union[str, Any] = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowerCAmelCase__ : List[str] = {} if accepts_eta: lowerCAmelCase__ : List[Any] = eta for t in self.progress_bar(a ): # concat latents and low resolution image in the channel dimension. lowerCAmelCase__ : Union[str, Any] = torch.cat([latents, image] , dim=1 ) lowerCAmelCase__ : Dict = self.scheduler.scale_model_input(a , a ) # predict the noise residual lowerCAmelCase__ : Tuple = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowerCAmelCase__ : List[str] = self.scheduler.step(a , a , a , **a ).prev_sample # decode the image latents with the VQVAE lowerCAmelCase__ : Dict = self.vqvae.decode(a ).sample lowerCAmelCase__ : Tuple = torch.clamp(a , -1.0 , 1.0 ) lowerCAmelCase__ : Tuple = image / 2 + 0.5 lowerCAmelCase__ : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
69
0
'''simple docstring''' import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
716
import os from collections import deque import torch from torch.utils.data import Dataset class A__ ( __magic_name__ ): def __init__( self : Union[str, Any] , a : str="" , a : str="train" ): '''simple docstring''' assert os.path.isdir(a ) lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Dict = os.listdir(a ) for story_filename in story_filenames_list: if "summary" in story_filename: continue lowerCAmelCase__ : Union[str, Any] = os.path.join(a , a ) if not os.path.isfile(a ): continue self.documents.append(a ) def __len__( self : Any ): '''simple docstring''' return len(self.documents ) def __getitem__( self : Dict , a : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.documents[idx] lowerCAmelCase__ : Union[str, Any] = document_path.split('/' )[-1] with open(a , encoding='utf-8' ) as source: lowerCAmelCase__ : List[Any] = source.read() lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = process_story(a ) return document_name, story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Optional[int] = list(filter(lambda SCREAMING_SNAKE_CASE_ : len(SCREAMING_SNAKE_CASE_ ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it lowerCAmelCase__ : List[Any] = [_add_missing_period(SCREAMING_SNAKE_CASE_ ) for line in nonempty_lines] # gather article lines lowerCAmelCase__ : int = [] lowerCAmelCase__ : Any = deque(SCREAMING_SNAKE_CASE_ ) while True: try: lowerCAmelCase__ : int = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(SCREAMING_SNAKE_CASE_ ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines lowerCAmelCase__ : Tuple = list(filter(lambda SCREAMING_SNAKE_CASE_ : not t.startswith('@highlight' ) , SCREAMING_SNAKE_CASE_ ) ) return story_lines, summary_lines def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Any: lowerCAmelCase__ : int = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: if len(SCREAMING_SNAKE_CASE_ ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(SCREAMING_SNAKE_CASE_ )) ) return sequence def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : str = torch.ones_like(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = sequence == pad_token_id lowerCAmelCase__ : Optional[int] = 0 return mask def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: lowerCAmelCase__ : Any = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in story_lines] lowerCAmelCase__ : str = [token for sentence in story_lines_token_ids for token in sentence] lowerCAmelCase__ : Dict = [tokenizer.encode(SCREAMING_SNAKE_CASE_ ) for line in summary_lines] lowerCAmelCase__ : str = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: lowerCAmelCase__ : Optional[int] = [] for sequence in batch: lowerCAmelCase__ : Union[str, Any] = -1 lowerCAmelCase__ : int = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(SCREAMING_SNAKE_CASE_ ) return torch.tensor(SCREAMING_SNAKE_CASE_ )
69
0
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None ) -> Dict: # set parameter of one layer assert torch_layer.weight.shape == weight.shape, F'''{torch_layer} layer.weight does not match''' lowerCAmelCase__ : str = nn.Parameter(SCREAMING_SNAKE_CASE_ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'''{torch_layer} layer.bias does not match''' lowerCAmelCase__ : List[str] = nn.Parameter(SCREAMING_SNAKE_CASE_ ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> List[str]: # set torch weights for 1-to-1 comparison lowerCAmelCase__ : str = np.asarray(weights[0] ) lowerCAmelCase__ : Dict = np.asarray(weights[1] ) lowerCAmelCase__ : Any = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(SCREAMING_SNAKE_CASE_ ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE_ ) , ) set_param( torch_layer.self_attention.value , torch.tensor(SCREAMING_SNAKE_CASE_ ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE_ ) , ) set_param( torch_layer.output.dense , torch.tensor(SCREAMING_SNAKE_CASE_ ).view(-1 , SCREAMING_SNAKE_CASE_ ).contiguous().transpose(0 , 1 ) , ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: # set torch weights for 1-to-1 comparison lowerCAmelCase__ : Optional[int] = np.asarray(weights[0] ) lowerCAmelCase__ : Any = np.asarray(weights[1] ) lowerCAmelCase__ : Tuple = np.asarray(weights[2] ) lowerCAmelCase__ : Any = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(SCREAMING_SNAKE_CASE_ ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE_ ) , ) set_param( torch_layer.self_attention.key , torch.tensor(SCREAMING_SNAKE_CASE_ ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE_ ) , ) set_param( torch_layer.self_attention.value , torch.tensor(SCREAMING_SNAKE_CASE_ ).transpose(1 , 2 ).contiguous().view(-1 , SCREAMING_SNAKE_CASE_ ) , ) set_param( torch_layer.output.dense , torch.tensor(SCREAMING_SNAKE_CASE_ ).view(-1 , SCREAMING_SNAKE_CASE_ ).contiguous().transpose(0 , 1 ) , ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: # layernorm 1 lowerCAmelCase__ : List[Any] = weights[0][0][0] lowerCAmelCase__ : int = np.asarray(layer_norm_a[0] ) lowerCAmelCase__ : int = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(SCREAMING_SNAKE_CASE_ ) , torch.tensor(SCREAMING_SNAKE_CASE_ ) , ) # lsh weights + output lowerCAmelCase__ : Optional[Any] = weights[0][1] if len(SCREAMING_SNAKE_CASE_ ) < 4: set_layer_weights_in_torch_lsh(SCREAMING_SNAKE_CASE_ , torch_block.attention , SCREAMING_SNAKE_CASE_ ) else: set_layer_weights_in_torch_local(SCREAMING_SNAKE_CASE_ , torch_block.attention , SCREAMING_SNAKE_CASE_ ) # intermediate weighs lowerCAmelCase__ : List[Any] = weights[2][0][1][2] # Chunked Feed Forward if len(SCREAMING_SNAKE_CASE_ ) == 4: lowerCAmelCase__ : Dict = intermediate_weights[2] # layernorm 2 lowerCAmelCase__ : List[str] = np.asarray(intermediate_weights[0][0] ) lowerCAmelCase__ : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(SCREAMING_SNAKE_CASE_ ) , torch.tensor(SCREAMING_SNAKE_CASE_ ) , ) # intermediate dense lowerCAmelCase__ : Union[str, Any] = np.asarray(intermediate_weights[1][0] ) lowerCAmelCase__ : List[Any] = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(SCREAMING_SNAKE_CASE_ ).transpose(0 , 1 ).contiguous() , torch.tensor(SCREAMING_SNAKE_CASE_ ) , ) # intermediate out lowerCAmelCase__ : Any = np.asarray(intermediate_weights[4][0] ) lowerCAmelCase__ : Tuple = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(SCREAMING_SNAKE_CASE_ ).transpose(0 , 1 ).contiguous() , torch.tensor(SCREAMING_SNAKE_CASE_ ) , ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Any: # reformer model lowerCAmelCase__ : Union[str, Any] = torch_model.reformer # word embeds lowerCAmelCase__ : Tuple = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(SCREAMING_SNAKE_CASE_ ) , ) if isinstance(weights[3] , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase__ : List[Any] = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): lowerCAmelCase__ : int = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'''{position_embeddings[emb_idx]} emb does not match''' lowerCAmelCase__ : int = nn.Parameter(torch.tensor(SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase__ : List[Any] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( SCREAMING_SNAKE_CASE_ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): lowerCAmelCase__ : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # output layer norm lowerCAmelCase__ : Any = np.asarray(weights[7][0] ) lowerCAmelCase__ : Union[str, Any] = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(SCREAMING_SNAKE_CASE_ ) , torch.tensor(SCREAMING_SNAKE_CASE_ ) , ) # output embeddings lowerCAmelCase__ : List[str] = np.asarray(weights[9][0] ) lowerCAmelCase__ : int = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(SCREAMING_SNAKE_CASE_ ).transpose(0 , 1 ).contiguous() , torch.tensor(SCREAMING_SNAKE_CASE_ ) , ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: # Initialise PyTorch model lowerCAmelCase__ : Dict = ReformerConfig.from_json_file(SCREAMING_SNAKE_CASE_ ) print(F'''Building PyTorch model from configuration: {config}''' ) lowerCAmelCase__ : List[Any] = ReformerModelWithLMHead(SCREAMING_SNAKE_CASE_ ) with open(SCREAMING_SNAKE_CASE_ , 'rb' ) as f: lowerCAmelCase__ : Dict = pickle.load(SCREAMING_SNAKE_CASE_ )['weights'] set_model_weights_in_torch(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , config.hidden_size ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowerCamelCase__ = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
717
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCamelCase__ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase__ = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCamelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCamelCase__ = re.compile(r"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") lowerCamelCase__ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> List[str]: lowerCAmelCase__ : int = None # source code of `config_class` lowerCAmelCase__ : Optional[int] = inspect.getsource(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Dict = _re_checkpoint.findall(SCREAMING_SNAKE_CASE_ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('/' ): lowerCAmelCase__ : Union[str, Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowerCAmelCase__ : Dict = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: lowerCAmelCase__ : str = ckpt_name break return checkpoint def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : Union[str, Any] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowerCAmelCase__ : Union[str, Any] = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) > 0: lowerCAmelCase__ : List[str] = '\n'.join(sorted(SCREAMING_SNAKE_CASE_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
69
0
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class A__ : def __init__( self : int , a : int , a : MutableSequence[float] ): '''simple docstring''' if len(a ) != degree + 1: raise ValueError( 'The number of coefficients should be equal to the degree + 1.' ) lowerCAmelCase__ : list[float] = list(a ) lowerCAmelCase__ : Tuple = degree def __add__( self : Optional[int] , a : Polynomial ): '''simple docstring''' if self.degree > polynomial_a.degree: lowerCAmelCase__ : List[str] = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , a ) else: lowerCAmelCase__ : int = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , a ) def __sub__( self : Optional[Any] , a : Polynomial ): '''simple docstring''' return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : List[Any] ): '''simple docstring''' return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Dict , a : Polynomial ): '''simple docstring''' lowerCAmelCase__ : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , a ) def _lowerCamelCase ( self : Optional[int] , a : int | float ): '''simple docstring''' lowerCAmelCase__ : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = '' for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(a ) return polynomial def __repr__( self : str ): '''simple docstring''' return self.__str__() def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : list[float] = [0] * self.degree for i in range(self.degree ): lowerCAmelCase__ : Dict = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , a ) def _lowerCamelCase ( self : Tuple , a : int | float = 0 ): '''simple docstring''' lowerCAmelCase__ : list[float] = [0] * (self.degree + 2) lowerCAmelCase__ : Optional[Any] = constant for i in range(self.degree + 1 ): lowerCAmelCase__ : Tuple = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , a ) def __eq__( self : Optional[Any] , a : object ): '''simple docstring''' if not isinstance(a , a ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Optional[int] , a : object ): '''simple docstring''' return not self.__eq__(a )
718
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class A__ ( __magic_name__ ): def __init__( self : List[str] , a : Optional[Any] , a : int=13 , a : str=7 , a : Any=True , a : List[str]=True , a : Any=False , a : List[Any]=True , a : List[str]=99 , a : Optional[Any]=32 , a : List[str]=5 , a : List[Any]=4 , a : List[Any]=64 , a : List[Any]="gelu" , a : List[Any]=0.1 , a : List[Any]=0.1 , a : int=512 , a : Tuple=16 , a : List[str]=2 , a : int=0.0_2 , a : Union[str, Any]=3 , a : Any=4 , a : Union[str, Any]=None , a : Union[str, Any]=2 , a : List[str]=2 , a : int=2 , a : Dict=2 , a : List[str]=4 , a : str=1 , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = seq_length lowerCAmelCase__ : Tuple = is_training lowerCAmelCase__ : List[str] = use_input_mask lowerCAmelCase__ : Optional[int] = use_token_type_ids lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : str = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : int = intermediate_size lowerCAmelCase__ : Optional[int] = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Union[str, Any] = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Dict = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range lowerCAmelCase__ : List[Any] = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : str = scope lowerCAmelCase__ : Any = q_groups lowerCAmelCase__ : Any = k_groups lowerCAmelCase__ : Union[str, Any] = v_groups lowerCAmelCase__ : int = post_attention_groups lowerCAmelCase__ : str = intermediate_groups lowerCAmelCase__ : Union[str, Any] = output_groups def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : List[Any] = None lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : int = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : str ): '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Any , a : Optional[int] , a : str , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = SqueezeBertModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : str , a : Any , a : Tuple , a : int , a : Union[str, Any] , a : Tuple , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = SqueezeBertForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[int] , a : Union[str, Any] , a : Optional[Any] , a : str , a : Optional[Any] , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = SqueezeBertForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , start_positions=a , end_positions=a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : List[Any] , a : Optional[int] , a : Union[str, Any] , a : str , a : str , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Union[str, Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Any , a : int , a : Any , a : Dict , a : Any , a : Tuple , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : str = self.num_labels lowerCAmelCase__ : Dict = SqueezeBertForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : str , a : Optional[int] , a : List[Any] , a : int , a : List[Any] , a : Union[str, Any] , a : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Union[str, Any] = SqueezeBertForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : List[str] = model( a , attention_mask=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.prepare_config_and_inputs() (lowerCAmelCase__) : List[Any] = config_and_inputs lowerCAmelCase__ : Dict = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowercase = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = True lowercase = False def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = SqueezeBertModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=a , dim=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a ) def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a ) @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[int] = SqueezeBertModel.from_pretrained(a ) self.assertIsNotNone(a ) @require_sentencepiece @require_tokenizers @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : int = SqueezeBertForSequenceClassification.from_pretrained('squeezebert/squeezebert-mnli' ) lowerCAmelCase__ : str = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) lowerCAmelCase__ : Any = model(a )[0] lowerCAmelCase__ : Tuple = torch.Size((1, 3) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : int = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(a , a , atol=1E-4 ) )
719
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
69
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list[list[int]]: lowerCAmelCase__ : Union[str, Any] = [] if len(SCREAMING_SNAKE_CASE_ ) == 1: return [nums.copy()] for _ in range(len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ : Optional[int] = nums.pop(0 ) lowerCAmelCase__ : List[str] = permute(SCREAMING_SNAKE_CASE_ ) for perm in permutations: perm.append(SCREAMING_SNAKE_CASE_ ) result.extend(SCREAMING_SNAKE_CASE_ ) nums.append(SCREAMING_SNAKE_CASE_ ) return result def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: def backtrack(SCREAMING_SNAKE_CASE_ ): if start == len(SCREAMING_SNAKE_CASE_ ) - 1: output.append(nums[:] ) else: for i in range(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ): lowerCAmelCase__ : str = nums[i], nums[start] backtrack(start + 1 ) lowerCAmelCase__ : Tuple = nums[i], nums[start] # backtrack lowerCAmelCase__ : Optional[int] = [] backtrack(0 ) return output if __name__ == "__main__": import doctest # use res to print the data in permute2 function lowerCamelCase__ = permutea([1, 2, 3]) print(res) doctest.testmod()
720
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = """cuda""" if torch.cuda.is_available() else """cpu""" def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=100 , SCREAMING_SNAKE_CASE_=" " ) -> List[str]: lowerCAmelCase__ : Optional[Any] = text.split(SCREAMING_SNAKE_CASE_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ )] def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ , lowerCAmelCase__ : int = [], [] for title, text in zip(documents['title'] , documents['text'] ): if text is not None: for passage in split_text(SCREAMING_SNAKE_CASE_ ): titles.append(title if title is not None else '' ) texts.append(SCREAMING_SNAKE_CASE_ ) return {"title": titles, "text": texts} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> dict: lowerCAmelCase__ : List[str] = ctx_tokenizer( documents['title'] , documents['text'] , truncation=SCREAMING_SNAKE_CASE_ , padding='longest' , return_tensors='pt' )['input_ids'] lowerCAmelCase__ : Tuple = ctx_encoder(input_ids.to(device=SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> Optional[Any]: ###################################### logger.info('Step 1 - Create the dataset' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way lowerCAmelCase__ : str = load_dataset( 'csv' , data_files=[rag_example_args.csv_path] , split='train' , delimiter='\t' , column_names=['title', 'text'] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words lowerCAmelCase__ : Optional[Any] = dataset.map(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , num_proc=processing_args.num_proc ) # And compute the embeddings lowerCAmelCase__ : List[str] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : str = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) lowerCAmelCase__ : List[Any] = Features( {'text': Value('string' ), 'title': Value('string' ), 'embeddings': Sequence(Value('float32' ) )} ) # optional, save as float32 instead of float64 to save space lowerCAmelCase__ : List[Any] = dataset.map( partial(SCREAMING_SNAKE_CASE_ , ctx_encoder=SCREAMING_SNAKE_CASE_ , ctx_tokenizer=SCREAMING_SNAKE_CASE_ ) , batched=SCREAMING_SNAKE_CASE_ , batch_size=processing_args.batch_size , features=SCREAMING_SNAKE_CASE_ , ) # And finally save your dataset lowerCAmelCase__ : Optional[Any] = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset' ) dataset.save_to_disk(SCREAMING_SNAKE_CASE_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('Step 2 - Index the dataset' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search lowerCAmelCase__ : Optional[int] = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('embeddings' , custom_index=SCREAMING_SNAKE_CASE_ ) # And save the index lowerCAmelCase__ : str = os.path.join(rag_example_args.output_dir , 'my_knowledge_dataset_hnsw_index.faiss' ) dataset.get_index('embeddings' ).save(SCREAMING_SNAKE_CASE_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class A__ : lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' / 'my_knowledge_dataset.csv' ) , metadata={'help': 'Path to a tab-separated csv file with columns \'title\' and \'text\''} , ) lowercase = field( default=__magic_name__ , metadata={'help': 'Question that is passed as input to RAG. Default is \'What does Moses\' rod turn into ?\'.'} , ) lowercase = field( default='facebook/rag-sequence-nq' , metadata={'help': 'The RAG model to use. Either \'facebook/rag-sequence-nq\' or \'facebook/rag-token-nq\''} , ) lowercase = field( default='facebook/dpr-ctx_encoder-multiset-base' , metadata={ 'help': ( 'The DPR context encoder model to use. Either \'facebook/dpr-ctx_encoder-single-nq-base\' or' ' \'facebook/dpr-ctx_encoder-multiset-base\'' ) } , ) lowercase = field( default=str(Path(__magic_name__ ).parent / 'test_run' / 'dummy-kb' ) , metadata={'help': 'Path to a directory where the dataset passages and the index will be saved'} , ) @dataclass class A__ : lowercase = field( default=__magic_name__ , metadata={ 'help': 'The number of processes to use to split the documents into passages. Default is single process.' } , ) lowercase = field( default=16 , metadata={ 'help': 'The batch size to use when computing the passages embeddings using the DPR context encoder.' } , ) @dataclass class A__ : lowercase = field( default=768 , metadata={'help': 'The dimension of the embeddings to pass to the HNSW Faiss index.'} , ) lowercase = field( default=128 , metadata={ 'help': ( 'The number of bi-directional links created for every new element during the HNSW index construction.' ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
0
import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesseract, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class A__ : @staticmethod def _lowerCamelCase ( *a : List[Any] , **a : Any ): '''simple docstring''' pass @is_pipeline_test @require_vision @require_timm @require_torch class A__ ( unittest.TestCase ): lowercase = MODEL_FOR_OBJECT_DETECTION_MAPPING def _lowerCamelCase ( self : int , a : Any , a : Any , a : Any ): '''simple docstring''' lowerCAmelCase__ : Tuple = ObjectDetectionPipeline(model=a , image_processor=a ) return object_detector, ["./tests/fixtures/tests_samples/COCO/000000039769.png"] def _lowerCamelCase ( self : List[Any] , a : str , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Dict = object_detector('./tests/fixtures/tests_samples/COCO/000000039769.png' , threshold=0.0 ) self.assertGreater(len(a ) , 0 ) for detected_object in outputs: self.assertEqual( a , { 'score': ANY(a ), 'label': ANY(a ), 'box': {'xmin': ANY(a ), 'ymin': ANY(a ), 'xmax': ANY(a ), 'ymax': ANY(a )}, } , ) import datasets lowerCAmelCase__ : Any = datasets.load_dataset('hf-internal-testing/fixtures_image_utils' , 'image' , split='test' ) lowerCAmelCase__ : Union[str, Any] = [ Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ), 'http://images.cocodataset.org/val2017/000000039769.jpg', # RGBA dataset[0]['file'], # LA dataset[1]['file'], # L dataset[2]['file'], ] lowerCAmelCase__ : Union[str, Any] = object_detector(a , threshold=0.0 ) self.assertEqual(len(a ) , len(a ) ) for outputs in batch_outputs: self.assertGreater(len(a ) , 0 ) for detected_object in outputs: self.assertEqual( a , { 'score': ANY(a ), 'label': ANY(a ), 'box': {'xmin': ANY(a ), 'ymin': ANY(a ), 'xmax': ANY(a ), 'ymax': ANY(a )}, } , ) @require_tf @unittest.skip('Object detection not implemented in TF' ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' pass @require_torch def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'hf-internal-testing/tiny-detr-mobilenetsv3' lowerCAmelCase__ : Optional[Any] = AutoModelForObjectDetection.from_pretrained(a ) lowerCAmelCase__ : List[Any] = AutoFeatureExtractor.from_pretrained(a ) lowerCAmelCase__ : Optional[Any] = ObjectDetectionPipeline(model=a , feature_extractor=a ) lowerCAmelCase__ : Optional[int] = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' , threshold=0.0 ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, ] , ) lowerCAmelCase__ : Optional[int] = object_detector( [ 'http://images.cocodataset.org/val2017/000000039769.jpg', 'http://images.cocodataset.org/val2017/000000039769.jpg', ] , threshold=0.0 , ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ [ {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, ], [ {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, {'score': 0.3_3_7_6, 'label': 'LABEL_0', 'box': {'xmin': 159, 'ymin': 120, 'xmax': 480, 'ymax': 359}}, ], ] , ) @require_torch @slow def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = 'facebook/detr-resnet-50' lowerCAmelCase__ : Optional[Any] = AutoModelForObjectDetection.from_pretrained(a ) lowerCAmelCase__ : List[str] = AutoFeatureExtractor.from_pretrained(a ) lowerCAmelCase__ : Any = ObjectDetectionPipeline(model=a , feature_extractor=a ) lowerCAmelCase__ : List[str] = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ] , ) lowerCAmelCase__ : int = object_detector( [ 'http://images.cocodataset.org/val2017/000000039769.jpg', 'http://images.cocodataset.org/val2017/000000039769.jpg', ] ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], ] , ) @require_torch @slow def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Dict = 'facebook/detr-resnet-50' lowerCAmelCase__ : Tuple = pipeline('object-detection' , model=a ) lowerCAmelCase__ : List[str] = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ] , ) lowerCAmelCase__ : Tuple = object_detector( [ 'http://images.cocodataset.org/val2017/000000039769.jpg', 'http://images.cocodataset.org/val2017/000000039769.jpg', ] ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], [ {'score': 0.9_9_8_2, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 70, 'xmax': 175, 'ymax': 117}}, {'score': 0.9_9_6_0, 'label': 'remote', 'box': {'xmin': 333, 'ymin': 72, 'xmax': 368, 'ymax': 187}}, {'score': 0.9_9_5_5, 'label': 'couch', 'box': {'xmin': 0, 'ymin': 1, 'xmax': 639, 'ymax': 473}}, {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ], ] , ) @require_torch @slow def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = 0.9_9_8_5 lowerCAmelCase__ : str = 'facebook/detr-resnet-50' lowerCAmelCase__ : Optional[Any] = pipeline('object-detection' , model=a ) lowerCAmelCase__ : Dict = object_detector('http://images.cocodataset.org/val2017/000000039769.jpg' , threshold=a ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.9_9_8_8, 'label': 'cat', 'box': {'xmin': 13, 'ymin': 52, 'xmax': 314, 'ymax': 470}}, {'score': 0.9_9_8_7, 'label': 'cat', 'box': {'xmin': 345, 'ymin': 23, 'xmax': 640, 'ymax': 368}}, ] , ) @require_torch @require_pytesseract @slow def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Dict = 'Narsil/layoutlmv3-finetuned-funsd' lowerCAmelCase__ : str = 0.9_9_9_3 lowerCAmelCase__ : int = pipeline('object-detection' , model=a , threshold=a ) lowerCAmelCase__ : Union[str, Any] = object_detector( 'https://huggingface.co/spaces/impira/docquery/resolve/2359223c1837a7587402bda0f2643382a6eefeab/invoice.png' ) self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'score': 0.9_9_9_3, 'label': 'I-ANSWER', 'box': {'xmin': 294, 'ymin': 254, 'xmax': 343, 'ymax': 264}}, {'score': 0.9_9_9_3, 'label': 'I-ANSWER', 'box': {'xmin': 294, 'ymin': 254, 'xmax': 343, 'ymax': 264}}, ] , )
721
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A__ ( __magic_name__ ): lowercase = (DDPMParallelScheduler,) def _lowerCamelCase ( self : str , **a : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = { 'num_train_timesteps': 1_000, 'beta_start': 0.0_0_0_1, 'beta_end': 0.0_2, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _lowerCamelCase ( self : Tuple ): '''simple docstring''' for timesteps in [1, 5, 100, 1_000]: self.check_over_configs(num_train_timesteps=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _lowerCamelCase ( self : Any ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : List[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : List[str] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Any = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Tuple = self.dummy_model() lowerCAmelCase__ : Optional[Any] = self.dummy_sample_deter lowerCAmelCase__ : int = self.dummy_sample_deter + 0.1 lowerCAmelCase__ : Union[str, Any] = self.dummy_sample_deter - 0.1 lowerCAmelCase__ : Tuple = samplea.shape[0] lowerCAmelCase__ : List[Any] = torch.stack([samplea, samplea, samplea] , dim=0 ) lowerCAmelCase__ : Optional[Any] = torch.arange(a )[0:3, None].repeat(1 , a ) lowerCAmelCase__ : List[str] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) lowerCAmelCase__ : Tuple = scheduler.batch_step_no_noise(a , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) lowerCAmelCase__ : str = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowerCamelCase ( self : Any ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : List[Any] = self.get_scheduler_config() lowerCAmelCase__ : Dict = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Any = self.dummy_model() lowerCAmelCase__ : int = self.dummy_sample_deter lowerCAmelCase__ : Tuple = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : int = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : List[str] = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Optional[Any] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : str = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase__ : int = scheduler_class(**a ) lowerCAmelCase__ : str = len(a ) lowerCAmelCase__ : Optional[int] = self.dummy_model() lowerCAmelCase__ : List[str] = self.dummy_sample_deter lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowerCAmelCase__ : List[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowerCAmelCase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample lowerCAmelCase__ : str = pred_prev_sample lowerCAmelCase__ : Optional[int] = torch.sum(torch.abs(a ) ) lowerCAmelCase__ : Any = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase__ : Any = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : Optional[Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=a ) lowerCAmelCase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowerCAmelCase__ : Tuple = -1 else: lowerCAmelCase__ : Dict = timesteps[i + 1] lowerCAmelCase__ : str = scheduler.previous_timestep(a ) lowerCAmelCase__ : int = prev_t.item() self.assertEqual(a , a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : Optional[int] = self.get_scheduler_config() lowerCAmelCase__ : Optional[Any] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 51, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.scheduler_classes[0] lowerCAmelCase__ : str = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [100, 87, 50, 1, 0] lowerCAmelCase__ : int = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Dict = self.scheduler_classes[0] lowerCAmelCase__ : Dict = self.get_scheduler_config() lowerCAmelCase__ : Optional[int] = scheduler_class(**a ) lowerCAmelCase__ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
69
0
import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCallback, TrainingArguments, set_seed, ) def UpperCamelCase ( ): '''simple docstring''' A_ : Any = argparse.ArgumentParser() parser.add_argument('--model_ckpt' ,type=__lowercase ,default='microsoft/unixcoder-base-nine' ) parser.add_argument('--num_epochs' ,type=__lowercase ,default=5 ) parser.add_argument('--batch_size' ,type=__lowercase ,default=6 ) parser.add_argument('--gradient_accumulation_steps' ,type=__lowercase ,default=1 ) parser.add_argument('--freeze' ,type=__lowercase ,default=__lowercase ) parser.add_argument('--learning_rate' ,type=__lowercase ,default=5e-4 ) parser.add_argument('--seed' ,type=__lowercase ,default=0 ) parser.add_argument('--lr_scheduler_type' ,type=__lowercase ,default='cosine' ) parser.add_argument('--num_warmup_steps' ,type=__lowercase ,default=10 ) parser.add_argument('--weight_decay' ,type=__lowercase ,default=0.01 ) parser.add_argument('--output_dir' ,type=__lowercase ,default='./results' ) return parser.parse_args() _UpperCAmelCase = load("""accuracy""") def UpperCamelCase ( __lowercase : Optional[Any] ): '''simple docstring''' A_ , A_ : Tuple = eval_pred A_ : int = np.argmax(__lowercase ,axis=1 ) return metric.compute(predictions=__lowercase ,references=__lowercase ) class UpperCAmelCase ( __A ): '''simple docstring''' def __init__( self , lowercase ): """simple docstring""" super().__init__() A_ : Union[str, Any] = trainer def lowerCAmelCase_ ( self , lowercase , lowercase , lowercase , **lowercase ): """simple docstring""" if control.should_evaluate: A_ : Any = deepcopy(lowercase ) self._trainer.evaluate(eval_dataset=self._trainer.train_dataset , metric_key_prefix='train' ) return control_copy def UpperCamelCase ( ): '''simple docstring''' A_ : Optional[Any] = get_args() set_seed(args.seed ) A_ : List[str] = load_dataset('codeparrot/codecomplex' ,split='train' ) A_ : Union[str, Any] = dataset.train_test_split(test_size=0.2 ) A_ : Optional[int] = train_test['test'].train_test_split(test_size=0.5 ) A_ : Optional[int] = DatasetDict( { 'train': train_test['train'], 'test': test_validation['train'], 'valid': test_validation['test'], } ) print('Loading tokenizer and model' ) A_ : Dict = AutoTokenizer.from_pretrained(args.model_ckpt ) A_ : int = tokenizer.eos_token A_ : Union[str, Any] = AutoModelForSequenceClassification.from_pretrained(args.model_ckpt ,num_labels=7 ) A_ : Union[str, Any] = model.config.eos_token_id if args.freeze: for param in model.roberta.parameters(): A_ : List[Any] = False A_ : Dict = ClassLabel(num_classes=7 ,names=list(set(train_test_validation['train']['complexity'] ) ) ) def tokenize(__lowercase : Tuple ): A_ : List[Any] = tokenizer(example['src'] ,truncation=__lowercase ,max_length=10_24 ) A_ : Dict = labels.straint(example['complexity'] ) return { "input_ids": inputs["input_ids"], "attention_mask": inputs["attention_mask"], "label": label, } A_ : int = train_test_validation.map( __lowercase ,batched=__lowercase ,remove_columns=train_test_validation['train'].column_names ,) A_ : Union[str, Any] = DataCollatorWithPadding(tokenizer=__lowercase ) A_ : List[str] = TrainingArguments( output_dir=args.output_dir ,learning_rate=args.learning_rate ,lr_scheduler_type=args.lr_scheduler_type ,evaluation_strategy='epoch' ,save_strategy='epoch' ,logging_strategy='epoch' ,per_device_train_batch_size=args.batch_size ,per_device_eval_batch_size=args.batch_size ,num_train_epochs=args.num_epochs ,gradient_accumulation_steps=args.gradient_accumulation_steps ,weight_decay=0.01 ,metric_for_best_model='accuracy' ,run_name='complexity-java' ,report_to='wandb' ,) A_ : Union[str, Any] = Trainer( model=__lowercase ,args=__lowercase ,train_dataset=tokenized_datasets['train'] ,eval_dataset=tokenized_datasets['valid'] ,tokenizer=__lowercase ,data_collator=__lowercase ,compute_metrics=__lowercase ,) print('Training...' ) trainer.add_callback(CustomCallback(__lowercase ) ) trainer.train() if __name__ == "__main__": main()
70
from math import sqrt def UpperCamelCase ( __lowercase : int = 1_00_00_00 ): '''simple docstring''' A_ : int = 0 A_ : int = 0 A_ : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 ,2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowercase ,sum_shortest_sides // 2 ) - max(1 ,sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
70
1