code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import _LazyModule lowerCAmelCase : List[str] = {"""tokenization_byt5""": ["""ByT5Tokenizer"""]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys lowerCAmelCase : List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
704
'''simple docstring''' import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase : int = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = {"""vocab_file""": """spiece.model"""} lowerCAmelCase : Optional[int] = { """vocab_file""": { """AI-Sweden/gpt-sw3-126m""": """https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-350m""": """https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-1.6b""": """https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-6.7b""": """https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-20b""": """https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model""", } } lowerCAmelCase : Union[str, Any] = { """AI-Sweden/gpt-sw3-126m""": 20_48, """AI-Sweden/gpt-sw3-350m""": 20_48, """AI-Sweden/gpt-sw3-1.6b""": 20_48, """AI-Sweden/gpt-sw3-6.7b""": 20_48, """AI-Sweden/gpt-sw3-20b""": 20_48, } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = ["input_ids", "attention_mask"] def __init__( self , snake_case__ , snake_case__=False , snake_case__=False , snake_case__=False , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Tuple = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase : List[Any] = kwargs.get('name_or_path' ) if name_or_path is None: logger.warning( 'name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,' ' you are testing the model, this can safely be ignored' ) _lowerCAmelCase : Any = 'None' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing _lowerCAmelCase : str = '<|endoftext|>' if eos_token is None else eos_token _lowerCAmelCase : Tuple = '<unk>' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: _lowerCAmelCase : List[str] = unk_token if pad_token is None else pad_token _lowerCAmelCase : Optional[int] = eos_token if bos_token is None else bos_token else: _lowerCAmelCase : Tuple = '<pad>' if pad_token is None else pad_token _lowerCAmelCase : Union[str, Any] = '<s>' if bos_token is None else bos_token super().__init__( do_lower_case=snake_case__ , remove_space=snake_case__ , keep_accents=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , pad_token=snake_case__ , sp_model_kwargs=self.sp_model_kwargs , **snake_case__ , ) _lowerCAmelCase : Union[str, Any] = do_lower_case _lowerCAmelCase : Optional[int] = remove_space _lowerCAmelCase : Any = keep_accents _lowerCAmelCase : Optional[int] = vocab_file _lowerCAmelCase : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case__ ) # Used for whitespace normalization in input texts # fmt : off _lowerCAmelCase : Optional[Any] = {' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', '', '„'} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing _lowerCAmelCase : Optional[Any] = re.compile( F'[{"".join(map(snake_case__ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]' ) def __getstate__( self ): '''simple docstring''' _lowerCAmelCase : List[str] = self.__dict__.copy() _lowerCAmelCase : int = None return state def __setstate__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowerCAmelCase : int = {} _lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def a ( self ): '''simple docstring''' return len(self.sp_model ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.non_printing_characters_re.sub('' , snake_case__ ) # Normalize whitespaces _lowerCAmelCase : Tuple = ''.join([char if char not in self.whitespaces else ' ' for char in text] ) # NFC Unicode normalization _lowerCAmelCase : Union[str, Any] = unicodedata.normalize('NFC' , snake_case__ ) return text def a ( self , snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : str = self.preprocess_text(snake_case__ ) return self.sp_model.encode(snake_case__ , out_type=snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.PieceToId(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case__ ) @staticmethod def a ( snake_case__ ): '''simple docstring''' return out_string def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = [] _lowerCAmelCase : Optional[Any] = '' _lowerCAmelCase : Tuple = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(snake_case__ ) + token _lowerCAmelCase : Union[str, Any] = True _lowerCAmelCase : List[Any] = [] else: current_sub_tokens.append(snake_case__ ) _lowerCAmelCase : List[Any] = False out_string += self.sp_model.decode(snake_case__ ) return out_string def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = {self.convert_ids_to_tokens(snake_case__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCAmelCase : int = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case__ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case__ , 'wb' ) as fi: _lowerCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (out_vocab_file,) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' if isinstance(snake_case__ , snake_case__ ): _lowerCAmelCase : Optional[Any] = self.preprocess_text(snake_case__ ) _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) else: _lowerCAmelCase : Tuple = [self.preprocess_text(snake_case__ ) for t in text] _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) if return_tensors is True or return_tensors == "pt": _lowerCAmelCase : int = torch.tensor(snake_case__ ) return token_ids def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.decode(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = [F'User: {text}' if is_user else F'Bot: {text}' for is_user, text in conversation.iter_texts()] _lowerCAmelCase : str = ( F'{self.eos_token}{self.bos_token}' + F'{self.bos_token}'.join(snake_case__ ) + F'{self.bos_token}Bot:' ) return self.encode(text=snake_case__ )
630
0
import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class UpperCamelCase__ ( _UpperCAmelCase ): """simple docstring""" __magic_name__ = 42 __magic_name__ = 42 class UpperCamelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" __magic_name__ = 1 @register_to_config def __init__( self , snake_case__ = 2000 , snake_case__ = 0.15 , snake_case__ = 0.01 , snake_case__ = 1348.0 , snake_case__ = 1E-5 , snake_case__ = 1 , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = sigma_max # setable values _lowerCAmelCase : List[Any] = None self.set_sigmas(A_ , A_ , A_ , A_ ) def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' return sample def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : Tuple = sampling_eps if sampling_eps is not None else self.config.sampling_eps _lowerCAmelCase : Tuple = torch.linspace(1 , A_ , A_ , device=A_ ) def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : Tuple = sigma_min if sigma_min is not None else self.config.sigma_min _lowerCAmelCase : Tuple = sigma_max if sigma_max is not None else self.config.sigma_max _lowerCAmelCase : str = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(A_ , A_ ) _lowerCAmelCase : Dict = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) _lowerCAmelCase : List[Any] = torch.exp(torch.linspace(math.log(A_ ) , math.log(A_ ) , A_ ) ) _lowerCAmelCase : Optional[int] = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = True , ): '''simple docstring''' if self.timesteps is None: raise ValueError( '`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler' ) _lowerCAmelCase : int = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) _lowerCAmelCase : Optional[int] = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda _lowerCAmelCase : Any = timesteps.to(self.discrete_sigmas.device ) _lowerCAmelCase : Optional[int] = self.discrete_sigmas[timesteps].to(sample.device ) _lowerCAmelCase : Optional[Any] = self.get_adjacent_sigma(A_ , A_ ).to(sample.device ) _lowerCAmelCase : str = torch.zeros_like(A_ ) _lowerCAmelCase : str = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods _lowerCAmelCase : Optional[Any] = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): _lowerCAmelCase : Optional[Any] = diffusion.unsqueeze(-1 ) _lowerCAmelCase : List[Any] = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of _lowerCAmelCase : Optional[int] = randn_tensor( sample.shape , layout=sample.layout , generator=A_ , device=sample.device , dtype=sample.dtype ) _lowerCAmelCase : int = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? _lowerCAmelCase : Union[str, Any] = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=A_ , prev_sample_mean=A_ ) def a ( self , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = True , ): '''simple docstring''' if self.timesteps is None: raise ValueError( '`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler' ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction _lowerCAmelCase : Dict = randn_tensor(sample.shape , layout=sample.layout , generator=A_ ).to(sample.device ) # compute step size from the model_output, the noise, and the snr _lowerCAmelCase : Any = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() _lowerCAmelCase : Optional[Any] = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() _lowerCAmelCase : Tuple = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 _lowerCAmelCase : str = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term _lowerCAmelCase : List[Any] = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): _lowerCAmelCase : Any = step_size.unsqueeze(-1 ) _lowerCAmelCase : Optional[Any] = sample + step_size * model_output _lowerCAmelCase : Any = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=A_ ) def a ( self , snake_case__ , snake_case__ , snake_case__ , ): '''simple docstring''' _lowerCAmelCase : str = timesteps.to(original_samples.device ) _lowerCAmelCase : List[Any] = self.discrete_sigmas.to(original_samples.device )[timesteps] _lowerCAmelCase : Union[str, Any] = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(A_ ) * sigmas[:, None, None, None] ) _lowerCAmelCase : List[str] = noise + original_samples return noisy_samples def __len__( self ): '''simple docstring''' return self.config.num_train_timesteps
705
'''simple docstring''' import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = (DDPMScheduler,) def a ( self , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**snake_case__ ) return config def a ( self ): '''simple docstring''' for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=snake_case__ ) def a ( self ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=snake_case__ , beta_end=snake_case__ ) def a ( self ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=snake_case__ ) def a ( self ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=snake_case__ ) def a ( self ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=snake_case__ ) def a ( self ): '''simple docstring''' self.check_over_configs(thresholding=snake_case__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=snake_case__ , prediction_type=snake_case__ , sample_max_value=snake_case__ , ) def a ( self ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=snake_case__ ) def a ( self ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1E-5 def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : str = self.dummy_model() _lowerCAmelCase : str = self.dummy_sample_deter _lowerCAmelCase : Union[str, Any] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : List[Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Any = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Dict = pred_prev_sample _lowerCAmelCase : Dict = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : List[str] = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 258.9606 ) < 1E-2 assert abs(result_mean.item() - 0.3372 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.scheduler_classes[0] _lowerCAmelCase : Any = self.get_scheduler_config(prediction_type='v_prediction' ) _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : Any = self.dummy_model() _lowerCAmelCase : Tuple = self.dummy_sample_deter _lowerCAmelCase : Optional[int] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : Union[str, Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Dict = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Tuple = pred_prev_sample _lowerCAmelCase : Any = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : Dict = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 202.0296 ) < 1E-2 assert abs(result_mean.item() - 0.2631 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[int] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=snake_case__ ) _lowerCAmelCase : Union[str, Any] = scheduler.timesteps for i, timestep in enumerate(snake_case__ ): if i == len(snake_case__ ) - 1: _lowerCAmelCase : str = -1 else: _lowerCAmelCase : Optional[Any] = timesteps[i + 1] _lowerCAmelCase : int = scheduler.previous_timestep(snake_case__ ) _lowerCAmelCase : int = prev_t.item() self.assertEqual(snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : Tuple = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 51, 0] with self.assertRaises(snake_case__ , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : List[str] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 1, 0] _lowerCAmelCase : int = len(snake_case__ ) with self.assertRaises(snake_case__ , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=snake_case__ , timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : int = self.get_scheduler_config() _lowerCAmelCase : Any = scheduler_class(**snake_case__ ) _lowerCAmelCase : Any = [scheduler.config.num_train_timesteps] with self.assertRaises( snake_case__ , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=snake_case__ )
630
0
'''simple docstring''' import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class UpperCamelCase__ ( __lowercase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = dataset _lowerCAmelCase : List[str] = process _lowerCAmelCase : List[Any] = params def __len__( self ): '''simple docstring''' return len(self.dataset ) def __getitem__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[str] = self.dataset[i] _lowerCAmelCase : Union[str, Any] = self.process(snake_case__ , **self.params ) return processed class UpperCamelCase__ ( __lowercase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__=None ): '''simple docstring''' _lowerCAmelCase : int = loader _lowerCAmelCase : Optional[int] = infer _lowerCAmelCase : Any = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _lowerCAmelCase : Optional[int] = None _lowerCAmelCase : Any = loader_batch_size # Internal bookkeeping _lowerCAmelCase : Optional[Any] = None _lowerCAmelCase : Union[str, Any] = None def __len__( self ): '''simple docstring''' return len(self.loader ) def __iter__( self ): '''simple docstring''' _lowerCAmelCase : str = iter(self.loader ) return self def a ( self ): '''simple docstring''' if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice _lowerCAmelCase : List[Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _lowerCAmelCase : Tuple = {} for k, element in self._loader_batch_data.items(): if isinstance(snake_case__ , snake_case__ ): # Convert ModelOutput to tuple first _lowerCAmelCase : List[str] = element.to_tuple() if isinstance(element[0] , torch.Tensor ): _lowerCAmelCase : Optional[int] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _lowerCAmelCase : str = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(snake_case__ , snake_case__ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): _lowerCAmelCase : Optional[Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _lowerCAmelCase : int = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _lowerCAmelCase : Optional[Any] = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _lowerCAmelCase : int = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _lowerCAmelCase : Tuple = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. _lowerCAmelCase : str = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _lowerCAmelCase : Tuple = self._loader_batch_data.__class__(snake_case__ ) self._loader_batch_index += 1 return result def a ( self ): '''simple docstring''' if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _lowerCAmelCase : Any = next(self.iterator ) _lowerCAmelCase : Optional[int] = self.infer(snake_case__ , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(snake_case__ , torch.Tensor ): _lowerCAmelCase : Optional[Any] = processed else: _lowerCAmelCase : Tuple = list(processed.keys() )[0] _lowerCAmelCase : List[str] = processed[key] if isinstance(snake_case__ , snake_case__ ): _lowerCAmelCase : str = len(snake_case__ ) else: _lowerCAmelCase : Any = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _lowerCAmelCase : int = observed_batch_size # Setting internal index to unwrap the batch _lowerCAmelCase : Tuple = processed _lowerCAmelCase : List[Any] = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class UpperCamelCase__ ( __lowercase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__=None ): '''simple docstring''' super().__init__(snake_case__ , snake_case__ , snake_case__ ) def __iter__( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = iter(self.loader ) _lowerCAmelCase : Optional[int] = None return self def a ( self ): '''simple docstring''' if self.subiterator is None: _lowerCAmelCase : Union[str, Any] = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item _lowerCAmelCase : List[str] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _lowerCAmelCase : Union[str, Any] = self.infer(next(self.iterator ) , **self.params ) _lowerCAmelCase : List[Any] = next(self.subiterator ) return processed class UpperCamelCase__ ( __lowercase ): """simple docstring""" def __iter__( self ): '''simple docstring''' _lowerCAmelCase : List[str] = iter(self.loader ) return self def a ( self ): '''simple docstring''' _lowerCAmelCase : str = False _lowerCAmelCase : List[Any] = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _lowerCAmelCase : Optional[Any] = self.loader_batch_item() _lowerCAmelCase : Union[str, Any] = item.pop('is_last' ) accumulator.append(snake_case__ ) if is_last: return accumulator while not is_last: _lowerCAmelCase : Any = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(snake_case__ , torch.Tensor ): _lowerCAmelCase : str = processed else: _lowerCAmelCase : Union[str, Any] = list(processed.keys() )[0] _lowerCAmelCase : Tuple = processed[key] if isinstance(snake_case__ , snake_case__ ): _lowerCAmelCase : str = len(snake_case__ ) else: _lowerCAmelCase : Optional[int] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _lowerCAmelCase : Optional[int] = observed_batch_size _lowerCAmelCase : Tuple = processed _lowerCAmelCase : List[str] = 0 while self._loader_batch_index < self.loader_batch_size: _lowerCAmelCase : List[str] = self.loader_batch_item() _lowerCAmelCase : str = item.pop('is_last' ) accumulator.append(snake_case__ ) if is_last: return accumulator else: _lowerCAmelCase : Optional[int] = processed _lowerCAmelCase : List[Any] = item.pop('is_last' ) accumulator.append(snake_case__ ) return accumulator class UpperCamelCase__ ( __lowercase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = dataset _lowerCAmelCase : Union[str, Any] = key def __len__( self ): '''simple docstring''' return len(self.dataset ) def __getitem__( self , snake_case__ ): '''simple docstring''' return self.dataset[i][self.key] class UpperCamelCase__ ( __lowercase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = dataset _lowerCAmelCase : Optional[int] = keya _lowerCAmelCase : Dict = keya def __len__( self ): '''simple docstring''' return len(self.dataset ) def __getitem__( self , snake_case__ ): '''simple docstring''' return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
706
'''simple docstring''' import socket def lowercase (): """simple docstring""" _lowerCAmelCase : Tuple = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCAmelCase : Optional[int] = socket.gethostname() _lowerCAmelCase : Any = 1_2_3_1_2 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCAmelCase : Union[str, Any] = sock.recv(1_0_2_4 ) if not data: break out_file.write(_A ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
630
0
'''simple docstring''' import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets lowerCAmelCase : Any = "\\n@article{hendrycksmath2021,\n title={Measuring Mathematical Problem Solving With the MATH Dataset},\n author={Dan Hendrycks\n and Collin Burns\n and Saurav Kadavath\n and Akul Arora\n and Steven Basart\n and Eric Tang\n and Dawn Song\n and Jacob Steinhardt},\n journal={arXiv preprint arXiv:2103.03874},\n year={2021}\n}\n" lowerCAmelCase : Optional[Any] = "\\nThis metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset.\nIt first canonicalizes the inputs (e.g., converting \"1/2\" to \"\\frac{1}{2}\") and then computes accuracy.\n" lowerCAmelCase : Tuple = r"\nCalculates accuracy after canonicalizing inputs.\n\nArgs:\n predictions: list of predictions to score. Each prediction\n is a string that contains natural language and LaTex.\n references: list of reference for each prediction. Each\n reference is a string that contains natural language\n and LaTex.\nReturns:\n accuracy: accuracy after canonicalizing inputs\n (e.g., converting \"1/2\" to \"\\frac{1}{2}\")\n\nExamples:\n >>> metric = datasets.load_metric(\"competition_math\")\n >>> results = metric.compute(references=[\"\\frac{1}{2}\"], predictions=[\"1/2\"])\n >>> print(results)\n {'accuracy': 1.0}\n" @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase__ ( datasets.Metric ): """simple docstring""" def a ( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/hendrycks/math' , codebase_urls=['https://github.com/hendrycks/math'] , ) def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = 0.0 for i, j in zip(__UpperCamelCase , __UpperCamelCase ): n_correct += 1.0 if math_equivalence.is_equiv(__UpperCamelCase , __UpperCamelCase ) else 0.0 _lowerCAmelCase : Optional[int] = n_correct / len(__UpperCamelCase ) return { "accuracy": accuracy, }
707
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase : Tuple = False lowerCAmelCase : str = True lowerCAmelCase : List[Any] = False if __name__ == "__main__": lowerCAmelCase : Any = argparse.ArgumentParser() parser.add_argument( """--repo_path""", default=None, type=str, required=True, help="""The config json file corresponding to the architecture.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") lowerCAmelCase : Optional[int] = parser.parse_args() lowerCAmelCase : int = { """image_size""": """sample_size""", """num_res_blocks""": """layers_per_block""", """block_channels""": """block_out_channels""", """down_blocks""": """down_block_types""", """up_blocks""": """up_block_types""", """downscale_freq_shift""": """freq_shift""", """resnet_num_groups""": """norm_num_groups""", """resnet_act_fn""": """act_fn""", """resnet_eps""": """norm_eps""", """num_head_channels""": """attention_head_dim""", } lowerCAmelCase : int = { """time_steps""": """time_proj""", """mid""": """mid_block""", """downsample_blocks""": """down_blocks""", """upsample_blocks""": """up_blocks""", } lowerCAmelCase : Optional[Any] = """""" if has_file(args.repo_path, """config.json""") else """unet""" with open(os.path.join(args.repo_path, subfolder, """config.json"""), """r""", encoding="""utf-8""") as reader: lowerCAmelCase : int = reader.read() lowerCAmelCase : List[str] = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, """config.json"""): lowerCAmelCase : str = UNetaDModel(**config) else: lowerCAmelCase : Union[str, Any] = UNetaDConditionModel if """ldm-text2im-large-256""" in args.repo_path else UNetaDModel lowerCAmelCase : Dict = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase : Union[str, Any] = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase : str = config[key] del config[key] lowerCAmelCase : Optional[int] = [k.replace("""UNetRes""", """""") for k in config["""down_block_types"""]] lowerCAmelCase : Dict = [k.replace("""UNetRes""", """""") for k in config["""up_block_types"""]] if do_only_weights: lowerCAmelCase : Tuple = torch.load(os.path.join(args.repo_path, subfolder, """diffusion_pytorch_model.bin""")) lowerCAmelCase : str = {} for param_key, param_value in state_dict.items(): if param_key.endswith(""".op.bias""") or param_key.endswith(""".op.weight"""): continue lowerCAmelCase : str = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split(""".""")[0] == key: lowerCAmelCase : Dict = param_value lowerCAmelCase : Tuple = True if not has_changed: lowerCAmelCase : Tuple = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
630
0
'''simple docstring''' import cva import numpy as np class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__ ): '''simple docstring''' if k in (0.04, 0.06): _lowerCAmelCase : int = k _lowerCAmelCase : Optional[int] = window_size else: raise ValueError('invalid k value' ) def __str__( self ): '''simple docstring''' return str(self.k ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = cva.imread(snake_case__ , 0 ) _lowerCAmelCase : Optional[Any] = img.shape _lowerCAmelCase : list[list[int]] = [] _lowerCAmelCase : Optional[int] = img.copy() _lowerCAmelCase : str = cva.cvtColor(snake_case__ , cva.COLOR_GRAY2RGB ) _lowerCAmelCase : Optional[Any] = np.gradient(snake_case__ ) _lowerCAmelCase : Any = dx**2 _lowerCAmelCase : Tuple = dy**2 _lowerCAmelCase : str = dx * dy _lowerCAmelCase : Union[str, Any] = 0.04 _lowerCAmelCase : List[str] = self.window_size // 2 for y in range(snake_case__ , h - offset ): for x in range(snake_case__ , w - offset ): _lowerCAmelCase : Union[str, Any] = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _lowerCAmelCase : Dict = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _lowerCAmelCase : int = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _lowerCAmelCase : Optional[Any] = (wxx * wyy) - (wxy**2) _lowerCAmelCase : Optional[Any] = wxx + wyy _lowerCAmelCase : Optional[Any] = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": lowerCAmelCase : Tuple = HarrisCorner(0.04, 3) lowerCAmelCase , lowerCAmelCase : List[Any] = edge_detect.detect("""path_to_image""") cva.imwrite("""detect.png""", color_img)
708
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = None ): '''simple docstring''' super().__init__() _lowerCAmelCase : Union[str, Any] = pad_token_id _lowerCAmelCase : List[Any] = max_length _lowerCAmelCase : Tuple = vocab _lowerCAmelCase : str = merges _lowerCAmelCase : List[str] = BytePairTokenizer(snake_case__ , snake_case__ , sequence_length=snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = [' '.join(snake_case__ ) for m in tokenizer.bpe_ranks.keys()] _lowerCAmelCase : Any = tokenizer.get_vocab() return cls(snake_case__ , snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = GPTaTokenizer.from_pretrained(snake_case__ , *snake_case__ , **snake_case__ ) return cls.from_tokenizer(snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ ): '''simple docstring''' return cls(**snake_case__ ) def a ( self ): '''simple docstring''' return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : str = self.tf_tokenizer(snake_case__ ) _lowerCAmelCase : str = tf.ones_like(snake_case__ ) if self.pad_token_id is not None: # pad the tokens up to max length _lowerCAmelCase : Optional[int] = max_length if max_length is not None else self.max_length if max_length is not None: _lowerCAmelCase , _lowerCAmelCase : str = pad_model_inputs( snake_case__ , max_seq_length=snake_case__ , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
630
0
'''simple docstring''' import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def lowercase (_A ): """simple docstring""" assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def lowercase (): """simple docstring""" assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def lowercase (): """simple docstring""" _lowerCAmelCase : int = 'mock-s3-bucket' _lowerCAmelCase : str = f's3://{mock_bucket}' _lowerCAmelCase : Any = extract_path_from_uri(lowerCAmelCase__ ) assert dataset_path.startswith('s3://' ) is False _lowerCAmelCase : List[Any] = './local/path' _lowerCAmelCase : str = extract_path_from_uri(lowerCAmelCase__ ) assert dataset_path == new_dataset_path def lowercase (_A ): """simple docstring""" _lowerCAmelCase : int = is_remote_filesystem(lowerCAmelCase__ ) assert is_remote is True _lowerCAmelCase : Union[str, Any] = fsspec.filesystem('file' ) _lowerCAmelCase : List[str] = is_remote_filesystem(lowerCAmelCase__ ) assert is_remote is False @pytest.mark.parametrize('compression_fs_class' , lowerCAmelCase__ ) def lowercase (_A , _A , _A , _A , _A , _A , _A ): """simple docstring""" _lowerCAmelCase : List[str] = {'gzip': gz_file, 'xz': xz_file, 'zstd': zstd_file, 'bz2': bza_file, 'lz4': lza_file} _lowerCAmelCase : int = input_paths[compression_fs_class.protocol] if input_path is None: _lowerCAmelCase : int = f'for \'{compression_fs_class.protocol}\' compression protocol, ' if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(lowerCAmelCase__ ) _lowerCAmelCase : List[Any] = fsspec.filesystem(compression_fs_class.protocol , fo=lowerCAmelCase__ ) assert isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) _lowerCAmelCase : Tuple = os.path.basename(lowerCAmelCase__ ) _lowerCAmelCase : Any = expected_filename[: expected_filename.rindex('.' )] assert fs.glob('*' ) == [expected_filename] with fs.open(lowerCAmelCase__ , 'r' , encoding='utf-8' ) as f, open(lowerCAmelCase__ , encoding='utf-8' ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize('protocol' , ['zip', 'gzip'] ) def lowercase (_A , _A , _A ): """simple docstring""" _lowerCAmelCase : Dict = {'zip': zip_jsonl_path, 'gzip': jsonl_gz_path} _lowerCAmelCase : int = compressed_file_paths[protocol] _lowerCAmelCase : List[Any] = 'dataset.jsonl' _lowerCAmelCase : Any = f'{protocol}://{member_file_path}::{compressed_file_path}' _lowerCAmelCase : List[str] = fsspec.get_fs_token_paths(lowerCAmelCase__ ) assert fs.isfile(lowerCAmelCase__ ) assert not fs.isfile('non_existing_' + member_file_path ) @pytest.mark.integration def lowercase (_A , _A , _A , _A ): """simple docstring""" _lowerCAmelCase : Tuple = hf_api.dataset_info(lowerCAmelCase__ , token=lowerCAmelCase__ ) _lowerCAmelCase : Union[str, Any] = HfFileSystem(repo_info=lowerCAmelCase__ , token=lowerCAmelCase__ ) assert sorted(hffs.glob('*' ) ) == [".gitattributes", "data"] assert hffs.isdir('data' ) assert hffs.isfile('.gitattributes' ) and hffs.isfile('data/text_data.txt' ) with open(lowerCAmelCase__ ) as f: assert hffs.open('data/text_data.txt' , 'r' ).read() == f.read() def lowercase (): """simple docstring""" _lowerCAmelCase : int = 'bz2' # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(lowerCAmelCase__ , lowerCAmelCase__ , clobber=lowerCAmelCase__ ) with pytest.warns(lowerCAmelCase__ ) as warning_info: importlib.reload(datasets.filesystems ) assert len(lowerCAmelCase__ ) == 1 assert ( str(warning_info[0].message ) == f'A filesystem protocol was already set for {protocol} and will be overwritten.' )
709
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Optional[int] = { """configuration_mctct""": ["""MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MCTCTConfig"""], """feature_extraction_mctct""": ["""MCTCTFeatureExtractor"""], """processing_mctct""": ["""MCTCTProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : List[str] = [ """MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MCTCTForCTC""", """MCTCTModel""", """MCTCTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys lowerCAmelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
710
'''simple docstring''' lowerCAmelCase : Optional[int] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_00_00)] def lowercase (_A ): """simple docstring""" _lowerCAmelCase : str = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 1_0_0_0_0_0] number //= 1_0_0_0_0_0 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution lowerCAmelCase : list[bool | None] = [None] * 10_00_00_00 lowerCAmelCase : List[str] = True lowerCAmelCase : Union[str, Any] = False def lowercase (_A ): """simple docstring""" if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _lowerCAmelCase : Any = chain(next_number(_A ) ) _lowerCAmelCase : List[str] = number_chain while number < 1_0_0_0_0_0_0_0: _lowerCAmelCase : Tuple = number_chain number *= 1_0 return number_chain def lowercase (_A = 1_0_0_0_0_0_0_0 ): """simple docstring""" for i in range(1 , _A ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(_A ) if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution() = }''')
630
0
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = FlaxXLMRobertaModel.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Any = AutoTokenizer.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Optional[int] = 'The dog is cute and lives in the garden house' _lowerCAmelCase : List[str] = jnp.array([tokenizer.encode(UpperCamelCase__ )] ) _lowerCAmelCase : Optional[int] = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim _lowerCAmelCase : Tuple = jnp.array( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) _lowerCAmelCase : Optional[Any] = model(UpperCamelCase__ )['last_hidden_state'] self.assertEqual(output.shape , UpperCamelCase__ ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , UpperCamelCase__ , atol=1E-3 ) )
711
'''simple docstring''' import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(snake_case__ , 'width_multiplier' ) ) class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=64 , snake_case__=2 , snake_case__=3 , snake_case__="swish" , snake_case__=3 , snake_case__=32 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=True , snake_case__=True , snake_case__=10 , snake_case__=None , snake_case__=0.25 , snake_case__=0.0 , snake_case__=0.0 , ): '''simple docstring''' _lowerCAmelCase : Tuple = parent _lowerCAmelCase : Optional[int] = batch_size _lowerCAmelCase : List[Any] = image_size _lowerCAmelCase : List[Any] = patch_size _lowerCAmelCase : List[str] = num_channels _lowerCAmelCase : Optional[Any] = make_divisible(512 * width_multiplier , divisor=8 ) _lowerCAmelCase : Optional[Any] = hidden_act _lowerCAmelCase : List[Any] = conv_kernel_size _lowerCAmelCase : Optional[Any] = output_stride _lowerCAmelCase : List[Any] = classifier_dropout_prob _lowerCAmelCase : str = use_labels _lowerCAmelCase : List[str] = is_training _lowerCAmelCase : Optional[int] = num_labels _lowerCAmelCase : List[str] = initializer_range _lowerCAmelCase : str = scope _lowerCAmelCase : Any = width_multiplier _lowerCAmelCase : Union[str, Any] = ffn_dropout _lowerCAmelCase : Optional[int] = attn_dropout def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase : Optional[Any] = None _lowerCAmelCase : Dict = None if self.use_labels: _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _lowerCAmelCase : Optional[int] = self.get_config() return config, pixel_values, labels, pixel_labels def a ( self ): '''simple docstring''' return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : str = model(snake_case__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.num_labels _lowerCAmelCase : List[Any] = MobileViTVaForImageClassification(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : int = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.num_labels _lowerCAmelCase : Optional[int] = MobileViTVaForSemanticSegmentation(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Dict = model(snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) _lowerCAmelCase : Any = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Optional[int] = config_and_inputs _lowerCAmelCase : Tuple = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = ( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) __magic_name__ = ( { "feature-extraction": MobileViTVaModel, "image-classification": MobileViTVaForImageClassification, "image-segmentation": MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False def a ( self ): '''simple docstring''' _lowerCAmelCase : int = MobileViTVaModelTester(self ) _lowerCAmelCase : Dict = MobileViTVaConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ ) def a ( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='MobileViTV2 does not use inputs_embeds' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not support input and output embeddings' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not output attentions' ) def a ( self ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason='Got `CUDA error: misaligned address` for tests after this one being run.' ) def a ( self ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : str = model_class(snake_case__ ) _lowerCAmelCase : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCAmelCase : int = [*signature.parameters.keys()] _lowerCAmelCase : Any = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def a ( self ): '''simple docstring''' def check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ): _lowerCAmelCase : Dict = model_class(snake_case__ ) model.to(snake_case__ ) model.eval() with torch.no_grad(): _lowerCAmelCase : Optional[Any] = model(**self._prepare_for_class(snake_case__ , snake_case__ ) ) _lowerCAmelCase : List[str] = outputs.hidden_states _lowerCAmelCase : List[str] = 5 self.assertEqual(len(snake_case__ ) , snake_case__ ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. _lowerCAmelCase : List[Any] = 2 for i in range(len(snake_case__ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : Optional[int] = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCAmelCase : Any = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*snake_case__ ) @slow def a ( self ): '''simple docstring''' for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase : Dict = MobileViTVaModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) def lowercase (): """simple docstring""" _lowerCAmelCase : Any = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def a ( self ): '''simple docstring''' return ( MobileViTImageProcessor.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ) if is_vision_available() else None ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = MobileViTVaForImageClassification.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ).to( snake_case__ ) _lowerCAmelCase : str = self.default_image_processor _lowerCAmelCase : Any = prepare_img() _lowerCAmelCase : Optional[int] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Tuple = model(**snake_case__ ) # verify the logits _lowerCAmelCase : Optional[Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , snake_case__ ) _lowerCAmelCase : Tuple = torch.tensor([-1.6336E00, -7.3204E-02, -5.1883E-01] ).to(snake_case__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Any = model.to(snake_case__ ) _lowerCAmelCase : int = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Optional[int] = prepare_img() _lowerCAmelCase : Dict = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : int = model(**snake_case__ ) _lowerCAmelCase : Dict = outputs.logits # verify the logits _lowerCAmelCase : str = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , snake_case__ ) _lowerCAmelCase : Any = torch.tensor( [ [[7.0863, 7.1525, 6.8201], [6.6931, 6.8770, 6.8933], [6.2978, 7.0366, 6.9636]], [[-3.7134, -3.6712, -3.6675], [-3.5825, -3.3549, -3.4777], [-3.3435, -3.3979, -3.2857]], [[-2.9329, -2.8003, -2.7369], [-3.0564, -2.4780, -2.0207], [-2.6889, -1.9298, -1.7640]], ] , device=snake_case__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : List[Any] = model.to(snake_case__ ) _lowerCAmelCase : str = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Tuple = prepare_img() _lowerCAmelCase : List[str] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Any = model(**snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.logits.detach().cpu() _lowerCAmelCase : Any = image_processor.post_process_semantic_segmentation(outputs=snake_case__ , target_sizes=[(50, 60)] ) _lowerCAmelCase : List[Any] = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , snake_case__ ) _lowerCAmelCase : List[str] = image_processor.post_process_semantic_segmentation(outputs=snake_case__ ) _lowerCAmelCase : Tuple = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , snake_case__ )
630
0
'''simple docstring''' lowerCAmelCase : str = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] lowerCAmelCase : List[Any] = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] lowerCAmelCase : Union[str, Any] = { 0: "Sunday", 1: "Monday", 2: "Tuesday", 3: "Wednesday", 4: "Thursday", 5: "Friday", 6: "Saturday", } def lowercase (_A , _A , _A ): """simple docstring""" assert len(str(__lowerCAmelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 1_2, "month should be between 1 to 12" assert 1 <= day <= 3_1, "day should be between 1 to 31" # Doomsday algorithm: _lowerCAmelCase : int = year // 1_0_0 _lowerCAmelCase : int = (5 * (century % 4) + 2) % 7 _lowerCAmelCase : Optional[int] = year % 1_0_0 _lowerCAmelCase : List[Any] = centurian % 1_2 _lowerCAmelCase : Optional[int] = ( (centurian // 1_2) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 _lowerCAmelCase : Union[str, Any] = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 4_0_0) == 0) else DOOMSDAY_LEAP[month - 1] ) _lowerCAmelCase : Tuple = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
712
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = FlaxXLMRobertaModel.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Dict = 'The dog is cute and lives in the garden house' _lowerCAmelCase : List[str] = jnp.array([tokenizer.encode(snake_case__ )] ) _lowerCAmelCase : Optional[int] = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim _lowerCAmelCase : Tuple = jnp.array( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) _lowerCAmelCase : Union[str, Any] = model(snake_case__ )['last_hidden_state'] self.assertEqual(output.shape , snake_case__ ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , snake_case__ , atol=1E-3 ) )
630
0
'''simple docstring''' import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class UpperCamelCase__ ( nn.Module ): """simple docstring""" def __init__( self ): '''simple docstring''' super().__init__() _lowerCAmelCase : Any = nn.Linear(3 , 4 ) _lowerCAmelCase : Tuple = nn.BatchNormad(4 ) _lowerCAmelCase : List[Any] = nn.Linear(4 , 5 ) def a ( self , snake_case__ ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(_UpperCAmelCase ) ) ) class UpperCamelCase__ ( _UpperCamelCase ): """simple docstring""" def a ( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' return (args[0] + 1,) + args[1:], kwargs class UpperCamelCase__ ( _UpperCamelCase ): """simple docstring""" def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' return output + 1 class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = ModelForTest() _lowerCAmelCase : Dict = ModelHook() add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase ) self.assertEqual(test_model._hf_hook , _UpperCAmelCase ) self.assertTrue(hasattr(_UpperCAmelCase , '_old_forward' ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , 'forward' ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ['x'] ) remove_hook_from_module(_UpperCAmelCase ) self.assertFalse(hasattr(_UpperCAmelCase , '_hf_hook' ) ) self.assertFalse(hasattr(_UpperCAmelCase , '_old_forward' ) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = ModelForTest() _lowerCAmelCase : List[str] = ModelHook() add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase ) add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase , append=_UpperCAmelCase ) self.assertEqual(isinstance(test_model._hf_hook , _UpperCAmelCase ) , _UpperCAmelCase ) self.assertEqual(len(test_model._hf_hook.hooks ) , 2 ) self.assertTrue(hasattr(_UpperCAmelCase , '_old_forward' ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , 'forward' ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ['x'] ) remove_hook_from_module(_UpperCAmelCase ) self.assertFalse(hasattr(_UpperCAmelCase , '_hf_hook' ) ) self.assertFalse(hasattr(_UpperCAmelCase , '_old_forward' ) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = ModelForTest() _lowerCAmelCase : List[Any] = torch.randn(2 , 3 ) _lowerCAmelCase : Optional[Any] = test_model(x + 1 ) _lowerCAmelCase : Optional[int] = test_model(x + 2 ) _lowerCAmelCase : Optional[int] = PreForwardHook() add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : Tuple = test_model(_UpperCAmelCase ) self.assertTrue(torch.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _lowerCAmelCase : Union[str, Any] = PreForwardHook() add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : Optional[int] = test_model(_UpperCAmelCase ) self.assertTrue(torch.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _lowerCAmelCase : List[str] = SequentialHook(PreForwardHook() , PreForwardHook() ) add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : List[str] = test_model(_UpperCAmelCase ) assert torch.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-5 ) def a ( self ): '''simple docstring''' _lowerCAmelCase : str = ModelForTest() _lowerCAmelCase : Tuple = torch.randn(2 , 3 ) _lowerCAmelCase : List[Any] = test_model(_UpperCAmelCase ) _lowerCAmelCase : int = PostForwardHook() add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : List[str] = test_model(_UpperCAmelCase ) self.assertTrue(torch.allclose(_UpperCAmelCase , output + 1 , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _lowerCAmelCase : Any = PostForwardHook() add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : Dict = test_model(_UpperCAmelCase ) self.assertTrue(torch.allclose(_UpperCAmelCase , output + 1 , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _lowerCAmelCase : List[Any] = SequentialHook(PostForwardHook() , PostForwardHook() ) add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : List[Any] = test_model(_UpperCAmelCase ) assert torch.allclose(_UpperCAmelCase , output + 2 , atol=1E-5 ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = ModelForTest() _lowerCAmelCase : Optional[int] = torch.randn(2 , 3 ) _lowerCAmelCase : int = test_model(_UpperCAmelCase ) _lowerCAmelCase : List[Any] = PostForwardHook() add_hook_to_module(_UpperCAmelCase , _UpperCAmelCase ) _lowerCAmelCase : Union[str, Any] = test_model(_UpperCAmelCase ) self.assertTrue(torch.allclose(_UpperCAmelCase , output + 1 ) ) self.assertTrue(outputa.requires_grad ) _lowerCAmelCase : Union[str, Any] = True _lowerCAmelCase : List[Any] = test_model(_UpperCAmelCase ) self.assertFalse(outputa.requires_grad ) @require_multi_gpu def a ( self ): '''simple docstring''' _lowerCAmelCase : str = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1 ) ) self.assertEqual(model.lineara.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0 ) ) self.assertEqual(model.lineara.weight.device , torch.device(1 ) ) # We can still make a forward pass. The input does not need to be on any particular device _lowerCAmelCase : Dict = torch.randn(2 , 3 ) _lowerCAmelCase : Dict = model(_UpperCAmelCase ) self.assertEqual(output.device , torch.device(1 ) ) # We can add a general hook to put back output on same device as input. add_hook_to_module(_UpperCAmelCase , AlignDevicesHook(io_same_device=_UpperCAmelCase ) ) _lowerCAmelCase : Union[str, Any] = torch.randn(2 , 3 ).to(0 ) _lowerCAmelCase : Optional[int] = model(_UpperCAmelCase ) self.assertEqual(output.device , torch.device(0 ) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # This will move each submodule on different devices _lowerCAmelCase : int = {'''execution_device''': 0 if torch.cuda.is_available() else '''cpu''', '''offload''': True} add_hook_to_module(model.lineara , AlignDevicesHook(**_UpperCAmelCase ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**_UpperCAmelCase ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**_UpperCAmelCase ) ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) # Buffers are not included in the offload by default, so are on the execution device _lowerCAmelCase : Tuple = torch.device(hook_kwargs['execution_device'] ) self.assertEqual(model.batchnorm.running_mean.device , _UpperCAmelCase ) _lowerCAmelCase : Tuple = torch.randn(2 , 3 ) _lowerCAmelCase : Optional[int] = model(_UpperCAmelCase ) self.assertEqual(output.device , _UpperCAmelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # Now test with buffers included in the offload _lowerCAmelCase : Optional[Any] = { '''execution_device''': 0 if torch.cuda.is_available() else '''cpu''', '''offload''': True, '''offload_buffers''': True, } add_hook_to_module(model.lineara , AlignDevicesHook(**_UpperCAmelCase ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**_UpperCAmelCase ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**_UpperCAmelCase ) ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device('meta' ) ) _lowerCAmelCase : int = torch.randn(2 , 3 ) _lowerCAmelCase : Tuple = model(_UpperCAmelCase ) self.assertEqual(output.device , _UpperCAmelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # This will move each submodule on different devices _lowerCAmelCase : str = 0 if torch.cuda.is_available() else '''cpu''' attach_align_device_hook(_UpperCAmelCase , execution_device=_UpperCAmelCase , offload=_UpperCAmelCase ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) # Buffers are not included in the offload by default, so are on the execution device _lowerCAmelCase : List[str] = torch.device(_UpperCAmelCase ) self.assertEqual(model.batchnorm.running_mean.device , _UpperCAmelCase ) _lowerCAmelCase : Dict = torch.randn(2 , 3 ) _lowerCAmelCase : Dict = model(_UpperCAmelCase ) self.assertEqual(output.device , _UpperCAmelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(_UpperCAmelCase ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # Now test with buffers included in the offload attach_align_device_hook(_UpperCAmelCase , execution_device=_UpperCAmelCase , offload=_UpperCAmelCase , offload_buffers=_UpperCAmelCase ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device('meta' ) ) _lowerCAmelCase : Tuple = torch.randn(2 , 3 ) _lowerCAmelCase : str = model(_UpperCAmelCase ) self.assertEqual(output.device , _UpperCAmelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(_UpperCAmelCase ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # This will move each submodule on different devices _lowerCAmelCase : List[Any] = 0 if torch.cuda.is_available() else '''cpu''' attach_align_device_hook( _UpperCAmelCase , execution_device=_UpperCAmelCase , offload=_UpperCAmelCase , weights_map=model.state_dict() ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) # Buffers are not included in the offload by default, so are on the execution device _lowerCAmelCase : List[Any] = torch.device(_UpperCAmelCase ) self.assertEqual(model.batchnorm.running_mean.device , _UpperCAmelCase ) _lowerCAmelCase : Dict = torch.randn(2 , 3 ) _lowerCAmelCase : Union[str, Any] = model(_UpperCAmelCase ) self.assertEqual(output.device , _UpperCAmelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(_UpperCAmelCase ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) # Now test with buffers included in the offload attach_align_device_hook( _UpperCAmelCase , execution_device=_UpperCAmelCase , offload=_UpperCAmelCase , weights_map=model.state_dict() , offload_buffers=_UpperCAmelCase , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('meta' ) ) self.assertEqual(model.lineara.weight.device , torch.device('meta' ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device('meta' ) ) _lowerCAmelCase : Optional[Any] = torch.randn(2 , 3 ) _lowerCAmelCase : Optional[Any] = model(_UpperCAmelCase ) self.assertEqual(output.device , _UpperCAmelCase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(_UpperCAmelCase ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('cpu' ) ) self.assertEqual(model.lineara.weight.device , torch.device('cpu' ) )
713
'''simple docstring''' def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Dict = len(_A ) while cur > 1: # Find the maximum number in arr _lowerCAmelCase : int = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi _lowerCAmelCase : Dict = arr[mi::-1] + arr[mi + 1 : len(_A )] # Reverse whole list _lowerCAmelCase : Optional[int] = arr[cur - 1 :: -1] + arr[cur : len(_A )] cur -= 1 return arr if __name__ == "__main__": lowerCAmelCase : Tuple = input("""Enter numbers separated by a comma:\n""").strip() lowerCAmelCase : Tuple = [int(item) for item in user_input.split(""",""")] print(pancake_sort(unsorted))
630
0
import math def lowercase (_A = 1_0_0 ): """simple docstring""" _lowerCAmelCase : Any = sum(i * i for i in range(1 , n + 1 ) ) _lowerCAmelCase : Optional[int] = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F'''{solution() = }''')
714
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) lowerCAmelCase : str = { """EleutherAI/gpt-j-6B""": """https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json""", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "gptj" __magic_name__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , snake_case__=5_0400 , snake_case__=2048 , snake_case__=4096 , snake_case__=28 , snake_case__=16 , snake_case__=64 , snake_case__=None , snake_case__="gelu_new" , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=1E-5 , snake_case__=0.02 , snake_case__=True , snake_case__=5_0256 , snake_case__=5_0256 , snake_case__=False , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : int = vocab_size _lowerCAmelCase : Optional[int] = n_positions _lowerCAmelCase : Optional[int] = n_embd _lowerCAmelCase : Optional[int] = n_layer _lowerCAmelCase : str = n_head _lowerCAmelCase : Tuple = n_inner _lowerCAmelCase : Tuple = rotary_dim _lowerCAmelCase : Optional[int] = activation_function _lowerCAmelCase : Any = resid_pdrop _lowerCAmelCase : List[str] = embd_pdrop _lowerCAmelCase : int = attn_pdrop _lowerCAmelCase : Any = layer_norm_epsilon _lowerCAmelCase : Optional[int] = initializer_range _lowerCAmelCase : List[str] = use_cache _lowerCAmelCase : Dict = bos_token_id _lowerCAmelCase : Any = eos_token_id super().__init__( bos_token_id=snake_case__ , eos_token_id=snake_case__ , tie_word_embeddings=snake_case__ , **snake_case__ ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = "default" , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' super().__init__(snake_case__ , task=snake_case__ , patching_specs=snake_case__ , use_past=snake_case__ ) if not getattr(self._config , 'pad_token_id' , snake_case__ ): # TODO: how to do that better? _lowerCAmelCase : Any = 0 @property def a ( self ): '''simple docstring''' _lowerCAmelCase : str = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(snake_case__ , direction='inputs' ) _lowerCAmelCase : int = {0: 'batch', 1: 'past_sequence + sequence'} else: _lowerCAmelCase : int = {0: 'batch', 1: 'sequence'} return common_inputs @property def a ( self ): '''simple docstring''' return self._config.n_layer @property def a ( self ): '''simple docstring''' return self._config.n_head def a ( self , snake_case__ , snake_case__ = -1 , snake_case__ = -1 , snake_case__ = False , snake_case__ = None , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = super(snake_case__ , self ).generate_dummy_inputs( snake_case__ , batch_size=snake_case__ , seq_length=snake_case__ , is_pair=snake_case__ , framework=snake_case__ ) # We need to order the input in the way they appears in the forward() _lowerCAmelCase : Any = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = common_inputs['input_ids'].shape # Not using the same length for past_key_values _lowerCAmelCase : Any = seqlen + 2 _lowerCAmelCase : Optional[int] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _lowerCAmelCase : Tuple = [ (torch.zeros(snake_case__ ), torch.zeros(snake_case__ )) for _ in range(self.num_layers ) ] _lowerCAmelCase : Tuple = common_inputs['attention_mask'] if self.use_past: _lowerCAmelCase : Any = ordered_inputs['attention_mask'].dtype _lowerCAmelCase : Union[str, Any] = torch.cat( [ordered_inputs['attention_mask'], torch.ones(snake_case__ , snake_case__ , dtype=snake_case__ )] , dim=1 ) return ordered_inputs @property def a ( self ): '''simple docstring''' return 13
630
0
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging lowerCAmelCase : Tuple = logging.get_logger(__name__) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = ["input_features"] def __init__( self , snake_case__=80 , snake_case__=1_6000 , snake_case__=160 , snake_case__=30 , snake_case__=400 , snake_case__=0.0 , snake_case__=False , **snake_case__ , ): '''simple docstring''' super().__init__( feature_size=__UpperCamelCase , sampling_rate=__UpperCamelCase , padding_value=__UpperCamelCase , return_attention_mask=__UpperCamelCase , **__UpperCamelCase , ) _lowerCAmelCase : List[Any] = n_fft _lowerCAmelCase : List[Any] = hop_length _lowerCAmelCase : Dict = chunk_length _lowerCAmelCase : Dict = chunk_length * sampling_rate _lowerCAmelCase : Union[str, Any] = self.n_samples // hop_length _lowerCAmelCase : Union[str, Any] = sampling_rate _lowerCAmelCase : int = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=__UpperCamelCase , min_frequency=0.0 , max_frequency=8000.0 , sampling_rate=__UpperCamelCase , norm='slaney' , mel_scale='slaney' , ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = spectrogram( __UpperCamelCase , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel='log10' , ) _lowerCAmelCase : List[Any] = log_spec[:, :-1] _lowerCAmelCase : Optional[Any] = np.maximum(__UpperCamelCase , log_spec.max() - 8.0 ) _lowerCAmelCase : Optional[int] = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def a ( snake_case__ , snake_case__ , snake_case__ = 0.0 ): '''simple docstring''' if attention_mask is not None: _lowerCAmelCase : List[str] = np.array(__UpperCamelCase , np.intaa ) _lowerCAmelCase : List[str] = [] for vector, length in zip(__UpperCamelCase , attention_mask.sum(-1 ) ): _lowerCAmelCase : Union[str, Any] = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: _lowerCAmelCase : List[str] = padding_value normed_input_values.append(__UpperCamelCase ) else: _lowerCAmelCase : Any = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__( self , snake_case__ , snake_case__ = True , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = "max_length" , snake_case__ = None , snake_case__ = None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a' F' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input' F' was sampled with {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) _lowerCAmelCase : Tuple = isinstance(__UpperCamelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'Only mono-channel audio is supported for input to {self}' ) _lowerCAmelCase : Optional[int] = is_batched_numpy or ( isinstance(__UpperCamelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: _lowerCAmelCase : Dict = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__UpperCamelCase , np.ndarray ): _lowerCAmelCase : Optional[int] = np.asarray(__UpperCamelCase , dtype=np.floataa ) elif isinstance(__UpperCamelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): _lowerCAmelCase : List[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: _lowerCAmelCase : int = [np.asarray([raw_speech] ).T] _lowerCAmelCase : List[Any] = BatchFeature({'input_features': raw_speech} ) # convert into correct format for padding _lowerCAmelCase : Tuple = self.pad( __UpperCamelCase , padding=__UpperCamelCase , max_length=max_length if max_length else self.n_samples , truncation=__UpperCamelCase , pad_to_multiple_of=__UpperCamelCase , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: _lowerCAmelCase : Any = self.zero_mean_unit_var_norm( padded_inputs['input_features'] , attention_mask=padded_inputs['attention_mask'] , padding_value=self.padding_value , ) _lowerCAmelCase : List[Any] = np.stack(padded_inputs['input_features'] , axis=0 ) # make sure list is in array format _lowerCAmelCase : Dict = padded_inputs.get('input_features' ).transpose(2 , 0 , 1 ) _lowerCAmelCase : Dict = [self._np_extract_fbank_features(__UpperCamelCase ) for waveform in input_features[0]] if isinstance(input_features[0] , __UpperCamelCase ): _lowerCAmelCase : Dict = [np.asarray(__UpperCamelCase , dtype=np.floataa ) for feature in input_features] else: _lowerCAmelCase : List[Any] = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) _lowerCAmelCase : Optional[int] = padded_inputs['attention_mask'][:, :: self.hop_length] if return_tensors is not None: _lowerCAmelCase : Optional[int] = padded_inputs.convert_to_tensors(__UpperCamelCase ) return padded_inputs def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = copy.deepcopy(self.__dict__ ) _lowerCAmelCase : Tuple = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
715
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Any = { """configuration_x_clip""": [ """XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XCLIPConfig""", """XCLIPTextConfig""", """XCLIPVisionConfig""", ], """processing_x_clip""": ["""XCLIPProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """XCLIPModel""", """XCLIPPreTrainedModel""", """XCLIPTextModel""", """XCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys lowerCAmelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' from typing import List from .keymap import KEYMAP, get_character def lowercase (_A ): """simple docstring""" def decorator(_A ): _lowerCAmelCase : str = getattr(_UpperCamelCase , 'handle_key' , [] ) handle += [key] setattr(_UpperCamelCase , 'handle_key' , _UpperCamelCase ) return func return decorator def lowercase (*_A ): """simple docstring""" def decorator(_A ): _lowerCAmelCase : Any = getattr(_UpperCamelCase , 'handle_key' , [] ) handle += keys setattr(_UpperCamelCase , 'handle_key' , _UpperCamelCase ) return func return decorator class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __new__( cls , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = super().__new__(cls , snake_case__ , snake_case__ , snake_case__ ) if not hasattr(snake_case__ , 'key_handler' ): setattr(snake_case__ , 'key_handler' , {} ) setattr(snake_case__ , 'handle_input' , KeyHandler.handle_input ) for value in attrs.values(): _lowerCAmelCase : Tuple = getattr(snake_case__ , 'handle_key' , [] ) for key in handled_keys: _lowerCAmelCase : Union[str, Any] = value return new_cls @staticmethod def a ( cls ): '''simple docstring''' _lowerCAmelCase : Any = get_character() if char != KEYMAP["undefined"]: _lowerCAmelCase : Optional[int] = ord(snake_case__ ) _lowerCAmelCase : Union[str, Any] = cls.key_handler.get(snake_case__ ) if handler: _lowerCAmelCase : Tuple = char return handler(cls ) else: return None def lowercase (cls ): """simple docstring""" return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
716
'''simple docstring''' import math from datetime import datetime, timedelta def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Optional[Any] = year % 1_9 _lowerCAmelCase : Any = year % 4 _lowerCAmelCase : Optional[int] = year % 7 _lowerCAmelCase : int = math.floor(year / 1_0_0 ) _lowerCAmelCase : Dict = math.floor((1_3 + 8 * leap_day_inhibits) / 2_5 ) _lowerCAmelCase : Optional[Any] = leap_day_inhibits / 4 _lowerCAmelCase : Dict = ( 1_5 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 3_0 _lowerCAmelCase : List[Any] = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 _lowerCAmelCase : Dict = (1_9 * metonic_cycle + secular_moon_shift) % 3_0 # PHM -> Paschal Full Moon _lowerCAmelCase : Union[str, Any] = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 2_9 and days_from_phm_to_sunday == 6: return datetime(_A , 4 , 1_9 ) elif days_to_add == 2_8 and days_from_phm_to_sunday == 6: return datetime(_A , 4 , 1_8 ) else: return datetime(_A , 3 , 2_2 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (19_94, 20_00, 20_10, 20_21, 20_23): lowerCAmelCase : List[str] = """will be""" if year > datetime.now().year else """was""" print(F'''Easter in {year} {tense} {gauss_easter(year)}''')
630
0
'''simple docstring''' from __future__ import annotations import math class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = size # approximate the overall size of segment tree with given value _lowerCAmelCase : int = [0 for i in range(0 , 4 * size )] # create array to store lazy update _lowerCAmelCase : Dict = [0 for i in range(0 , 4 * size )] _lowerCAmelCase : Optional[Any] = [0 for i in range(0 , 4 * size )] # flag for lazy update def a ( self , snake_case__ ): '''simple docstring''' return idx * 2 def a ( self , snake_case__ ): '''simple docstring''' return idx * 2 + 1 def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' if left_element == right_element: _lowerCAmelCase : Optional[int] = a[left_element - 1] else: _lowerCAmelCase : Tuple = (left_element + right_element) // 2 self.build(self.left(snake_case__ ) , snake_case__ , snake_case__ , snake_case__ ) self.build(self.right(snake_case__ ) , mid + 1 , snake_case__ , snake_case__ ) _lowerCAmelCase : List[Any] = max( self.segment_tree[self.left(snake_case__ )] , self.segment_tree[self.right(snake_case__ )] ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' if self.flag[idx] is True: _lowerCAmelCase : List[Any] = self.lazy[idx] _lowerCAmelCase : Union[str, Any] = False if left_element != right_element: _lowerCAmelCase : Any = self.lazy[idx] _lowerCAmelCase : Optional[int] = self.lazy[idx] _lowerCAmelCase : Optional[Any] = True _lowerCAmelCase : Optional[Any] = True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: _lowerCAmelCase : List[Any] = val if left_element != right_element: _lowerCAmelCase : Dict = val _lowerCAmelCase : Union[str, Any] = val _lowerCAmelCase : Tuple = True _lowerCAmelCase : List[Any] = True return True _lowerCAmelCase : Optional[Any] = (left_element + right_element) // 2 self.update(self.left(snake_case__ ) , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) self.update(self.right(snake_case__ ) , mid + 1 , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) _lowerCAmelCase : Any = max( self.segment_tree[self.left(snake_case__ )] , self.segment_tree[self.right(snake_case__ )] ) return True def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' if self.flag[idx] is True: _lowerCAmelCase : Union[str, Any] = self.lazy[idx] _lowerCAmelCase : Dict = False if left_element != right_element: _lowerCAmelCase : Any = self.lazy[idx] _lowerCAmelCase : List[str] = self.lazy[idx] _lowerCAmelCase : Union[str, Any] = True _lowerCAmelCase : Optional[int] = True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] _lowerCAmelCase : Optional[int] = (left_element + right_element) // 2 _lowerCAmelCase : Any = self.query(self.left(snake_case__ ) , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) _lowerCAmelCase : Tuple = self.query(self.right(snake_case__ ) , mid + 1 , snake_case__ , snake_case__ , snake_case__ ) return max(snake_case__ , snake_case__ ) def __str__( self ): '''simple docstring''' return str([self.query(1 , 1 , self.size , snake_case__ , snake_case__ ) for i in range(1 , self.size + 1 )] ) if __name__ == "__main__": lowerCAmelCase : str = [1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] lowerCAmelCase : str = 15 lowerCAmelCase : Union[str, Any] = SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 1_11) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 2_35) print(segt)
717
'''simple docstring''' import unittest from knapsack import greedy_knapsack as kp class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [10, 20, 30, 40, 50, 60] _lowerCAmelCase : Union[str, Any] = [2, 4, 6, 8, 10, 12] _lowerCAmelCase : Dict = 100 self.assertEqual(kp.calc_profit(snake_case__ , snake_case__ , snake_case__ ) , 210 ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'max_weight must greater than zero.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'Weight can not be negative.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'Profit can not be negative.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'max_weight must greater than zero.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex( snake_case__ , 'The length of profit and weight must be same.' ) if __name__ == "__main__": unittest.main()
630
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Tuple = { "configuration_upernet": ["UperNetConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Union[str, Any] = [ "UperNetForSemanticSegmentation", "UperNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_upernet import UperNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_upernet import UperNetForSemanticSegmentation, UperNetPreTrainedModel else: import sys lowerCAmelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
718
'''simple docstring''' def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : Optional[int] = (boundary[1] - boundary[0]) / steps _lowerCAmelCase : Any = boundary[0] _lowerCAmelCase : List[str] = boundary[1] _lowerCAmelCase : Tuple = make_points(_A , _A , _A ) _lowerCAmelCase : Tuple = 0.0 y += (h / 2.0) * f(_A ) for i in x_i: # print(i) y += h * f(_A ) y += (h / 2.0) * f(_A ) return y def lowercase (_A , _A , _A ): """simple docstring""" _lowerCAmelCase : Tuple = a + h while x < (b - h): yield x _lowerCAmelCase : Any = x + h def lowercase (_A ): # enter your function here """simple docstring""" _lowerCAmelCase : int = (x - 0) * (x - 0) return y def lowercase (): """simple docstring""" _lowerCAmelCase : Optional[Any] = 0.0 # Lower bound of integration _lowerCAmelCase : Dict = 1.0 # Upper bound of integration _lowerCAmelCase : Optional[Any] = 10.0 # define number of steps or resolution _lowerCAmelCase : Optional[int] = [a, b] # define boundary of integration _lowerCAmelCase : List[Any] = method_a(_A , _A ) print(f'y = {y}' ) if __name__ == "__main__": main()
630
0
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) lowerCAmelCase : Tuple = '▁' lowerCAmelCase : Optional[Any] = {'vocab_file': 'sentencepiece.bpe.model'} lowerCAmelCase : str = { 'vocab_file': { 'facebook/mbart-large-50-one-to-many-mmt': ( 'https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt/resolve/main/sentencepiece.bpe.model' ), } } lowerCAmelCase : List[str] = { 'facebook/mbart-large-50-one-to-many-mmt': 10_24, } # fmt: off lowerCAmelCase : Union[str, Any] = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN', 'af_ZA', 'az_AZ', 'bn_IN', 'fa_IR', 'he_IL', 'hr_HR', 'id_ID', 'ka_GE', 'km_KH', 'mk_MK', 'ml_IN', 'mn_MN', 'mr_IN', 'pl_PL', 'ps_AF', 'pt_XX', 'sv_SE', 'sw_KE', 'ta_IN', 'te_IN', 'th_TH', 'tl_XX', 'uk_UA', 'ur_PK', 'xh_ZA', 'gl_ES', 'sl_SI'] class UpperCamelCase__ ( UpperCamelCase_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = ["input_ids", "attention_mask"] __magic_name__ = [] __magic_name__ = [] def __init__( self , snake_case__ , snake_case__=None , snake_case__=None , snake_case__="</s>" , snake_case__="</s>" , snake_case__="<s>" , snake_case__="<unk>" , snake_case__="<pad>" , snake_case__="<mask>" , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Dict = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token _lowerCAmelCase : Dict = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase : Dict = kwargs.get('additional_special_tokens' , [] ) kwargs["additional_special_tokens"] += [ code for code in FAIRSEQ_LANGUAGE_CODES if code not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=snake_case__ , tgt_lang=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , cls_token=snake_case__ , pad_token=snake_case__ , mask_token=snake_case__ , sp_model_kwargs=self.sp_model_kwargs , **snake_case__ , ) _lowerCAmelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(snake_case__ ) ) _lowerCAmelCase : int = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token _lowerCAmelCase : List[str] = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _lowerCAmelCase : int = 1 _lowerCAmelCase : Dict = len(self.sp_model ) _lowerCAmelCase : int = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(snake_case__ ) } _lowerCAmelCase : int = {v: k for k, v in self.lang_code_to_id.items()} _lowerCAmelCase : int = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) _lowerCAmelCase : List[Any] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} _lowerCAmelCase : Optional[Any] = src_lang if src_lang is not None else '''en_XX''' _lowerCAmelCase : Tuple = self.lang_code_to_id[self._src_lang] _lowerCAmelCase : List[str] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def a ( self ): '''simple docstring''' return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def a ( self ): '''simple docstring''' return self._src_lang @src_lang.setter def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self ): '''simple docstring''' _lowerCAmelCase : Tuple = self.__dict__.copy() _lowerCAmelCase : Dict = None return state def __setstate__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowerCAmelCase : List[str] = {} _lowerCAmelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = {self.convert_ids_to_tokens(snake_case__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.encode(snake_case__ , out_type=snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _lowerCAmelCase : int = self.sp_model.PieceToId(snake_case__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def a ( self , snake_case__ ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = [] _lowerCAmelCase : Any = '''''' _lowerCAmelCase : Dict = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(snake_case__ ) + token _lowerCAmelCase : Tuple = True _lowerCAmelCase : Optional[int] = [] else: current_sub_tokens.append(snake_case__ ) _lowerCAmelCase : Tuple = False out_string += self.sp_model.decode(snake_case__ ) return out_string.strip() def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCAmelCase : Dict = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case__ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case__ , 'wb' ) as fi: _lowerCAmelCase : Tuple = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (out_vocab_file,) def a ( self , snake_case__ , snake_case__ = None , snake_case__ = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case__ , token_ids_a=snake_case__ , already_has_special_tokens=snake_case__ ) _lowerCAmelCase : Optional[Any] = [1] * len(self.prefix_tokens ) _lowerCAmelCase : Union[str, Any] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(snake_case__ )) + suffix_ones return prefix_ones + ([0] * len(snake_case__ )) + ([0] * len(snake_case__ )) + suffix_ones def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , **snake_case__ ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) _lowerCAmelCase : Dict = src_lang _lowerCAmelCase : Union[str, Any] = self(snake_case__ , add_special_tokens=snake_case__ , return_tensors=snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[Any] = self.convert_tokens_to_ids(snake_case__ ) _lowerCAmelCase : Optional[int] = tgt_lang_id return inputs def a ( self , snake_case__ , snake_case__ = "en_XX" , snake_case__ = None , snake_case__ = "ro_RO" , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : str = src_lang _lowerCAmelCase : int = tgt_lang return super().prepare_seqaseq_batch(snake_case__ , snake_case__ , **snake_case__ ) def a ( self ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def a ( self ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[str] = self.lang_code_to_id[src_lang] _lowerCAmelCase : Dict = [self.cur_lang_code_id] _lowerCAmelCase : Union[str, Any] = [self.eos_token_id] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.lang_code_to_id[tgt_lang] _lowerCAmelCase : Tuple = [self.cur_lang_code_id] _lowerCAmelCase : List[Any] = [self.eos_token_id]
719
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) lowerCAmelCase : int = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys lowerCAmelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase : Dict = logging.get_logger(__name__) class UpperCamelCase__ ( lowercase_ ): """simple docstring""" __magic_name__ = ['''pixel_values'''] def __init__( self , snake_case__ = True , snake_case__ = None , snake_case__ = PIL.Image.BICUBIC , snake_case__ = True , snake_case__ = None , snake_case__ = 1 / 255 , snake_case__ = True , snake_case__ = True , snake_case__ = None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Optional[Any] = size if size is not None else {'height': 256, 'width': 256} _lowerCAmelCase : Optional[int] = get_size_dict(snake_case__ ) _lowerCAmelCase : Tuple = crop_size if crop_size is not None else {'height': 224, 'width': 224} _lowerCAmelCase : Optional[Any] = get_size_dict(snake_case__ , param_name='crop_size' ) _lowerCAmelCase : str = do_resize _lowerCAmelCase : str = size _lowerCAmelCase : Any = resample _lowerCAmelCase : str = do_center_crop _lowerCAmelCase : str = crop_size _lowerCAmelCase : List[Any] = do_rescale _lowerCAmelCase : List[Any] = rescale_factor _lowerCAmelCase : Tuple = do_normalize _lowerCAmelCase : List[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _lowerCAmelCase : Any = image_std if image_std is not None else IMAGENET_STANDARD_STD def a ( self , snake_case__ , snake_case__ , snake_case__ = PIL.Image.BICUBIC , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : List[Any] = get_size_dict(snake_case__ ) if "height" not in size or "width" not in size: raise ValueError(F'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return resize( snake_case__ , size=(size['height'], size['width']) , resample=snake_case__ , data_format=snake_case__ , **snake_case__ ) def a ( self , snake_case__ , snake_case__ , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = get_size_dict(snake_case__ ) if "height" not in size or "width" not in size: raise ValueError(F'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return center_crop(snake_case__ , size=(size['height'], size['width']) , data_format=snake_case__ , **snake_case__ ) def a ( self , snake_case__ , snake_case__ , snake_case__ = None , **snake_case__ , ): '''simple docstring''' return rescale(snake_case__ , scale=snake_case__ , data_format=snake_case__ , **snake_case__ ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = None , **snake_case__ , ): '''simple docstring''' return normalize(snake_case__ , mean=snake_case__ , std=snake_case__ , data_format=snake_case__ , **snake_case__ ) def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__=None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = ChannelDimension.FIRST , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = do_resize if do_resize is not None else self.do_resize _lowerCAmelCase : Dict = resample if resample is not None else self.resample _lowerCAmelCase : List[str] = do_center_crop if do_center_crop is not None else self.do_center_crop _lowerCAmelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale _lowerCAmelCase : str = rescale_factor if rescale_factor is not None else self.rescale_factor _lowerCAmelCase : str = do_normalize if do_normalize is not None else self.do_normalize _lowerCAmelCase : Dict = image_mean if image_mean is not None else self.image_mean _lowerCAmelCase : Tuple = image_std if image_std is not None else self.image_std _lowerCAmelCase : List[str] = size if size is not None else self.size _lowerCAmelCase : Union[str, Any] = get_size_dict(snake_case__ ) _lowerCAmelCase : Union[str, Any] = crop_size if crop_size is not None else self.crop_size _lowerCAmelCase : Dict = get_size_dict(snake_case__ , param_name='crop_size' ) _lowerCAmelCase : Optional[int] = make_list_of_images(snake_case__ ) if not valid_images(snake_case__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. _lowerCAmelCase : str = [to_numpy_array(snake_case__ ) for image in images] if do_resize: _lowerCAmelCase : str = [self.resize(image=snake_case__ , size=snake_case__ , resample=snake_case__ ) for image in images] if do_center_crop: _lowerCAmelCase : str = [self.center_crop(image=snake_case__ , size=snake_case__ ) for image in images] if do_rescale: _lowerCAmelCase : Optional[Any] = [self.rescale(image=snake_case__ , scale=snake_case__ ) for image in images] if do_normalize: _lowerCAmelCase : int = [self.normalize(image=snake_case__ , mean=snake_case__ , std=snake_case__ ) for image in images] _lowerCAmelCase : Optional[Any] = [to_channel_dimension_format(snake_case__ , snake_case__ ) for image in images] _lowerCAmelCase : Dict = {'pixel_values': images} return BatchFeature(data=snake_case__ , tensor_type=snake_case__ )
720
'''simple docstring''' from collections import Counter from timeit import timeit def lowercase (_A = "" , ): """simple docstring""" return sum(c % 2 for c in Counter(input_str.replace(' ' , '' ).lower() ).values() ) < 2 def lowercase (_A = "" ): """simple docstring""" if len(_A ) == 0: return True _lowerCAmelCase : Union[str, Any] = input_str.replace(' ' , '' ).lower() # character_freq_dict: Stores the frequency of every character in the input string _lowerCAmelCase : dict[str, int] = {} for character in lower_case_input_str: _lowerCAmelCase : Union[str, Any] = character_freq_dict.get(_A , 0 ) + 1 _lowerCAmelCase : List[Any] = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def lowercase (_A = "" ): """simple docstring""" print('\nFor string = ' , _A , ':' ) print( '> can_string_be_rearranged_as_palindrome_counter()' , '\tans =' , can_string_be_rearranged_as_palindrome_counter(_A ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome_counter(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) print( '> can_string_be_rearranged_as_palindrome()' , '\tans =' , can_string_be_rearranged_as_palindrome(_A ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) if __name__ == "__main__": lowerCAmelCase : Tuple = input( """Enter string to determine if it can be rearranged as a palindrome or not: """ ).strip() benchmark(check_str) lowerCAmelCase : Optional[Any] = can_string_be_rearranged_as_palindrome_counter(check_str) print(F'''{check_str} can {"" if status else "not "}be rearranged as a palindrome''')
630
0
'''simple docstring''' import numpy as np class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None ): '''simple docstring''' self.set_matricies(red=snake_case__ , green=snake_case__ , blue=snake_case__ , red_edge=snake_case__ , nir=snake_case__ ) def a ( self , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None ): '''simple docstring''' if red is not None: _lowerCAmelCase : List[Any] = red if green is not None: _lowerCAmelCase : Any = green if blue is not None: _lowerCAmelCase : Any = blue if red_edge is not None: _lowerCAmelCase : str = red_edge if nir is not None: _lowerCAmelCase : Union[str, Any] = nir return True def a ( self , snake_case__="" , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None ): '''simple docstring''' self.set_matricies(red=snake_case__ , green=snake_case__ , blue=snake_case__ , red_edge=snake_case__ , nir=snake_case__ ) _lowerCAmelCase : str = { 'ARVI2': self.arvaa, 'CCCI': self.ccci, 'CVI': self.cvi, 'GLI': self.gli, 'NDVI': self.ndvi, 'BNDVI': self.bndvi, 'redEdgeNDVI': self.red_edge_ndvi, 'GNDVI': self.gndvi, 'GBNDVI': self.gbndvi, 'GRNDVI': self.grndvi, 'RBNDVI': self.rbndvi, 'PNDVI': self.pndvi, 'ATSAVI': self.atsavi, 'BWDRVI': self.bwdrvi, 'CIgreen': self.ci_green, 'CIrededge': self.ci_rededge, 'CI': self.ci, 'CTVI': self.ctvi, 'GDVI': self.gdvi, 'EVI': self.evi, 'GEMI': self.gemi, 'GOSAVI': self.gosavi, 'GSAVI': self.gsavi, 'Hue': self.hue, 'IVI': self.ivi, 'IPVI': self.ipvi, 'I': self.i, 'RVI': self.rvi, 'MRVI': self.mrvi, 'MSAVI': self.m_savi, 'NormG': self.norm_g, 'NormNIR': self.norm_nir, 'NormR': self.norm_r, 'NGRDI': self.ngrdi, 'RI': self.ri, 'S': self.s, 'IF': self._if, 'DVI': self.dvi, 'TVI': self.tvi, 'NDRE': self.ndre, } try: return funcs[index]() except KeyError: print('Index not in the list!' ) return False def a ( self ): '''simple docstring''' return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def a ( self ): '''simple docstring''' return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def a ( self ): '''simple docstring''' return self.nir * (self.red / (self.green**2)) def a ( self ): '''simple docstring''' return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def a ( self ): '''simple docstring''' return (self.nir - self.red) / (self.nir + self.red) def a ( self ): '''simple docstring''' return (self.nir - self.blue) / (self.nir + self.blue) def a ( self ): '''simple docstring''' return (self.redEdge - self.red) / (self.redEdge + self.red) def a ( self ): '''simple docstring''' return (self.nir - self.green) / (self.nir + self.green) def a ( self ): '''simple docstring''' return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def a ( self ): '''simple docstring''' return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def a ( self ): '''simple docstring''' return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def a ( self ): '''simple docstring''' return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def a ( self , snake_case__=0.08 , snake_case__=1.22 , snake_case__=0.03 ): '''simple docstring''' return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def a ( self ): '''simple docstring''' return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def a ( self ): '''simple docstring''' return (self.nir / self.green) - 1 def a ( self ): '''simple docstring''' return (self.nir / self.redEdge) - 1 def a ( self ): '''simple docstring''' return (self.red - self.blue) / self.red def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def a ( self ): '''simple docstring''' return self.nir - self.green def a ( self ): '''simple docstring''' return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.125) / (1 - self.red) def a ( self , snake_case__=0.16 ): '''simple docstring''' return (self.nir - self.green) / (self.nir + self.green + y) def a ( self , snake_case__=0.5 ): '''simple docstring''' return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def a ( self ): '''simple docstring''' return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def a ( self , snake_case__=None , snake_case__=None ): '''simple docstring''' return (self.nir - b) / (a * self.red) def a ( self ): '''simple docstring''' return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def a ( self ): '''simple docstring''' return (self.red + self.green + self.blue) / 30.5 def a ( self ): '''simple docstring''' return self.nir / self.red def a ( self ): '''simple docstring''' return (self.rvi() - 1) / (self.rvi() + 1) def a ( self ): '''simple docstring''' return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def a ( self ): '''simple docstring''' return self.green / (self.nir + self.red + self.green) def a ( self ): '''simple docstring''' return self.nir / (self.nir + self.red + self.green) def a ( self ): '''simple docstring''' return self.red / (self.nir + self.red + self.green) def a ( self ): '''simple docstring''' return (self.green - self.red) / (self.green + self.red) def a ( self ): '''simple docstring''' return (self.red - self.green) / (self.red + self.green) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) _lowerCAmelCase : Any = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def a ( self ): '''simple docstring''' return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def a ( self ): '''simple docstring''' return self.nir / self.red def a ( self ): '''simple docstring''' return (self.ndvi() + 0.5) ** (1 / 2) def a ( self ): '''simple docstring''' return (self.nir - self.redEdge) / (self.nir + self.redEdge)
721
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase : str = logging.get_logger(__name__) lowerCAmelCase : int = { """facebook/data2vec-text-base""": """https://huggingface.co/data2vec/resolve/main/config.json""", } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "data2vec-text" def __init__( self , snake_case__=3_0522 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=2 , snake_case__=0.02 , snake_case__=1E-12 , snake_case__=1 , snake_case__=0 , snake_case__=2 , snake_case__="absolute" , snake_case__=True , snake_case__=None , **snake_case__ , ): '''simple docstring''' super().__init__(pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ , **snake_case__ ) _lowerCAmelCase : List[Any] = vocab_size _lowerCAmelCase : Tuple = hidden_size _lowerCAmelCase : Dict = num_hidden_layers _lowerCAmelCase : int = num_attention_heads _lowerCAmelCase : str = hidden_act _lowerCAmelCase : Any = intermediate_size _lowerCAmelCase : Any = hidden_dropout_prob _lowerCAmelCase : Optional[int] = attention_probs_dropout_prob _lowerCAmelCase : str = max_position_embeddings _lowerCAmelCase : Any = type_vocab_size _lowerCAmelCase : int = initializer_range _lowerCAmelCase : List[str] = layer_norm_eps _lowerCAmelCase : List[Any] = position_embedding_type _lowerCAmelCase : str = use_cache _lowerCAmelCase : Union[str, Any] = classifier_dropout class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" @property def a ( self ): '''simple docstring''' if self.task == "multiple-choice": _lowerCAmelCase : Union[str, Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _lowerCAmelCase : List[Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
630
0
'''simple docstring''' from typing import TYPE_CHECKING from ..utils import _LazyModule lowerCAmelCase : Optional[int] = { """config""": [ """EXTERNAL_DATA_FORMAT_SIZE_LIMIT""", """OnnxConfig""", """OnnxConfigWithPast""", """OnnxSeq2SeqConfigWithPast""", """PatchingSpec""", ], """convert""": ["""export""", """validate_model_outputs"""], """features""": ["""FeaturesManager"""], """utils""": ["""ParameterFormat""", """compute_serialized_parameters_size"""], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys lowerCAmelCase : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
700
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins lowerCAmelCase : List[str] = ["""tests.fixtures.files""", """tests.fixtures.hub""", """tests.fixtures.fsspec"""] def lowercase (_A , _A ): """simple docstring""" for item in items: if any(marker in item.keywords for marker in ['integration', 'unit'] ): continue item.add_marker(pytest.mark.unit ) def lowercase (_A ): """simple docstring""" config.addinivalue_line('markers' , 'torchaudio_latest: mark test to run with torchaudio>=0.12' ) @pytest.fixture(autouse=_A ) def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : str = tmp_path_factory.getbasetemp() / 'cache' _lowerCAmelCase : Dict = test_hf_cache_home / 'datasets' _lowerCAmelCase : List[Any] = test_hf_cache_home / 'metrics' _lowerCAmelCase : List[Any] = test_hf_cache_home / 'modules' monkeypatch.setattr('datasets.config.HF_DATASETS_CACHE' , str(_A ) ) monkeypatch.setattr('datasets.config.HF_METRICS_CACHE' , str(_A ) ) monkeypatch.setattr('datasets.config.HF_MODULES_CACHE' , str(_A ) ) _lowerCAmelCase : Dict = test_hf_datasets_cache / 'downloads' monkeypatch.setattr('datasets.config.DOWNLOADED_DATASETS_PATH' , str(_A ) ) _lowerCAmelCase : Union[str, Any] = test_hf_datasets_cache / 'downloads' / 'extracted' monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH' , str(_A ) ) @pytest.fixture(autouse=_A , scope='session' ) def lowercase (): """simple docstring""" datasets.disable_progress_bar() @pytest.fixture(autouse=_A ) def lowercase (_A ): """simple docstring""" monkeypatch.setattr('datasets.config.HF_UPDATE_DOWNLOAD_COUNTS' , _A ) @pytest.fixture def lowercase (_A ): """simple docstring""" monkeypatch.setattr('sqlalchemy.util.deprecations.SILENCE_UBER_WARNING' , _A )
630
0
from cva import destroyAllWindows, imread, imshow, waitKey def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Dict = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(_A ): for j in range(_A ): _lowerCAmelCase : str = [2_5_5, 2_5_5, 2_5_5] - img[i][j] return img if __name__ == "__main__": # read original image lowerCAmelCase : str = imread("""image_data/lena.jpg""", 1) # convert to its negative lowerCAmelCase : Any = convert_to_negative(img) # show result image imshow("""negative of original image""", img) waitKey(0) destroyAllWindows()
701
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCAmelCase : str = logging.get_logger(__name__) # General docstring lowerCAmelCase : Optional[Any] = """RegNetConfig""" # Base docstring lowerCAmelCase : int = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = [1, 10_88, 7, 7] # Image classification docstring lowerCAmelCase : Any = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = """tabby, tabby cat""" lowerCAmelCase : Tuple = [ """facebook/regnet-y-040""", # See all regnet models at https://huggingface.co/models?filter=regnet ] class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 3 , snake_case__ = 1 , snake_case__ = 1 , snake_case__ = "relu" , **snake_case__ , ): '''simple docstring''' super().__init__(**snake_case__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _lowerCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _lowerCAmelCase : List[Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=snake_case__ , strides=snake_case__ , padding='VALID' , groups=snake_case__ , use_bias=snake_case__ , name='convolution' , ) _lowerCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) _lowerCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.convolution(self.padding(snake_case__ ) ) _lowerCAmelCase : Union[str, Any] = self.normalization(snake_case__ ) _lowerCAmelCase : int = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = config.num_channels _lowerCAmelCase : List[Any] = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name='embedder' , ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = shape_list(snake_case__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( 'Make sure that the channel dimension of the pixel values match with the one set in the configuration.' ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _lowerCAmelCase : List[Any] = tf.transpose(snake_case__ , perm=(0, 2, 3, 1) ) _lowerCAmelCase : Tuple = self.embedder(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=1 , strides=snake_case__ , use_bias=snake_case__ , name='convolution' ) _lowerCAmelCase : Tuple = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' return self.normalization(self.convolution(snake_case__ ) , training=snake_case__ ) class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Any = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) _lowerCAmelCase : str = [ tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='relu' , name='attention.0' ), tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='sigmoid' , name='attention.2' ), ] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.pooler(snake_case__ ) for layer_module in self.attention: _lowerCAmelCase : Tuple = layer_module(snake_case__ ) _lowerCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Optional[int] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _lowerCAmelCase : Any = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.2' ), ] _lowerCAmelCase : List[str] = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = hidden_state for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) _lowerCAmelCase : int = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : Tuple = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : List[str] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) _lowerCAmelCase : Tuple = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetSELayer(snake_case__ , reduced_channels=int(round(in_channels / 4 ) ) , name='layer.2' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.3' ), ] _lowerCAmelCase : Tuple = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = hidden_state for layer_module in self.layers: _lowerCAmelCase : List[Any] = layer_module(snake_case__ ) _lowerCAmelCase : Tuple = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : str = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 2 , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Dict = TFRegNetXLayer if config.layer_type == 'x' else TFRegNetYLayer _lowerCAmelCase : Optional[int] = [ # downsampling is done in the first layer with stride of 2 layer(snake_case__ , snake_case__ , snake_case__ , stride=snake_case__ , name='layers.0' ), *[layer(snake_case__ , snake_case__ , snake_case__ , name=F'layers.{i+1}' ) for i in range(depth - 1 )], ] def a ( self , snake_case__ ): '''simple docstring''' for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( snake_case__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name='stages.0' , ) ) _lowerCAmelCase : Union[str, Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(snake_case__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(snake_case__ , snake_case__ , snake_case__ , depth=snake_case__ , name=F'stages.{i+1}' ) ) def a ( self , snake_case__ , snake_case__ = False , snake_case__ = True ): '''simple docstring''' _lowerCAmelCase : List[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _lowerCAmelCase : str = hidden_states + (hidden_state,) _lowerCAmelCase : List[str] = stage_module(snake_case__ ) if output_hidden_states: _lowerCAmelCase : Dict = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=snake_case__ , hidden_states=snake_case__ ) @keras_serializable class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" __magic_name__ = RegNetConfig def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = config _lowerCAmelCase : Union[str, Any] = TFRegNetEmbeddings(snake_case__ , name='embedder' ) _lowerCAmelCase : Optional[int] = TFRegNetEncoder(snake_case__ , name='encoder' ) _lowerCAmelCase : Dict = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) @unpack_inputs def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' _lowerCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : int = self.embedder(snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[str] = self.encoder( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[Any] = encoder_outputs[0] _lowerCAmelCase : Tuple = self.pooler(snake_case__ ) # Change to NCHW output format have uniformity in the modules _lowerCAmelCase : Optional[int] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) _lowerCAmelCase : Optional[Any] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _lowerCAmelCase : Union[str, Any] = tuple([tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=snake_case__ , pooler_output=snake_case__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = RegNetConfig __magic_name__ = "regnet" __magic_name__ = "pixel_values" @property def a ( self ): '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 224, 224) , dtype=tf.floataa )} lowerCAmelCase : List[Any] = r""" Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. """ lowerCAmelCase : Dict = r""" Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : List[str] = TFRegNetMainLayer(snake_case__ , name='regnet' ) @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Any = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : str = self.regnet( pixel_values=snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[Any] = config.num_labels _lowerCAmelCase : Optional[Any] = TFRegNetMainLayer(snake_case__ , name='regnet' ) # classification head _lowerCAmelCase : Optional[int] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name='classifier.1' ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def a ( self , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : Dict = self.regnet( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.pooler_output if return_dict else outputs[1] _lowerCAmelCase : List[Any] = self.classifier[0](snake_case__ ) _lowerCAmelCase : Tuple = self.classifier[1](snake_case__ ) _lowerCAmelCase : int = None if labels is None else self.hf_compute_loss(labels=snake_case__ , logits=snake_case__ ) if not return_dict: _lowerCAmelCase : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=snake_case__ , logits=snake_case__ , hidden_states=outputs.hidden_states )
630
0
'''simple docstring''' import numpy as np lowerCAmelCase : Optional[Any] = [ ["""a""", """b""", """c""", """d""", """e"""], ["""f""", """g""", """h""", """i""", """k"""], ["""l""", """m""", """n""", """o""", """p"""], ["""q""", """r""", """s""", """t""", """u"""], ["""v""", """w""", """x""", """y""", """z"""], ] class UpperCamelCase__ : """simple docstring""" def __init__( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = np.array(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = np.where(letter == self.SQUARE ) _lowerCAmelCase : Dict = np.concatenate([indexa + 1, indexa + 1] ) return indexes def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Tuple = self.SQUARE[indexa - 1, indexa - 1] return letter def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = message.lower() _lowerCAmelCase : str = message.replace(' ' , '' ) _lowerCAmelCase : List[str] = message.replace('j' , 'i' ) _lowerCAmelCase : List[Any] = np.empty((2, len(snake_case__ )) ) for letter_index in range(len(snake_case__ ) ): _lowerCAmelCase : Optional[Any] = self.letter_to_numbers(message[letter_index] ) _lowerCAmelCase : int = numbers[0] _lowerCAmelCase : str = numbers[1] _lowerCAmelCase : List[Any] = first_step.reshape(2 * len(snake_case__ ) ) _lowerCAmelCase : Tuple = '' for numbers_index in range(len(snake_case__ ) ): _lowerCAmelCase : List[Any] = int(second_step[numbers_index * 2] ) _lowerCAmelCase : Optional[Any] = int(second_step[(numbers_index * 2) + 1] ) _lowerCAmelCase : List[str] = self.numbers_to_letter(snake_case__ , snake_case__ ) _lowerCAmelCase : Optional[int] = encoded_message + letter return encoded_message def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Tuple = message.lower() message.replace(' ' , '' ) _lowerCAmelCase : int = np.empty(2 * len(snake_case__ ) ) for letter_index in range(len(snake_case__ ) ): _lowerCAmelCase : Optional[Any] = self.letter_to_numbers(message[letter_index] ) _lowerCAmelCase : Tuple = numbers[0] _lowerCAmelCase : str = numbers[1] _lowerCAmelCase : Optional[Any] = first_step.reshape((2, len(snake_case__ )) ) _lowerCAmelCase : Optional[int] = '' for numbers_index in range(len(snake_case__ ) ): _lowerCAmelCase : Tuple = int(second_step[0, numbers_index] ) _lowerCAmelCase : str = int(second_step[1, numbers_index] ) _lowerCAmelCase : str = self.numbers_to_letter(snake_case__ , snake_case__ ) _lowerCAmelCase : int = decoded_message + letter return decoded_message
702
'''simple docstring''' from typing import Any def lowercase (_A ): """simple docstring""" if not input_list: return [] _lowerCAmelCase : Optional[int] = [input_list.count(_A ) for value in input_list] _lowerCAmelCase : int = max(_A ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(_A ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
630
0
'''simple docstring''' import argparse from ...utils.dataclasses import ( ComputeEnvironment, DistributedType, DynamoBackend, PrecisionType, SageMakerDistributedType, ) from ..menu import BulletMenu lowerCAmelCase : int = [ """EAGER""", """AOT_EAGER""", """INDUCTOR""", """NVFUSER""", """AOT_NVFUSER""", """AOT_CUDAGRAPHS""", """OFI""", """FX2TRT""", """ONNXRT""", """IPEX""", ] def lowercase (_A , _A=None , _A=None , _A=None ): """simple docstring""" _lowerCAmelCase : Optional[Any] = True while ask_again: _lowerCAmelCase : Dict = input(_A ) try: if default is not None and len(_A ) == 0: return default return convert_value(_A ) if convert_value is not None else result except Exception: if error_message is not None: print(_A ) def lowercase (_A , _A=[] , _A=None , _A=0 ): """simple docstring""" _lowerCAmelCase : Dict = BulletMenu(_A , _A ) _lowerCAmelCase : Optional[int] = menu.run(default_choice=_A ) return convert_value(_A ) if convert_value is not None else result def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Optional[int] = int(_A ) return ComputeEnvironment(['LOCAL_MACHINE', 'AMAZON_SAGEMAKER'][value] ) def lowercase (_A ): """simple docstring""" _lowerCAmelCase : str = int(_A ) return DistributedType(['NO', 'MULTI_CPU', 'MULTI_XPU', 'MULTI_GPU', 'MULTI_NPU', 'TPU'][value] ) def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Optional[int] = int(_A ) return DynamoBackend(DYNAMO_BACKENDS[value] ).value def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Optional[int] = int(_A ) return PrecisionType(['no', 'fp16', 'bf16', 'fp8'][value] ) def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Union[str, Any] = int(_A ) return SageMakerDistributedType(['NO', 'DATA_PARALLEL', 'MODEL_PARALLEL'][value] ) def lowercase (_A ): """simple docstring""" return {"yes": True, "no": False}[value.lower()] class UpperCamelCase__ ( argparse.RawDescriptionHelpFormatter ): """simple docstring""" def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[str] = super()._format_usage(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) _lowerCAmelCase : int = usage.replace('<command> [<args>] ' , '' ) return usage
703
'''simple docstring''' from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
630
0
'''simple docstring''' import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=99 , snake_case__=32 , snake_case__=5 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=16 , snake_case__=2 , snake_case__=0.02 , snake_case__=4 , ): '''simple docstring''' _lowerCAmelCase : List[Any] = parent _lowerCAmelCase : Union[str, Any] = batch_size _lowerCAmelCase : Union[str, Any] = seq_length _lowerCAmelCase : Any = is_training _lowerCAmelCase : List[Any] = use_attention_mask _lowerCAmelCase : Any = use_token_type_ids _lowerCAmelCase : Optional[Any] = use_labels _lowerCAmelCase : Tuple = vocab_size _lowerCAmelCase : Union[str, Any] = hidden_size _lowerCAmelCase : int = num_hidden_layers _lowerCAmelCase : List[Any] = num_attention_heads _lowerCAmelCase : Union[str, Any] = intermediate_size _lowerCAmelCase : List[Any] = hidden_act _lowerCAmelCase : List[Any] = hidden_dropout_prob _lowerCAmelCase : int = attention_probs_dropout_prob _lowerCAmelCase : int = max_position_embeddings _lowerCAmelCase : Optional[int] = type_vocab_size _lowerCAmelCase : Dict = type_sequence_label_size _lowerCAmelCase : List[str] = initializer_range _lowerCAmelCase : Optional[Any] = num_choices def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase : Dict = None if self.use_attention_mask: _lowerCAmelCase : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCAmelCase : Any = None if self.use_token_type_ids: _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCAmelCase : Tuple = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.prepare_config_and_inputs() _lowerCAmelCase : Dict = config_and_inputs _lowerCAmelCase : Optional[Any] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.prepare_config_and_inputs() _lowerCAmelCase : Any = config_and_inputs _lowerCAmelCase : Optional[Any] = True _lowerCAmelCase : Dict = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) _lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = True __magic_name__ = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def a ( self ): '''simple docstring''' _lowerCAmelCase : str = FlaxBertModelTester(self ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = FlaxBertModel.from_pretrained('bert-base-cased' ) _lowerCAmelCase : List[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(snake_case__ )
704
'''simple docstring''' import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase : int = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = {"""vocab_file""": """spiece.model"""} lowerCAmelCase : Optional[int] = { """vocab_file""": { """AI-Sweden/gpt-sw3-126m""": """https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-350m""": """https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-1.6b""": """https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-6.7b""": """https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-20b""": """https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model""", } } lowerCAmelCase : Union[str, Any] = { """AI-Sweden/gpt-sw3-126m""": 20_48, """AI-Sweden/gpt-sw3-350m""": 20_48, """AI-Sweden/gpt-sw3-1.6b""": 20_48, """AI-Sweden/gpt-sw3-6.7b""": 20_48, """AI-Sweden/gpt-sw3-20b""": 20_48, } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = ["input_ids", "attention_mask"] def __init__( self , snake_case__ , snake_case__=False , snake_case__=False , snake_case__=False , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Tuple = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase : List[Any] = kwargs.get('name_or_path' ) if name_or_path is None: logger.warning( 'name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,' ' you are testing the model, this can safely be ignored' ) _lowerCAmelCase : Any = 'None' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing _lowerCAmelCase : str = '<|endoftext|>' if eos_token is None else eos_token _lowerCAmelCase : Tuple = '<unk>' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: _lowerCAmelCase : List[str] = unk_token if pad_token is None else pad_token _lowerCAmelCase : Optional[int] = eos_token if bos_token is None else bos_token else: _lowerCAmelCase : Tuple = '<pad>' if pad_token is None else pad_token _lowerCAmelCase : Union[str, Any] = '<s>' if bos_token is None else bos_token super().__init__( do_lower_case=snake_case__ , remove_space=snake_case__ , keep_accents=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , pad_token=snake_case__ , sp_model_kwargs=self.sp_model_kwargs , **snake_case__ , ) _lowerCAmelCase : Union[str, Any] = do_lower_case _lowerCAmelCase : Optional[int] = remove_space _lowerCAmelCase : Any = keep_accents _lowerCAmelCase : Optional[int] = vocab_file _lowerCAmelCase : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case__ ) # Used for whitespace normalization in input texts # fmt : off _lowerCAmelCase : Optional[Any] = {' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', '', '„'} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing _lowerCAmelCase : Optional[Any] = re.compile( F'[{"".join(map(snake_case__ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]' ) def __getstate__( self ): '''simple docstring''' _lowerCAmelCase : List[str] = self.__dict__.copy() _lowerCAmelCase : int = None return state def __setstate__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowerCAmelCase : int = {} _lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def a ( self ): '''simple docstring''' return len(self.sp_model ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.non_printing_characters_re.sub('' , snake_case__ ) # Normalize whitespaces _lowerCAmelCase : Tuple = ''.join([char if char not in self.whitespaces else ' ' for char in text] ) # NFC Unicode normalization _lowerCAmelCase : Union[str, Any] = unicodedata.normalize('NFC' , snake_case__ ) return text def a ( self , snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : str = self.preprocess_text(snake_case__ ) return self.sp_model.encode(snake_case__ , out_type=snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.PieceToId(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case__ ) @staticmethod def a ( snake_case__ ): '''simple docstring''' return out_string def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = [] _lowerCAmelCase : Optional[Any] = '' _lowerCAmelCase : Tuple = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(snake_case__ ) + token _lowerCAmelCase : Union[str, Any] = True _lowerCAmelCase : List[Any] = [] else: current_sub_tokens.append(snake_case__ ) _lowerCAmelCase : List[Any] = False out_string += self.sp_model.decode(snake_case__ ) return out_string def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = {self.convert_ids_to_tokens(snake_case__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCAmelCase : int = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case__ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case__ , 'wb' ) as fi: _lowerCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (out_vocab_file,) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' if isinstance(snake_case__ , snake_case__ ): _lowerCAmelCase : Optional[Any] = self.preprocess_text(snake_case__ ) _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) else: _lowerCAmelCase : Tuple = [self.preprocess_text(snake_case__ ) for t in text] _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) if return_tensors is True or return_tensors == "pt": _lowerCAmelCase : int = torch.tensor(snake_case__ ) return token_ids def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.decode(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = [F'User: {text}' if is_user else F'Bot: {text}' for is_user, text in conversation.iter_texts()] _lowerCAmelCase : str = ( F'{self.eos_token}{self.bos_token}' + F'{self.bos_token}'.join(snake_case__ ) + F'{self.bos_token}Bot:' ) return self.encode(text=snake_case__ )
630
0
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = ["image_processor", "tokenizer"] __magic_name__ = "LayoutLMv3ImageProcessor" __magic_name__ = ("LayoutLMv3Tokenizer", "LayoutLMv3TokenizerFast") def __init__( self , snake_case__=None , snake_case__=None , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case__ , ) _lowerCAmelCase : int = kwargs.pop('feature_extractor' ) _lowerCAmelCase : Tuple = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case__ , snake_case__ ) def __call__( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = True , snake_case__ = False , snake_case__ = None , snake_case__ = None , snake_case__ = 0 , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = False , snake_case__ = False , snake_case__ = False , snake_case__ = False , snake_case__ = True , snake_case__ = None , **snake_case__ , ): '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.' ) # first, apply the image processor _lowerCAmelCase : Optional[Any] = self.image_processor(images=snake_case__ , return_tensors=snake_case__ ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(snake_case__ , snake_case__ ): _lowerCAmelCase : Union[str, Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) _lowerCAmelCase : List[str] = features['words'] _lowerCAmelCase : List[str] = self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=snake_case__ , add_special_tokens=snake_case__ , padding=snake_case__ , truncation=snake_case__ , max_length=snake_case__ , stride=snake_case__ , pad_to_multiple_of=snake_case__ , return_token_type_ids=snake_case__ , return_attention_mask=snake_case__ , return_overflowing_tokens=snake_case__ , return_special_tokens_mask=snake_case__ , return_offsets_mapping=snake_case__ , return_length=snake_case__ , verbose=snake_case__ , return_tensors=snake_case__ , **snake_case__ , ) # add pixel values _lowerCAmelCase : Dict = features.pop('pixel_values' ) if return_overflowing_tokens is True: _lowerCAmelCase : Any = self.get_overflowing_images(snake_case__ , encoded_inputs['overflow_to_sample_mapping'] ) _lowerCAmelCase : Any = images return encoded_inputs def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(snake_case__ ) != len(snake_case__ ): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' F' {len(snake_case__ )} and {len(snake_case__ )}' ) return images_with_overflow def a ( self , *snake_case__ , **snake_case__ ): '''simple docstring''' return self.tokenizer.batch_decode(*snake_case__ , **snake_case__ ) def a ( self , *snake_case__ , **snake_case__ ): '''simple docstring''' return self.tokenizer.decode(*snake_case__ , **snake_case__ ) @property def a ( self ): '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def a ( self ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case__ , ) return self.image_processor_class @property def a ( self ): '''simple docstring''' warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case__ , ) return self.image_processor
705
'''simple docstring''' import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = (DDPMScheduler,) def a ( self , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**snake_case__ ) return config def a ( self ): '''simple docstring''' for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=snake_case__ ) def a ( self ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=snake_case__ , beta_end=snake_case__ ) def a ( self ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=snake_case__ ) def a ( self ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=snake_case__ ) def a ( self ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=snake_case__ ) def a ( self ): '''simple docstring''' self.check_over_configs(thresholding=snake_case__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=snake_case__ , prediction_type=snake_case__ , sample_max_value=snake_case__ , ) def a ( self ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=snake_case__ ) def a ( self ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1E-5 def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : str = self.dummy_model() _lowerCAmelCase : str = self.dummy_sample_deter _lowerCAmelCase : Union[str, Any] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : List[Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Any = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Dict = pred_prev_sample _lowerCAmelCase : Dict = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : List[str] = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 258.9606 ) < 1E-2 assert abs(result_mean.item() - 0.3372 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.scheduler_classes[0] _lowerCAmelCase : Any = self.get_scheduler_config(prediction_type='v_prediction' ) _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : Any = self.dummy_model() _lowerCAmelCase : Tuple = self.dummy_sample_deter _lowerCAmelCase : Optional[int] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : Union[str, Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Dict = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Tuple = pred_prev_sample _lowerCAmelCase : Any = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : Dict = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 202.0296 ) < 1E-2 assert abs(result_mean.item() - 0.2631 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[int] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=snake_case__ ) _lowerCAmelCase : Union[str, Any] = scheduler.timesteps for i, timestep in enumerate(snake_case__ ): if i == len(snake_case__ ) - 1: _lowerCAmelCase : str = -1 else: _lowerCAmelCase : Optional[Any] = timesteps[i + 1] _lowerCAmelCase : int = scheduler.previous_timestep(snake_case__ ) _lowerCAmelCase : int = prev_t.item() self.assertEqual(snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : Tuple = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 51, 0] with self.assertRaises(snake_case__ , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : List[str] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 1, 0] _lowerCAmelCase : int = len(snake_case__ ) with self.assertRaises(snake_case__ , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=snake_case__ , timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : int = self.get_scheduler_config() _lowerCAmelCase : Any = scheduler_class(**snake_case__ ) _lowerCAmelCase : Any = [scheduler.config.num_train_timesteps] with self.assertRaises( snake_case__ , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=snake_case__ )
630
0
'''simple docstring''' import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=True , snake_case__=False , snake_case__=True , snake_case__=99 , snake_case__=32 , snake_case__=5 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=16 , snake_case__=2 , snake_case__=0.02 , snake_case__=3 , snake_case__=4 , snake_case__=None , ): '''simple docstring''' _lowerCAmelCase : Dict = parent _lowerCAmelCase : Tuple = batch_size _lowerCAmelCase : str = seq_length _lowerCAmelCase : Tuple = is_training _lowerCAmelCase : Optional[int] = use_input_mask _lowerCAmelCase : Any = use_token_type_ids _lowerCAmelCase : List[str] = use_labels _lowerCAmelCase : Optional[Any] = vocab_size _lowerCAmelCase : Optional[Any] = hidden_size _lowerCAmelCase : Optional[Any] = num_hidden_layers _lowerCAmelCase : Tuple = num_attention_heads _lowerCAmelCase : str = intermediate_size _lowerCAmelCase : Tuple = hidden_act _lowerCAmelCase : List[str] = hidden_dropout_prob _lowerCAmelCase : str = attention_probs_dropout_prob _lowerCAmelCase : Any = max_position_embeddings _lowerCAmelCase : Tuple = type_vocab_size _lowerCAmelCase : List[str] = type_sequence_label_size _lowerCAmelCase : Union[str, Any] = initializer_range _lowerCAmelCase : Union[str, Any] = num_labels _lowerCAmelCase : Optional[int] = num_choices _lowerCAmelCase : int = scope def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase : Optional[int] = None if self.use_input_mask: _lowerCAmelCase : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCAmelCase : Tuple = None if self.use_token_type_ids: _lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCAmelCase : Tuple = None _lowerCAmelCase : Union[str, Any] = None _lowerCAmelCase : List[Any] = None if self.use_labels: _lowerCAmelCase : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCAmelCase : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) _lowerCAmelCase : int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a ( self ): '''simple docstring''' return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = LlamaModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Any = model(snake_case__ , attention_mask=snake_case__ ) _lowerCAmelCase : int = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Tuple = True _lowerCAmelCase : Union[str, Any] = LlamaModel(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : List[Any] = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , ) _lowerCAmelCase : Dict = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , ) _lowerCAmelCase : Any = model(snake_case__ , attention_mask=snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = LlamaForCausalLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : str = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ): '''simple docstring''' _lowerCAmelCase : int = True _lowerCAmelCase : Union[str, Any] = True _lowerCAmelCase : Any = LlamaForCausalLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() # first forward pass _lowerCAmelCase : Dict = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , use_cache=snake_case__ , ) _lowerCAmelCase : Any = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _lowerCAmelCase : List[str] = ids_tensor((self.batch_size, 3) , config.vocab_size ) _lowerCAmelCase : int = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _lowerCAmelCase : List[str] = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCAmelCase : List[str] = torch.cat([input_mask, next_mask] , dim=-1 ) _lowerCAmelCase : Dict = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , output_hidden_states=snake_case__ , )['hidden_states'][0] _lowerCAmelCase : List[Any] = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , past_key_values=snake_case__ , output_hidden_states=snake_case__ , )['hidden_states'][0] # select random slice _lowerCAmelCase : Optional[int] = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCAmelCase : List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() _lowerCAmelCase : Optional[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1E-3 ) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.prepare_config_and_inputs() ( _lowerCAmelCase ) : str = config_and_inputs _lowerCAmelCase : List[str] = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () __magic_name__ = (LlamaForCausalLM,) if is_torch_available() else () __magic_name__ = ( { "feature-extraction": LlamaModel, "text-classification": LlamaForSequenceClassification, "text-generation": LlamaForCausalLM, "zero-shot": LlamaForSequenceClassification, } if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = LlamaModelTester(self ) _lowerCAmelCase : str = ConfigTester(self , config_class=snake_case__ , hidden_size=37 ) def a ( self ): '''simple docstring''' self.config_tester.run_common_tests() def a ( self ): '''simple docstring''' _lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowerCAmelCase : int = type self.model_tester.create_and_check_model(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase : Union[str, Any] = 3 _lowerCAmelCase : Optional[Any] = input_dict['input_ids'] _lowerCAmelCase : str = input_ids.ne(1 ).to(snake_case__ ) _lowerCAmelCase : List[str] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _lowerCAmelCase : Union[str, Any] = LlamaForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : str = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase : Tuple = 3 _lowerCAmelCase : Union[str, Any] = 'single_label_classification' _lowerCAmelCase : List[str] = input_dict['input_ids'] _lowerCAmelCase : Tuple = input_ids.ne(1 ).to(snake_case__ ) _lowerCAmelCase : int = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _lowerCAmelCase : List[Any] = LlamaForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Union[str, Any] = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase : Tuple = 3 _lowerCAmelCase : Optional[int] = 'multi_label_classification' _lowerCAmelCase : List[Any] = input_dict['input_ids'] _lowerCAmelCase : Optional[Any] = input_ids.ne(1 ).to(snake_case__ ) _lowerCAmelCase : Tuple = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) _lowerCAmelCase : Any = LlamaForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Optional[int] = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('LLaMA buffers include complex numbers, which breaks this test' ) def a ( self ): '''simple docstring''' pass @parameterized.expand([('linear',), ('dynamic',)] ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase : Union[str, Any] = ids_tensor([1, 10] , config.vocab_size ) _lowerCAmelCase : str = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _lowerCAmelCase : List[Any] = LlamaModel(snake_case__ ) original_model.to(snake_case__ ) original_model.eval() _lowerCAmelCase : List[str] = original_model(snake_case__ ).last_hidden_state _lowerCAmelCase : List[Any] = original_model(snake_case__ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _lowerCAmelCase : Tuple = {'type': scaling_type, 'factor': 10.0} _lowerCAmelCase : Optional[Any] = LlamaModel(snake_case__ ) scaled_model.to(snake_case__ ) scaled_model.eval() _lowerCAmelCase : Any = scaled_model(snake_case__ ).last_hidden_state _lowerCAmelCase : Dict = scaled_model(snake_case__ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(snake_case__ , snake_case__ , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case__ , snake_case__ , atol=1E-5 ) ) @require_torch class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = [1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCAmelCase : int = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-7b-hf' , device_map='auto' ) _lowerCAmelCase : Union[str, Any] = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 _lowerCAmelCase : Any = torch.tensor([[-6.6550, -4.1227, -4.9859, -3.2406, 0.8262, -3.0033, 1.2964, -3.3699]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off _lowerCAmelCase : Any = torch.tensor([-12.8281, -7.4453, -0.4639, -8.0625, -7.2500, -8.0000, -6.4883, -7.7695, -7.8438, -7.0312, -6.2188, -7.1328, -1.8496, 1.9961, -8.6250, -6.7227, -12.8281, -6.9492, -7.0742, -7.7852, -7.5820, -7.9062, -6.9375, -7.9805, -8.3438, -8.1562, -8.0469, -7.6250, -7.7422, -7.3398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , snake_case__ , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = [1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCAmelCase : Dict = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-hf' , device_map='auto' ) _lowerCAmelCase : str = model(torch.tensor(snake_case__ ) ) # Expected mean on dim = -1 _lowerCAmelCase : Tuple = torch.tensor([[-2.0622, -1.2794, -1.1638, -0.9788, -1.4603, -1.0238, -1.7893, -1.4411]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off _lowerCAmelCase : Dict = torch.tensor([-8.1406, -8.0547, 2.7461, -1.2344, -0.1448, -1.8262, -1.0020, -1.8154, -1.6895, -1.8516, -2.3574, -0.9277, 3.7598, 6.5742, -1.2998, -0.1177, -8.1406, -2.9688, -2.9199, -3.1699, -3.5254, -2.3555, -2.7988, -3.4141, -2.8262, -4.5195, -3.3379, -3.3164, -2.7832, -3.0273] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , snake_case__ , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCAmelCase : Dict = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-chat-hf' , device_map='auto' ) _lowerCAmelCase : Union[str, Any] = model(torch.tensor(snake_case__ ) ) # Expected mean on dim = -1 _lowerCAmelCase : Any = torch.tensor([[-0.8562, -1.8520, -0.7551, -0.4162, -1.5161, -1.2038, -2.4823, -2.3254]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off _lowerCAmelCase : Union[str, Any] = torch.tensor([-2.2227, 4.8828, 0.9023, -0.4578, -0.7871, -0.1033, -0.6221, -0.5786, -0.7803, -1.0674, -1.2920, -0.1570, 0.8008, 2.0723, -0.9497, 0.2771, -2.2227, -0.7612, -1.4346, -1.2061, -1.6426, -0.3000, -0.7139, -1.1934, -1.8691, -1.6973, -1.5947, -1.2705, -0.3523, -0.5513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1E-2 , rtol=1E-2 ) @unittest.skip( 'Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test' ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [1, 306, 4658, 278, 6593, 310, 2834, 338] _lowerCAmelCase : List[Any] = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-70b-hf' , device_map='auto' ) _lowerCAmelCase : List[Any] = model(torch.tensor(snake_case__ ) ) _lowerCAmelCase : List[Any] = torch.tensor( [[-4.2327, -3.3360, -4.6665, -4.7631, -1.8180, -3.4170, -1.4211, -3.1810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1E-2 , rtol=1E-2 ) # fmt: off _lowerCAmelCase : List[str] = torch.tensor([-9.4922, -3.9551, 1.7998, -5.6758, -5.1055, -5.8984, -4.8320, -6.8086, -6.5391, -5.6172, -5.5820, -5.5352, 1.7881, 3.6289, -6.5117, -3.4785, -9.5000, -6.0352, -6.8125, -6.0195, -6.6836, -5.4727, -6.2812, -6.0391, -7.3398, -7.4297, -7.4844, -6.5820, -5.8789, -5.5312] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , snake_case__ , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Model is curently gated' ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = 'Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi' _lowerCAmelCase : Dict = 'Simply put, the theory of relativity states that ' _lowerCAmelCase : List[str] = LlamaTokenizer.from_pretrained('meta-llama/Llama-2-13b-chat-hf' ) _lowerCAmelCase : List[Any] = tokenizer.encode(snake_case__ , return_tensors='pt' ) _lowerCAmelCase : str = LlamaForCausalLM.from_pretrained( 'meta-llama/Llama-2-13b-chat-hf' , device_map='sequential' , use_safetensors=snake_case__ ) # greedy generation outputs _lowerCAmelCase : List[Any] = model.generate(snake_case__ , max_new_tokens=64 , top_p=snake_case__ , temperature=1 , do_sample=snake_case__ ) _lowerCAmelCase : Tuple = tokenizer.decode(generated_ids[0] , skip_special_tokens=snake_case__ ) self.assertEqual(snake_case__ , snake_case__ )
706
'''simple docstring''' import socket def lowercase (): """simple docstring""" _lowerCAmelCase : Tuple = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCAmelCase : Optional[int] = socket.gethostname() _lowerCAmelCase : Any = 1_2_3_1_2 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCAmelCase : Union[str, Any] = sock.recv(1_0_2_4 ) if not data: break out_file.write(_A ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
630
0
'''simple docstring''' import itertools import json import os import unittest from transformers import AddedToken, RobertaTokenizer, RobertaTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = RobertaTokenizer __magic_name__ = RobertaTokenizerFast __magic_name__ = True __magic_name__ = {"cls_token": "<s>"} def a ( self ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _lowerCAmelCase : Any = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] _lowerCAmelCase : Optional[Any] = dict(zip(snake_case__ , range(len(snake_case__ ) ) ) ) _lowerCAmelCase : Any = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] _lowerCAmelCase : Optional[Any] = {'unk_token': '<unk>'} _lowerCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) _lowerCAmelCase : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(snake_case__ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(snake_case__ ) ) def a ( self , **snake_case__ ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **snake_case__ ) def a ( self , **snake_case__ ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return RobertaTokenizerFast.from_pretrained(self.tmpdirname , **snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = 'lower newer' _lowerCAmelCase : Dict = 'lower newer' return input_text, output_text def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) _lowerCAmelCase : Optional[Any] = 'lower newer' _lowerCAmelCase : Optional[int] = ['l', 'o', 'w', 'er', '\u0120', 'n', 'e', 'w', 'er'] _lowerCAmelCase : List[str] = tokenizer.tokenize(snake_case__ ) # , add_prefix_space=True) self.assertListEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : Tuple = tokens + [tokenizer.unk_token] _lowerCAmelCase : Optional[int] = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case__ ) , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : str = self.get_tokenizer() self.assertListEqual(tokenizer.encode('Hello world!' , add_special_tokens=snake_case__ ) , [0, 3_1414, 232, 328, 2] ) self.assertListEqual( tokenizer.encode('Hello world! cécé herlolip 418' , add_special_tokens=snake_case__ ) , [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2] , ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.tokenizer_class.from_pretrained('roberta-base' ) _lowerCAmelCase : Any = tokenizer.encode('sequence builders' , add_special_tokens=snake_case__ ) _lowerCAmelCase : Any = tokenizer.encode('multi-sequence build' , add_special_tokens=snake_case__ ) _lowerCAmelCase : List[Any] = tokenizer.encode( 'sequence builders' , add_special_tokens=snake_case__ , add_prefix_space=snake_case__ ) _lowerCAmelCase : Dict = tokenizer.encode( 'sequence builders' , 'multi-sequence build' , add_special_tokens=snake_case__ , add_prefix_space=snake_case__ ) _lowerCAmelCase : Dict = tokenizer.build_inputs_with_special_tokens(snake_case__ ) _lowerCAmelCase : str = tokenizer.build_inputs_with_special_tokens(snake_case__ , snake_case__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.get_tokenizer() _lowerCAmelCase : int = 'Encode this sequence.' _lowerCAmelCase : List[str] = tokenizer.byte_encoder[' '.encode('utf-8' )[0]] # Testing encoder arguments _lowerCAmelCase : Tuple = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ , add_prefix_space=snake_case__ ) _lowerCAmelCase : str = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : List[Any] = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ , add_prefix_space=snake_case__ ) _lowerCAmelCase : List[Any] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(snake_case__ , snake_case__ ) tokenizer.add_special_tokens({'bos_token': '<s>'} ) _lowerCAmelCase : Dict = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) _lowerCAmelCase : List[str] = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(snake_case__ , snake_case__ ) # Testing spaces after special tokens _lowerCAmelCase : Tuple = '<mask>' tokenizer.add_special_tokens( {'mask_token': AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ )} ) # mask token has a left space _lowerCAmelCase : List[str] = tokenizer.convert_tokens_to_ids(snake_case__ ) _lowerCAmelCase : Tuple = 'Encode <mask> sequence' _lowerCAmelCase : Dict = 'Encode <mask>sequence' _lowerCAmelCase : List[str] = tokenizer.encode(snake_case__ ) _lowerCAmelCase : Any = encoded.index(snake_case__ ) _lowerCAmelCase : Any = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : List[str] = tokenizer.encode(snake_case__ ) _lowerCAmelCase : int = encoded.index(snake_case__ ) _lowerCAmelCase : Any = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): _lowerCAmelCase : Optional[int] = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) _lowerCAmelCase : int = self.tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) _lowerCAmelCase : List[Any] = 'A, <mask> AllenNLP sentence.' _lowerCAmelCase : str = tokenizer_r.encode_plus(snake_case__ , add_special_tokens=snake_case__ , return_token_type_ids=snake_case__ ) _lowerCAmelCase : List[Any] = tokenizer_p.encode_plus(snake_case__ , add_special_tokens=snake_case__ , return_token_type_ids=snake_case__ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['token_type_ids'] ) , sum(tokens_p['token_type_ids'] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['attention_mask'] ) / len(tokens_r['attention_mask'] ) , sum(tokens_p['attention_mask'] ) / len(tokens_p['attention_mask'] ) , ) _lowerCAmelCase : Optional[Any] = tokenizer_r.convert_ids_to_tokens(tokens_r['input_ids'] ) _lowerCAmelCase : str = tokenizer_p.convert_ids_to_tokens(tokens_p['input_ids'] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['input_ids'] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r['input_ids'] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual( snake_case__ , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] ) self.assertSequenceEqual( snake_case__ , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] ) def a ( self ): '''simple docstring''' for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): _lowerCAmelCase : Any = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ ) _lowerCAmelCase : str = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) _lowerCAmelCase : Optional[int] = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state['add_prefix_space'] , snake_case__ ) self.assertEqual(post_processor_state['add_prefix_space'] , snake_case__ ) self.assertEqual(post_processor_state['trim_offsets'] , snake_case__ ) def a ( self ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): _lowerCAmelCase : Any = 'hello' # `hello` is a token in the vocabulary of `pretrained_name` _lowerCAmelCase : Optional[int] = F'{text_of_1_token} {text_of_1_token}' _lowerCAmelCase : str = self.rust_tokenizer_class.from_pretrained( snake_case__ , use_fast=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ ) _lowerCAmelCase : Optional[int] = tokenizer_r(snake_case__ , return_offsets_mapping=snake_case__ , add_special_tokens=snake_case__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(snake_case__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(snake_case__ ) + 1, len(snake_case__ ) + 1 + len(snake_case__ )) , ) _lowerCAmelCase : Dict = self.rust_tokenizer_class.from_pretrained( snake_case__ , use_fast=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ ) _lowerCAmelCase : Optional[Any] = tokenizer_r(snake_case__ , return_offsets_mapping=snake_case__ , add_special_tokens=snake_case__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(snake_case__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(snake_case__ ) + 1, len(snake_case__ ) + 1 + len(snake_case__ )) , ) _lowerCAmelCase : str = self.rust_tokenizer_class.from_pretrained( snake_case__ , use_fast=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ ) _lowerCAmelCase : Union[str, Any] = tokenizer_r(snake_case__ , return_offsets_mapping=snake_case__ , add_special_tokens=snake_case__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(snake_case__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(snake_case__ ), len(snake_case__ ) + 1 + len(snake_case__ )) , ) _lowerCAmelCase : List[Any] = self.rust_tokenizer_class.from_pretrained( snake_case__ , use_fast=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ ) _lowerCAmelCase : List[Any] = tokenizer_r(snake_case__ , return_offsets_mapping=snake_case__ , add_special_tokens=snake_case__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(snake_case__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(snake_case__ ), len(snake_case__ ) + 1 + len(snake_case__ )) , ) _lowerCAmelCase : str = F' {text}' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) _lowerCAmelCase : List[str] = self.rust_tokenizer_class.from_pretrained( snake_case__ , use_fast=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ ) _lowerCAmelCase : int = tokenizer_r(snake_case__ , return_offsets_mapping=snake_case__ , add_special_tokens=snake_case__ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(snake_case__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(snake_case__ ) + 1, 1 + len(snake_case__ ) + 1 + len(snake_case__ )) , ) _lowerCAmelCase : Any = self.rust_tokenizer_class.from_pretrained( snake_case__ , use_fast=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ ) _lowerCAmelCase : List[Any] = tokenizer_r(snake_case__ , return_offsets_mapping=snake_case__ , add_special_tokens=snake_case__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(snake_case__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(snake_case__ ), 1 + len(snake_case__ ) + 1 + len(snake_case__ )) , ) _lowerCAmelCase : str = self.rust_tokenizer_class.from_pretrained( snake_case__ , use_fast=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ ) _lowerCAmelCase : Optional[int] = tokenizer_r(snake_case__ , return_offsets_mapping=snake_case__ , add_special_tokens=snake_case__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(snake_case__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(snake_case__ ), 1 + len(snake_case__ ) + 1 + len(snake_case__ )) , )
707
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase : Tuple = False lowerCAmelCase : str = True lowerCAmelCase : List[Any] = False if __name__ == "__main__": lowerCAmelCase : Any = argparse.ArgumentParser() parser.add_argument( """--repo_path""", default=None, type=str, required=True, help="""The config json file corresponding to the architecture.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") lowerCAmelCase : Optional[int] = parser.parse_args() lowerCAmelCase : int = { """image_size""": """sample_size""", """num_res_blocks""": """layers_per_block""", """block_channels""": """block_out_channels""", """down_blocks""": """down_block_types""", """up_blocks""": """up_block_types""", """downscale_freq_shift""": """freq_shift""", """resnet_num_groups""": """norm_num_groups""", """resnet_act_fn""": """act_fn""", """resnet_eps""": """norm_eps""", """num_head_channels""": """attention_head_dim""", } lowerCAmelCase : int = { """time_steps""": """time_proj""", """mid""": """mid_block""", """downsample_blocks""": """down_blocks""", """upsample_blocks""": """up_blocks""", } lowerCAmelCase : Optional[Any] = """""" if has_file(args.repo_path, """config.json""") else """unet""" with open(os.path.join(args.repo_path, subfolder, """config.json"""), """r""", encoding="""utf-8""") as reader: lowerCAmelCase : int = reader.read() lowerCAmelCase : List[str] = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, """config.json"""): lowerCAmelCase : str = UNetaDModel(**config) else: lowerCAmelCase : Union[str, Any] = UNetaDConditionModel if """ldm-text2im-large-256""" in args.repo_path else UNetaDModel lowerCAmelCase : Dict = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase : Union[str, Any] = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase : str = config[key] del config[key] lowerCAmelCase : Optional[int] = [k.replace("""UNetRes""", """""") for k in config["""down_block_types"""]] lowerCAmelCase : Dict = [k.replace("""UNetRes""", """""") for k in config["""up_block_types"""]] if do_only_weights: lowerCAmelCase : Tuple = torch.load(os.path.join(args.repo_path, subfolder, """diffusion_pytorch_model.bin""")) lowerCAmelCase : str = {} for param_key, param_value in state_dict.items(): if param_key.endswith(""".op.bias""") or param_key.endswith(""".op.weight"""): continue lowerCAmelCase : str = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split(""".""")[0] == key: lowerCAmelCase : Dict = param_value lowerCAmelCase : Tuple = True if not has_changed: lowerCAmelCase : Tuple = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
630
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) lowerCAmelCase : int = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys lowerCAmelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
708
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = None ): '''simple docstring''' super().__init__() _lowerCAmelCase : Union[str, Any] = pad_token_id _lowerCAmelCase : List[Any] = max_length _lowerCAmelCase : Tuple = vocab _lowerCAmelCase : str = merges _lowerCAmelCase : List[str] = BytePairTokenizer(snake_case__ , snake_case__ , sequence_length=snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = [' '.join(snake_case__ ) for m in tokenizer.bpe_ranks.keys()] _lowerCAmelCase : Any = tokenizer.get_vocab() return cls(snake_case__ , snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = GPTaTokenizer.from_pretrained(snake_case__ , *snake_case__ , **snake_case__ ) return cls.from_tokenizer(snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ ): '''simple docstring''' return cls(**snake_case__ ) def a ( self ): '''simple docstring''' return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : str = self.tf_tokenizer(snake_case__ ) _lowerCAmelCase : str = tf.ones_like(snake_case__ ) if self.pad_token_id is not None: # pad the tokens up to max length _lowerCAmelCase : Optional[int] = max_length if max_length is not None else self.max_length if max_length is not None: _lowerCAmelCase , _lowerCAmelCase : str = pad_model_inputs( snake_case__ , max_seq_length=snake_case__ , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
630
0
'''simple docstring''' import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__=7 , snake_case__=3 , snake_case__=30 , snake_case__=400 , snake_case__=True , snake_case__=None , snake_case__=True , snake_case__=[0.5, 0.5, 0.5] , snake_case__=[0.5, 0.5, 0.5] , snake_case__=True , snake_case__=1 / 255 , snake_case__=True , ): '''simple docstring''' _lowerCAmelCase : str = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1333} _lowerCAmelCase : Optional[Any] = parent _lowerCAmelCase : Optional[int] = batch_size _lowerCAmelCase : Optional[Any] = num_channels _lowerCAmelCase : Optional[int] = min_resolution _lowerCAmelCase : Union[str, Any] = max_resolution _lowerCAmelCase : Optional[Any] = do_resize _lowerCAmelCase : str = size _lowerCAmelCase : Optional[Any] = do_normalize _lowerCAmelCase : str = image_mean _lowerCAmelCase : int = image_std _lowerCAmelCase : Union[str, Any] = do_rescale _lowerCAmelCase : Optional[int] = rescale_factor _lowerCAmelCase : List[str] = do_pad def a ( self ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def a ( self , snake_case__ , snake_case__=False ): '''simple docstring''' if not batched: _lowerCAmelCase : Optional[Any] = image_inputs[0] if isinstance(snake_case__ , Image.Image ): _lowerCAmelCase : str = image.size else: _lowerCAmelCase : Union[str, Any] = image.shape[1], image.shape[2] if w < h: _lowerCAmelCase : Any = int(self.size['shortest_edge'] * h / w ) _lowerCAmelCase : Optional[int] = self.size['shortest_edge'] elif w > h: _lowerCAmelCase : List[str] = self.size['shortest_edge'] _lowerCAmelCase : str = int(self.size['shortest_edge'] * w / h ) else: _lowerCAmelCase : Optional[int] = self.size['shortest_edge'] _lowerCAmelCase : List[str] = self.size['shortest_edge'] else: _lowerCAmelCase : Optional[Any] = [] for image in image_inputs: _lowerCAmelCase : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowerCAmelCase : int = max(snake_case__ , key=lambda snake_case__ : item[0] )[0] _lowerCAmelCase : List[str] = max(snake_case__ , key=lambda snake_case__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = ConditionalDetrImageProcessor if is_vision_available() else None def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = ConditionalDetrImageProcessingTester(self ) @property def a ( self ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case__ , 'image_mean' ) ) self.assertTrue(hasattr(snake_case__ , 'image_std' ) ) self.assertTrue(hasattr(snake_case__ , 'do_normalize' ) ) self.assertTrue(hasattr(snake_case__ , 'do_resize' ) ) self.assertTrue(hasattr(snake_case__ , 'size' ) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1333} ) self.assertEqual(image_processor.do_pad , snake_case__ ) _lowerCAmelCase : List[Any] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=snake_case__ ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , snake_case__ ) def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , Image.Image ) # Test not batched input _lowerCAmelCase : List[Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values _lowerCAmelCase : Dict = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCAmelCase : int = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) _lowerCAmelCase : Optional[int] = image_processing(snake_case__ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCAmelCase : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , numpify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , np.ndarray ) # Test not batched input _lowerCAmelCase : int = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values _lowerCAmelCase : List[Any] = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCAmelCase : int = image_processing(snake_case__ , return_tensors='pt' ).pixel_values _lowerCAmelCase : Optional[int] = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case__ , torchify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , torch.Tensor ) # Test not batched input _lowerCAmelCase : Tuple = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values _lowerCAmelCase : int = self.image_processor_tester.get_expected_values(snake_case__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCAmelCase : List[Any] = image_processing(snake_case__ , return_tensors='pt' ).pixel_values _lowerCAmelCase : Dict = self.image_processor_tester.get_expected_values(snake_case__ , batched=snake_case__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: _lowerCAmelCase : Optional[int] = json.loads(f.read() ) _lowerCAmelCase : str = {'image_id': 3_9769, 'annotations': target} # encode them _lowerCAmelCase : List[Any] = ConditionalDetrImageProcessor.from_pretrained('microsoft/conditional-detr-resnet-50' ) _lowerCAmelCase : List[Any] = image_processing(images=snake_case__ , annotations=snake_case__ , return_tensors='pt' ) # verify pixel values _lowerCAmelCase : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , snake_case__ ) _lowerCAmelCase : str = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , snake_case__ , atol=1E-4 ) ) # verify area _lowerCAmelCase : List[str] = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , snake_case__ ) ) # verify boxes _lowerCAmelCase : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , snake_case__ ) _lowerCAmelCase : Any = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , snake_case__ , atol=1E-3 ) ) # verify image_id _lowerCAmelCase : Optional[Any] = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , snake_case__ ) ) # verify is_crowd _lowerCAmelCase : Any = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , snake_case__ ) ) # verify class_labels _lowerCAmelCase : Optional[Any] = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , snake_case__ ) ) # verify orig_size _lowerCAmelCase : Optional[int] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , snake_case__ ) ) # verify size _lowerCAmelCase : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , snake_case__ ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: _lowerCAmelCase : Any = json.loads(f.read() ) _lowerCAmelCase : Optional[Any] = {'file_name': '000000039769.png', 'image_id': 3_9769, 'segments_info': target} _lowerCAmelCase : Union[str, Any] = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them _lowerCAmelCase : Dict = ConditionalDetrImageProcessor(format='coco_panoptic' ) _lowerCAmelCase : Dict = image_processing(images=snake_case__ , annotations=snake_case__ , masks_path=snake_case__ , return_tensors='pt' ) # verify pixel values _lowerCAmelCase : Optional[Any] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['pixel_values'].shape , snake_case__ ) _lowerCAmelCase : Dict = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , snake_case__ , atol=1E-4 ) ) # verify area _lowerCAmelCase : Any = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , snake_case__ ) ) # verify boxes _lowerCAmelCase : str = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , snake_case__ ) _lowerCAmelCase : Optional[Any] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , snake_case__ , atol=1E-3 ) ) # verify image_id _lowerCAmelCase : Optional[Any] = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , snake_case__ ) ) # verify is_crowd _lowerCAmelCase : Any = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , snake_case__ ) ) # verify class_labels _lowerCAmelCase : Optional[Any] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , snake_case__ ) ) # verify masks _lowerCAmelCase : str = 82_2873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , snake_case__ ) # verify orig_size _lowerCAmelCase : Dict = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , snake_case__ ) ) # verify size _lowerCAmelCase : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , snake_case__ ) )
709
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Optional[int] = { """configuration_mctct""": ["""MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MCTCTConfig"""], """feature_extraction_mctct""": ["""MCTCTFeatureExtractor"""], """processing_mctct""": ["""MCTCTProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : List[str] = [ """MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MCTCTForCTC""", """MCTCTModel""", """MCTCTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys lowerCAmelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' from __future__ import annotations import math def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : str = u for i in range(1 , _A ): _lowerCAmelCase : Optional[int] = temp * (u - i) return temp def lowercase (): """simple docstring""" _lowerCAmelCase : Any = int(input('enter the numbers of values: ' ) ) _lowerCAmelCase : list[list[float]] = [] for _ in range(_A ): y.append([] ) for i in range(_A ): for j in range(_A ): y[i].append(_A ) _lowerCAmelCase : Optional[Any] = 0 print('enter the values of parameters in a list: ' ) _lowerCAmelCase : List[str] = list(map(_A , input().split() ) ) print('enter the values of corresponding parameters: ' ) for i in range(_A ): _lowerCAmelCase : str = float(input() ) _lowerCAmelCase : Dict = int(input('enter the value to interpolate: ' ) ) _lowerCAmelCase : Optional[int] = (value - x[0]) / (x[1] - x[0]) # for calculating forward difference table for i in range(1 , _A ): for j in range(n - i ): _lowerCAmelCase : Dict = y[j + 1][i - 1] - y[j][i - 1] _lowerCAmelCase : int = y[0][0] for i in range(1 , _A ): summ += (ucal(_A , _A ) * y[0][i]) / math.factorial(_A ) print(f'the value at {value} is {summ}' ) if __name__ == "__main__": main()
710
'''simple docstring''' lowerCAmelCase : Optional[int] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_00_00)] def lowercase (_A ): """simple docstring""" _lowerCAmelCase : str = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 1_0_0_0_0_0] number //= 1_0_0_0_0_0 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution lowerCAmelCase : list[bool | None] = [None] * 10_00_00_00 lowerCAmelCase : List[str] = True lowerCAmelCase : Union[str, Any] = False def lowercase (_A ): """simple docstring""" if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _lowerCAmelCase : Any = chain(next_number(_A ) ) _lowerCAmelCase : List[str] = number_chain while number < 1_0_0_0_0_0_0_0: _lowerCAmelCase : Tuple = number_chain number *= 1_0 return number_chain def lowercase (_A = 1_0_0_0_0_0_0_0 ): """simple docstring""" for i in range(1 , _A ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(_A ) if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution() = }''')
630
0
'''simple docstring''' def lowercase (_A ): """simple docstring""" if not isinstance(_A , _A ): raise ValueError('Input must be an integer' ) if input_num <= 0: raise ValueError('Input must be positive' ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
711
'''simple docstring''' import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(snake_case__ , 'width_multiplier' ) ) class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=64 , snake_case__=2 , snake_case__=3 , snake_case__="swish" , snake_case__=3 , snake_case__=32 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=True , snake_case__=True , snake_case__=10 , snake_case__=None , snake_case__=0.25 , snake_case__=0.0 , snake_case__=0.0 , ): '''simple docstring''' _lowerCAmelCase : Tuple = parent _lowerCAmelCase : Optional[int] = batch_size _lowerCAmelCase : List[Any] = image_size _lowerCAmelCase : List[Any] = patch_size _lowerCAmelCase : List[str] = num_channels _lowerCAmelCase : Optional[Any] = make_divisible(512 * width_multiplier , divisor=8 ) _lowerCAmelCase : Optional[Any] = hidden_act _lowerCAmelCase : List[Any] = conv_kernel_size _lowerCAmelCase : Optional[Any] = output_stride _lowerCAmelCase : List[Any] = classifier_dropout_prob _lowerCAmelCase : str = use_labels _lowerCAmelCase : List[str] = is_training _lowerCAmelCase : Optional[int] = num_labels _lowerCAmelCase : List[str] = initializer_range _lowerCAmelCase : str = scope _lowerCAmelCase : Any = width_multiplier _lowerCAmelCase : Union[str, Any] = ffn_dropout _lowerCAmelCase : Optional[int] = attn_dropout def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase : Optional[Any] = None _lowerCAmelCase : Dict = None if self.use_labels: _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _lowerCAmelCase : Optional[int] = self.get_config() return config, pixel_values, labels, pixel_labels def a ( self ): '''simple docstring''' return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : str = model(snake_case__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.num_labels _lowerCAmelCase : List[Any] = MobileViTVaForImageClassification(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : int = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.num_labels _lowerCAmelCase : Optional[int] = MobileViTVaForSemanticSegmentation(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Dict = model(snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) _lowerCAmelCase : Any = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Optional[int] = config_and_inputs _lowerCAmelCase : Tuple = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = ( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) __magic_name__ = ( { "feature-extraction": MobileViTVaModel, "image-classification": MobileViTVaForImageClassification, "image-segmentation": MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False def a ( self ): '''simple docstring''' _lowerCAmelCase : int = MobileViTVaModelTester(self ) _lowerCAmelCase : Dict = MobileViTVaConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ ) def a ( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='MobileViTV2 does not use inputs_embeds' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not support input and output embeddings' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not output attentions' ) def a ( self ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason='Got `CUDA error: misaligned address` for tests after this one being run.' ) def a ( self ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : str = model_class(snake_case__ ) _lowerCAmelCase : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCAmelCase : int = [*signature.parameters.keys()] _lowerCAmelCase : Any = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def a ( self ): '''simple docstring''' def check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ): _lowerCAmelCase : Dict = model_class(snake_case__ ) model.to(snake_case__ ) model.eval() with torch.no_grad(): _lowerCAmelCase : Optional[Any] = model(**self._prepare_for_class(snake_case__ , snake_case__ ) ) _lowerCAmelCase : List[str] = outputs.hidden_states _lowerCAmelCase : List[str] = 5 self.assertEqual(len(snake_case__ ) , snake_case__ ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. _lowerCAmelCase : List[Any] = 2 for i in range(len(snake_case__ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : Optional[int] = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCAmelCase : Any = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*snake_case__ ) @slow def a ( self ): '''simple docstring''' for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase : Dict = MobileViTVaModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) def lowercase (): """simple docstring""" _lowerCAmelCase : Any = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def a ( self ): '''simple docstring''' return ( MobileViTImageProcessor.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ) if is_vision_available() else None ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = MobileViTVaForImageClassification.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ).to( snake_case__ ) _lowerCAmelCase : str = self.default_image_processor _lowerCAmelCase : Any = prepare_img() _lowerCAmelCase : Optional[int] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Tuple = model(**snake_case__ ) # verify the logits _lowerCAmelCase : Optional[Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , snake_case__ ) _lowerCAmelCase : Tuple = torch.tensor([-1.6336E00, -7.3204E-02, -5.1883E-01] ).to(snake_case__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Any = model.to(snake_case__ ) _lowerCAmelCase : int = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Optional[int] = prepare_img() _lowerCAmelCase : Dict = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : int = model(**snake_case__ ) _lowerCAmelCase : Dict = outputs.logits # verify the logits _lowerCAmelCase : str = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , snake_case__ ) _lowerCAmelCase : Any = torch.tensor( [ [[7.0863, 7.1525, 6.8201], [6.6931, 6.8770, 6.8933], [6.2978, 7.0366, 6.9636]], [[-3.7134, -3.6712, -3.6675], [-3.5825, -3.3549, -3.4777], [-3.3435, -3.3979, -3.2857]], [[-2.9329, -2.8003, -2.7369], [-3.0564, -2.4780, -2.0207], [-2.6889, -1.9298, -1.7640]], ] , device=snake_case__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : List[Any] = model.to(snake_case__ ) _lowerCAmelCase : str = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Tuple = prepare_img() _lowerCAmelCase : List[str] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Any = model(**snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.logits.detach().cpu() _lowerCAmelCase : Any = image_processor.post_process_semantic_segmentation(outputs=snake_case__ , target_sizes=[(50, 60)] ) _lowerCAmelCase : List[Any] = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , snake_case__ ) _lowerCAmelCase : List[str] = image_processor.post_process_semantic_segmentation(outputs=snake_case__ ) _lowerCAmelCase : Tuple = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , snake_case__ )
630
0
'''simple docstring''' lowerCAmelCase : Any = """ # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git """ lowerCAmelCase : Dict = [{"""type""": """code""", """content""": INSTALL_CONTENT}] lowerCAmelCase : str = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
712
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = FlaxXLMRobertaModel.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Dict = 'The dog is cute and lives in the garden house' _lowerCAmelCase : List[str] = jnp.array([tokenizer.encode(snake_case__ )] ) _lowerCAmelCase : Optional[int] = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim _lowerCAmelCase : Tuple = jnp.array( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) _lowerCAmelCase : Union[str, Any] = model(snake_case__ )['last_hidden_state'] self.assertEqual(output.shape , snake_case__ ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , snake_case__ , atol=1E-3 ) )
630
0
'''simple docstring''' from __future__ import annotations def lowercase (_A ): """simple docstring""" _lowerCAmelCase : List[str] = 2 _lowerCAmelCase : List[Any] = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(_A ) if n > 1: factors.append(_A ) return factors if __name__ == "__main__": import doctest doctest.testmod()
713
'''simple docstring''' def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Dict = len(_A ) while cur > 1: # Find the maximum number in arr _lowerCAmelCase : int = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi _lowerCAmelCase : Dict = arr[mi::-1] + arr[mi + 1 : len(_A )] # Reverse whole list _lowerCAmelCase : Optional[int] = arr[cur - 1 :: -1] + arr[cur : len(_A )] cur -= 1 return arr if __name__ == "__main__": lowerCAmelCase : Tuple = input("""Enter numbers separated by a comma:\n""").strip() lowerCAmelCase : Tuple = [int(item) for item in user_input.split(""",""")] print(pancake_sort(unsorted))
630
0
import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels lowerCAmelCase : Union[str, Any] = object() # For specifying empty leaf dict `{}` lowerCAmelCase : Any = object() def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : str = tuple((re.compile(x + '$' ) for x in qs) ) for i in range(len(_A ) - len(_A ) + 1 ): _lowerCAmelCase : Dict = [x.match(_A ) for x, y in zip(_A , ks[i:] )] if matches and all(_A ): return True return False def lowercase (_A ): """simple docstring""" def replace(_A , _A ): for rule, replacement in rules: if _match(_A , _A ): return replacement return val return replace def lowercase (): """simple docstring""" return [ # embeddings (("transformer", "wpe", "embedding"), P('mp' , _A )), (("transformer", "wte", "embedding"), P('mp' , _A )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(_A , 'mp' )), (("attention", "out_proj", "kernel"), P('mp' , _A )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(_A , 'mp' )), (("mlp", "c_fc", "bias"), P('mp' )), (("mlp", "c_proj", "kernel"), P('mp' , _A )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Any = _get_partition_rules() _lowerCAmelCase : Tuple = _replacement_rules(_A ) _lowerCAmelCase : Tuple = {k: _unmatched for k in flatten_dict(_A )} _lowerCAmelCase : List[Any] = {k: replace(_A , _A ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(_A ) )
714
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) lowerCAmelCase : str = { """EleutherAI/gpt-j-6B""": """https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json""", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "gptj" __magic_name__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , snake_case__=5_0400 , snake_case__=2048 , snake_case__=4096 , snake_case__=28 , snake_case__=16 , snake_case__=64 , snake_case__=None , snake_case__="gelu_new" , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=1E-5 , snake_case__=0.02 , snake_case__=True , snake_case__=5_0256 , snake_case__=5_0256 , snake_case__=False , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : int = vocab_size _lowerCAmelCase : Optional[int] = n_positions _lowerCAmelCase : Optional[int] = n_embd _lowerCAmelCase : Optional[int] = n_layer _lowerCAmelCase : str = n_head _lowerCAmelCase : Tuple = n_inner _lowerCAmelCase : Tuple = rotary_dim _lowerCAmelCase : Optional[int] = activation_function _lowerCAmelCase : Any = resid_pdrop _lowerCAmelCase : List[str] = embd_pdrop _lowerCAmelCase : int = attn_pdrop _lowerCAmelCase : Any = layer_norm_epsilon _lowerCAmelCase : Optional[int] = initializer_range _lowerCAmelCase : List[str] = use_cache _lowerCAmelCase : Dict = bos_token_id _lowerCAmelCase : Any = eos_token_id super().__init__( bos_token_id=snake_case__ , eos_token_id=snake_case__ , tie_word_embeddings=snake_case__ , **snake_case__ ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = "default" , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' super().__init__(snake_case__ , task=snake_case__ , patching_specs=snake_case__ , use_past=snake_case__ ) if not getattr(self._config , 'pad_token_id' , snake_case__ ): # TODO: how to do that better? _lowerCAmelCase : Any = 0 @property def a ( self ): '''simple docstring''' _lowerCAmelCase : str = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(snake_case__ , direction='inputs' ) _lowerCAmelCase : int = {0: 'batch', 1: 'past_sequence + sequence'} else: _lowerCAmelCase : int = {0: 'batch', 1: 'sequence'} return common_inputs @property def a ( self ): '''simple docstring''' return self._config.n_layer @property def a ( self ): '''simple docstring''' return self._config.n_head def a ( self , snake_case__ , snake_case__ = -1 , snake_case__ = -1 , snake_case__ = False , snake_case__ = None , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = super(snake_case__ , self ).generate_dummy_inputs( snake_case__ , batch_size=snake_case__ , seq_length=snake_case__ , is_pair=snake_case__ , framework=snake_case__ ) # We need to order the input in the way they appears in the forward() _lowerCAmelCase : Any = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = common_inputs['input_ids'].shape # Not using the same length for past_key_values _lowerCAmelCase : Any = seqlen + 2 _lowerCAmelCase : Optional[int] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _lowerCAmelCase : Tuple = [ (torch.zeros(snake_case__ ), torch.zeros(snake_case__ )) for _ in range(self.num_layers ) ] _lowerCAmelCase : Tuple = common_inputs['attention_mask'] if self.use_past: _lowerCAmelCase : Any = ordered_inputs['attention_mask'].dtype _lowerCAmelCase : Union[str, Any] = torch.cat( [ordered_inputs['attention_mask'], torch.ones(snake_case__ , snake_case__ , dtype=snake_case__ )] , dim=1 ) return ordered_inputs @property def a ( self ): '''simple docstring''' return 13
630
0
'''simple docstring''' from abc import ABC, abstractmethod from typing import List, Optional class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self ): '''simple docstring''' self.test() def a ( self ): '''simple docstring''' _lowerCAmelCase : str = 0 _lowerCAmelCase : List[str] = False while not completed: if counter == 1: self.reset() _lowerCAmelCase : Dict = self.advance() if not self.does_advance(snake_case__ ): raise Exception( 'Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.' ) _lowerCAmelCase : Optional[Any] = self.update(snake_case__ ) counter += 1 if counter > 1_0000: raise Exception('update() does not fulfill the constraint.' ) if self.remaining() != 0: raise Exception('Custom Constraint is not defined correctly.' ) @abstractmethod def a ( self ): '''simple docstring''' raise NotImplementedError( F'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def a ( self , snake_case__ ): '''simple docstring''' raise NotImplementedError( F'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def a ( self , snake_case__ ): '''simple docstring''' raise NotImplementedError( F'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def a ( self ): '''simple docstring''' raise NotImplementedError( F'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def a ( self ): '''simple docstring''' raise NotImplementedError( F'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def a ( self , snake_case__=False ): '''simple docstring''' raise NotImplementedError( F'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' super(snake_case__ , self ).__init__() if not isinstance(snake_case__ , snake_case__ ) or len(snake_case__ ) == 0: raise ValueError(F'`token_ids` has to be a non-empty list, but is {token_ids}.' ) if any((not isinstance(snake_case__ , snake_case__ ) or token_id < 0) for token_id in token_ids ): raise ValueError(F'Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.' ) _lowerCAmelCase : List[str] = token_ids _lowerCAmelCase : Optional[Any] = len(self.token_ids ) _lowerCAmelCase : List[str] = -1 # the index of the currently fulfilled step _lowerCAmelCase : int = False def a ( self ): '''simple docstring''' if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def a ( self , snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise ValueError(F'`token_id` has to be an `int`, but is {token_id} of type {type(snake_case__ )}' ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def a ( self , snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise ValueError(F'`token_id` has to be an `int`, but is {token_id} of type {type(snake_case__ )}' ) _lowerCAmelCase : Union[str, Any] = False _lowerCAmelCase : int = False _lowerCAmelCase : Optional[int] = False if self.does_advance(snake_case__ ): self.fulfilled_idx += 1 _lowerCAmelCase : str = True if self.fulfilled_idx == (self.seqlen - 1): _lowerCAmelCase : Optional[int] = True _lowerCAmelCase : Dict = completed else: # failed to make progress. _lowerCAmelCase : List[Any] = True self.reset() return stepped, completed, reset def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = False _lowerCAmelCase : Optional[int] = 0 def a ( self ): '''simple docstring''' return self.seqlen - (self.fulfilled_idx + 1) def a ( self , snake_case__=False ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = PhrasalConstraint(self.token_ids ) if stateful: _lowerCAmelCase : Dict = self.seqlen _lowerCAmelCase : Optional[int] = self.fulfilled_idx _lowerCAmelCase : List[Any] = self.completed return new_constraint class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__=True ): '''simple docstring''' _lowerCAmelCase : Any = max([len(snake_case__ ) for one in nested_token_ids] ) _lowerCAmelCase : int = {} for token_ids in nested_token_ids: _lowerCAmelCase : Tuple = root for tidx, token_id in enumerate(snake_case__ ): if token_id not in level: _lowerCAmelCase : Dict = {} _lowerCAmelCase : Tuple = level[token_id] if no_subsets and self.has_subsets(snake_case__ , snake_case__ ): raise ValueError( 'Each list in `nested_token_ids` can\'t be a complete subset of another list, but is' F' {nested_token_ids}.' ) _lowerCAmelCase : Optional[int] = root def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[str] = self.trie for current_token in current_seq: _lowerCAmelCase : List[Any] = start[current_token] _lowerCAmelCase : Optional[int] = list(start.keys() ) return next_tokens def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = self.next_tokens(snake_case__ ) return len(snake_case__ ) == 0 def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = list(root.values() ) if len(snake_case__ ) == 0: return 1 else: return sum([self.count_leaves(snake_case__ ) for nn in next_nodes] ) def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = self.count_leaves(snake_case__ ) return len(snake_case__ ) != leaf_count class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' super(snake_case__ , self ).__init__() if not isinstance(snake_case__ , snake_case__ ) or len(snake_case__ ) == 0: raise ValueError(F'`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.' ) if any(not isinstance(snake_case__ , snake_case__ ) for token_ids in nested_token_ids ): raise ValueError(F'`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.' ) if any( any((not isinstance(snake_case__ , snake_case__ ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( F'Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.' ) _lowerCAmelCase : Dict = DisjunctiveTrie(snake_case__ ) _lowerCAmelCase : List[str] = nested_token_ids _lowerCAmelCase : int = self.trie.max_height _lowerCAmelCase : str = [] _lowerCAmelCase : str = False def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = self.trie.next_tokens(self.current_seq ) if len(snake_case__ ) == 0: return None else: return token_list def a ( self , snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise ValueError(F'`token_id` is supposed to be type `int`, but is {token_id} of type {type(snake_case__ )}' ) _lowerCAmelCase : Optional[Any] = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def a ( self , snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise ValueError(F'`token_id` is supposed to be type `int`, but is {token_id} of type {type(snake_case__ )}' ) _lowerCAmelCase : str = False _lowerCAmelCase : str = False _lowerCAmelCase : Tuple = False if self.does_advance(snake_case__ ): self.current_seq.append(snake_case__ ) _lowerCAmelCase : List[str] = True else: _lowerCAmelCase : Optional[int] = True self.reset() _lowerCAmelCase : List[str] = self.trie.reached_leaf(self.current_seq ) _lowerCAmelCase : Optional[Any] = completed return stepped, completed, reset def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = False _lowerCAmelCase : int = [] def a ( self ): '''simple docstring''' if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def a ( self , snake_case__=False ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = DisjunctiveConstraint(self.token_ids ) if stateful: _lowerCAmelCase : List[str] = self.seqlen _lowerCAmelCase : Optional[int] = self.current_seq _lowerCAmelCase : Tuple = self.completed return new_constraint class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = constraints # max # of steps required to fulfill a given constraint _lowerCAmelCase : List[str] = max([c.seqlen for c in constraints] ) _lowerCAmelCase : int = len(snake_case__ ) _lowerCAmelCase : Dict = False self.init_state() def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = [] _lowerCAmelCase : str = None _lowerCAmelCase : Optional[Any] = [constraint.copy(stateful=snake_case__ ) for constraint in self.constraints] def a ( self ): '''simple docstring''' _lowerCAmelCase : int = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" _lowerCAmelCase : Any = constraint.advance() if isinstance(snake_case__ , snake_case__ ): token_list.append(snake_case__ ) elif isinstance(snake_case__ , snake_case__ ): token_list.extend(snake_case__ ) else: _lowerCAmelCase : Optional[Any] = self.inprogress_constraint.advance() if isinstance(snake_case__ , snake_case__ ): token_list.append(snake_case__ ) elif isinstance(snake_case__ , snake_case__ ): token_list.extend(snake_case__ ) if len(snake_case__ ) == 0: return None else: return token_list def a ( self , snake_case__ ): '''simple docstring''' self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint _lowerCAmelCase : Union[str, Any] = self.add(snake_case__ ) # the entire list of constraints are fulfilled if self.completed: break def a ( self , snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ): raise ValueError(F'`token_id` should be an `int`, but is `{token_id}`.' ) _lowerCAmelCase : List[Any] = False, False if self.completed: _lowerCAmelCase : List[Any] = True _lowerCAmelCase : List[Any] = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state _lowerCAmelCase : int = self.inprogress_constraint.update(snake_case__ ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=snake_case__ ) ) _lowerCAmelCase : int = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) _lowerCAmelCase : Optional[int] = None if len(self.pending_constraints ) == 0: # we're done! _lowerCAmelCase : Tuple = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(snake_case__ ): _lowerCAmelCase : str = pending_constraint.update(snake_case__ ) if not stepped: raise Exception( '`constraint.update(token_id)` is not yielding incremental progress, ' 'even though `constraint.does_advance(token_id)` is true.' ) if complete: self.complete_constraints.append(snake_case__ ) _lowerCAmelCase : str = None if not complete and stepped: _lowerCAmelCase : int = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". _lowerCAmelCase : Any = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. _lowerCAmelCase : List[str] = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def a ( self , snake_case__=True ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: _lowerCAmelCase : List[str] = [ constraint.copy(stateful=snake_case__ ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: _lowerCAmelCase : str = self.inprogress_constraint.copy(stateful=snake_case__ ) _lowerCAmelCase : List[str] = [constraint.copy() for constraint in self.pending_constraints] return new_state
715
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Any = { """configuration_x_clip""": [ """XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XCLIPConfig""", """XCLIPTextConfig""", """XCLIPVisionConfig""", ], """processing_x_clip""": ["""XCLIPProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """XCLIPModel""", """XCLIPPreTrainedModel""", """XCLIPTextModel""", """XCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys lowerCAmelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Dict = len(_A ) while cur > 1: # Find the maximum number in arr _lowerCAmelCase : int = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi _lowerCAmelCase : Dict = arr[mi::-1] + arr[mi + 1 : len(_A )] # Reverse whole list _lowerCAmelCase : Optional[int] = arr[cur - 1 :: -1] + arr[cur : len(_A )] cur -= 1 return arr if __name__ == "__main__": lowerCAmelCase : Tuple = input("""Enter numbers separated by a comma:\n""").strip() lowerCAmelCase : Tuple = [int(item) for item in user_input.split(""",""")] print(pancake_sort(unsorted))
716
'''simple docstring''' import math from datetime import datetime, timedelta def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Optional[Any] = year % 1_9 _lowerCAmelCase : Any = year % 4 _lowerCAmelCase : Optional[int] = year % 7 _lowerCAmelCase : int = math.floor(year / 1_0_0 ) _lowerCAmelCase : Dict = math.floor((1_3 + 8 * leap_day_inhibits) / 2_5 ) _lowerCAmelCase : Optional[Any] = leap_day_inhibits / 4 _lowerCAmelCase : Dict = ( 1_5 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 3_0 _lowerCAmelCase : List[Any] = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 _lowerCAmelCase : Dict = (1_9 * metonic_cycle + secular_moon_shift) % 3_0 # PHM -> Paschal Full Moon _lowerCAmelCase : Union[str, Any] = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 2_9 and days_from_phm_to_sunday == 6: return datetime(_A , 4 , 1_9 ) elif days_to_add == 2_8 and days_from_phm_to_sunday == 6: return datetime(_A , 4 , 1_8 ) else: return datetime(_A , 3 , 2_2 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (19_94, 20_00, 20_10, 20_21, 20_23): lowerCAmelCase : List[str] = """will be""" if year > datetime.now().year else """was""" print(F'''Easter in {year} {tense} {gauss_easter(year)}''')
630
0
'''simple docstring''' import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' super().tearDown() gc.collect() def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = FlaxStableDiffusionPipeline.from_pretrained( 'stabilityai/stable-diffusion-2' , revision='bf16' , dtype=jnp.bfloataa , ) _lowerCAmelCase : List[Any] = 'A painting of a squirrel eating a burger' _lowerCAmelCase : Optional[int] = jax.device_count() _lowerCAmelCase : Tuple = num_samples * [prompt] _lowerCAmelCase : Optional[int] = sd_pipe.prepare_inputs(snake_case__ ) _lowerCAmelCase : Tuple = replicate(snake_case__ ) _lowerCAmelCase : str = shard(snake_case__ ) _lowerCAmelCase : Tuple = jax.random.PRNGKey(0 ) _lowerCAmelCase : Any = jax.random.split(snake_case__ , jax.device_count() ) _lowerCAmelCase : Dict = sd_pipe(snake_case__ , snake_case__ , snake_case__ , num_inference_steps=25 , jit=snake_case__ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _lowerCAmelCase : Dict = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _lowerCAmelCase : Optional[int] = images[0, 253:256, 253:256, -1] _lowerCAmelCase : List[str] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _lowerCAmelCase : int = jnp.array([0.4238, 0.4414, 0.4395, 0.4453, 0.4629, 0.4590, 0.4531, 0.4_5508, 0.4512] ) print(F'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2 def a ( self ): '''simple docstring''' _lowerCAmelCase : int = 'stabilityai/stable-diffusion-2' _lowerCAmelCase : Optional[Any] = FlaxDPMSolverMultistepScheduler.from_pretrained(snake_case__ , subfolder='scheduler' ) _lowerCAmelCase : Optional[int] = FlaxStableDiffusionPipeline.from_pretrained( snake_case__ , scheduler=snake_case__ , revision='bf16' , dtype=jnp.bfloataa , ) _lowerCAmelCase : Any = scheduler_params _lowerCAmelCase : Tuple = 'A painting of a squirrel eating a burger' _lowerCAmelCase : Optional[int] = jax.device_count() _lowerCAmelCase : Any = num_samples * [prompt] _lowerCAmelCase : Dict = sd_pipe.prepare_inputs(snake_case__ ) _lowerCAmelCase : Tuple = replicate(snake_case__ ) _lowerCAmelCase : Optional[int] = shard(snake_case__ ) _lowerCAmelCase : Optional[int] = jax.random.PRNGKey(0 ) _lowerCAmelCase : Dict = jax.random.split(snake_case__ , jax.device_count() ) _lowerCAmelCase : List[Any] = sd_pipe(snake_case__ , snake_case__ , snake_case__ , num_inference_steps=25 , jit=snake_case__ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) _lowerCAmelCase : Dict = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _lowerCAmelCase : Optional[int] = images[0, 253:256, 253:256, -1] _lowerCAmelCase : Dict = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _lowerCAmelCase : Any = jnp.array([0.4336, 0.4_2969, 0.4453, 0.4199, 0.4297, 0.4531, 0.4434, 0.4434, 0.4297] ) print(F'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
717
'''simple docstring''' import unittest from knapsack import greedy_knapsack as kp class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [10, 20, 30, 40, 50, 60] _lowerCAmelCase : Union[str, Any] = [2, 4, 6, 8, 10, 12] _lowerCAmelCase : Dict = 100 self.assertEqual(kp.calc_profit(snake_case__ , snake_case__ , snake_case__ ) , 210 ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'max_weight must greater than zero.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'Weight can not be negative.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'Profit can not be negative.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'max_weight must greater than zero.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex( snake_case__ , 'The length of profit and weight must be same.' ) if __name__ == "__main__": unittest.main()
630
0
'''simple docstring''' lowerCAmelCase : Optional[int] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_00_00)] def lowercase (_A ): """simple docstring""" _lowerCAmelCase : str = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 1_0_0_0_0_0] number //= 1_0_0_0_0_0 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution lowerCAmelCase : list[bool | None] = [None] * 10_00_00_00 lowerCAmelCase : List[str] = True lowerCAmelCase : Union[str, Any] = False def lowercase (_A ): """simple docstring""" if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _lowerCAmelCase : Any = chain(next_number(_A ) ) _lowerCAmelCase : List[str] = number_chain while number < 1_0_0_0_0_0_0_0: _lowerCAmelCase : Tuple = number_chain number *= 1_0 return number_chain def lowercase (_A = 1_0_0_0_0_0_0_0 ): """simple docstring""" for i in range(1 , _A ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(_A ) if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution() = }''')
718
'''simple docstring''' def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : Optional[int] = (boundary[1] - boundary[0]) / steps _lowerCAmelCase : Any = boundary[0] _lowerCAmelCase : List[str] = boundary[1] _lowerCAmelCase : Tuple = make_points(_A , _A , _A ) _lowerCAmelCase : Tuple = 0.0 y += (h / 2.0) * f(_A ) for i in x_i: # print(i) y += h * f(_A ) y += (h / 2.0) * f(_A ) return y def lowercase (_A , _A , _A ): """simple docstring""" _lowerCAmelCase : Tuple = a + h while x < (b - h): yield x _lowerCAmelCase : Any = x + h def lowercase (_A ): # enter your function here """simple docstring""" _lowerCAmelCase : int = (x - 0) * (x - 0) return y def lowercase (): """simple docstring""" _lowerCAmelCase : Optional[Any] = 0.0 # Lower bound of integration _lowerCAmelCase : Dict = 1.0 # Upper bound of integration _lowerCAmelCase : Optional[Any] = 10.0 # define number of steps or resolution _lowerCAmelCase : Optional[int] = [a, b] # define boundary of integration _lowerCAmelCase : List[Any] = method_a(_A , _A ) print(f'y = {y}' ) if __name__ == "__main__": main()
630
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCAmelCase : Optional[Any] = { """configuration_poolformer""": [ """POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PoolFormerConfig""", """PoolFormerOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : List[Any] = ["""PoolFormerFeatureExtractor"""] lowerCAmelCase : List[Any] = ["""PoolFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : List[str] = [ """POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PoolFormerForImageClassification""", """PoolFormerModel""", """PoolFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys lowerCAmelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
719
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) lowerCAmelCase : int = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys lowerCAmelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LDMTextToImagePipeline, UNetaDConditionModel from diffusers.utils.testing_utils import ( enable_full_determinism, load_numpy, nightly, require_torch_gpu, slow, torch_device, ) from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = LDMTextToImagePipeline __magic_name__ = TEXT_TO_IMAGE_PARAMS - { "negative_prompt", "negative_prompt_embeds", "cross_attention_kwargs", "prompt_embeds", } __magic_name__ = PipelineTesterMixin.required_optional_params - { "num_images_per_prompt", "callback", "callback_steps", } __magic_name__ = TEXT_TO_IMAGE_BATCH_PARAMS __magic_name__ = False def a ( self ): '''simple docstring''' torch.manual_seed(0 ) _lowerCAmelCase : Optional[int] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) _lowerCAmelCase : Optional[int] = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , ) torch.manual_seed(0 ) _lowerCAmelCase : Optional[Any] = AutoencoderKL( block_out_channels=(32, 64) , in_channels=3 , out_channels=3 , down_block_types=('DownEncoderBlock2D', 'DownEncoderBlock2D') , up_block_types=('UpDecoderBlock2D', 'UpDecoderBlock2D') , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) _lowerCAmelCase : Union[str, Any] = CLIPTextModel(snake_case__ ) _lowerCAmelCase : Tuple = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _lowerCAmelCase : Union[str, Any] = { 'unet': unet, 'scheduler': scheduler, 'vqvae': vae, 'bert': text_encoder, 'tokenizer': tokenizer, } return components def a ( self , snake_case__ , snake_case__=0 ): '''simple docstring''' if str(snake_case__ ).startswith('mps' ): _lowerCAmelCase : List[str] = torch.manual_seed(snake_case__ ) else: _lowerCAmelCase : Any = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) _lowerCAmelCase : List[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase : Any = self.get_dummy_components() _lowerCAmelCase : Optional[int] = LDMTextToImagePipeline(**snake_case__ ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : Any = self.get_dummy_inputs(snake_case__ ) _lowerCAmelCase : str = pipe(**snake_case__ ).images _lowerCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 16, 16, 3) _lowerCAmelCase : Any = np.array([0.6101, 0.6156, 0.5622, 0.4895, 0.6661, 0.3804, 0.5748, 0.6136, 0.5014] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def a ( self , snake_case__ , snake_case__=torch.floataa , snake_case__=0 ): '''simple docstring''' _lowerCAmelCase : int = torch.manual_seed(snake_case__ ) _lowerCAmelCase : Optional[int] = np.random.RandomState(snake_case__ ).standard_normal((1, 4, 32, 32) ) _lowerCAmelCase : Dict = torch.from_numpy(snake_case__ ).to(device=snake_case__ , dtype=snake_case__ ) _lowerCAmelCase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = LDMTextToImagePipeline.from_pretrained('CompVis/ldm-text2im-large-256' ).to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : Tuple = self.get_inputs(snake_case__ ) _lowerCAmelCase : Tuple = pipe(**snake_case__ ).images _lowerCAmelCase : Tuple = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 256, 256, 3) _lowerCAmelCase : int = np.array([0.5_1825, 0.5_2850, 0.5_2543, 0.5_4258, 0.5_2304, 0.5_2569, 0.5_4363, 0.5_5276, 0.5_6878] ) _lowerCAmelCase : Optional[int] = np.abs(expected_slice - image_slice ).max() assert max_diff < 1E-3 @nightly @require_torch_gpu class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def a ( self , snake_case__ , snake_case__=torch.floataa , snake_case__=0 ): '''simple docstring''' _lowerCAmelCase : List[str] = torch.manual_seed(snake_case__ ) _lowerCAmelCase : Any = np.random.RandomState(snake_case__ ).standard_normal((1, 4, 32, 32) ) _lowerCAmelCase : Optional[Any] = torch.from_numpy(snake_case__ ).to(device=snake_case__ , dtype=snake_case__ ) _lowerCAmelCase : int = { 'prompt': 'A painting of a squirrel eating a burger', 'latents': latents, 'generator': generator, 'num_inference_steps': 50, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = LDMTextToImagePipeline.from_pretrained('CompVis/ldm-text2im-large-256' ).to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : Optional[int] = self.get_inputs(snake_case__ ) _lowerCAmelCase : Optional[Any] = pipe(**snake_case__ ).images[0] _lowerCAmelCase : int = load_numpy( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/ldm_text2img/ldm_large_256_ddim.npy' ) _lowerCAmelCase : List[Any] = np.abs(expected_image - image ).max() assert max_diff < 1E-3
720
'''simple docstring''' from collections import Counter from timeit import timeit def lowercase (_A = "" , ): """simple docstring""" return sum(c % 2 for c in Counter(input_str.replace(' ' , '' ).lower() ).values() ) < 2 def lowercase (_A = "" ): """simple docstring""" if len(_A ) == 0: return True _lowerCAmelCase : Union[str, Any] = input_str.replace(' ' , '' ).lower() # character_freq_dict: Stores the frequency of every character in the input string _lowerCAmelCase : dict[str, int] = {} for character in lower_case_input_str: _lowerCAmelCase : Union[str, Any] = character_freq_dict.get(_A , 0 ) + 1 _lowerCAmelCase : List[Any] = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def lowercase (_A = "" ): """simple docstring""" print('\nFor string = ' , _A , ':' ) print( '> can_string_be_rearranged_as_palindrome_counter()' , '\tans =' , can_string_be_rearranged_as_palindrome_counter(_A ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome_counter(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) print( '> can_string_be_rearranged_as_palindrome()' , '\tans =' , can_string_be_rearranged_as_palindrome(_A ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) if __name__ == "__main__": lowerCAmelCase : Tuple = input( """Enter string to determine if it can be rearranged as a palindrome or not: """ ).strip() benchmark(check_str) lowerCAmelCase : Optional[Any] = can_string_be_rearranged_as_palindrome_counter(check_str) print(F'''{check_str} can {"" if status else "not "}be rearranged as a palindrome''')
630
0
'''simple docstring''' from sklearn.metrics import recall_score import datasets lowerCAmelCase : Union[str, Any] = """ Recall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation: Recall = TP / (TP + FN) Where TP is the true positives and FN is the false negatives. """ lowerCAmelCase : List[Any] = """ Args: - **predictions** (`list` of `int`): The predicted labels. - **references** (`list` of `int`): The ground truth labels. - **labels** (`list` of `int`): The set of labels to include when `average` is not set to `binary`, and their order when average is `None`. Labels present in the data can be excluded in this input, for example to calculate a multiclass average ignoring a majority negative class, while labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in y_true and y_pred are used in sorted order. Defaults to None. - **pos_label** (`int`): The class label to use as the 'positive class' when calculating the recall. Defaults to `1`. - **average** (`string`): This parameter is required for multiclass/multilabel targets. If None, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `'binary'`. - `'binary'`: Only report results for the class specified by `pos_label`. This is applicable only if the target labels and predictions are binary. - `'micro'`: Calculate metrics globally by counting the total true positives, false negatives, and false positives. - `'macro'`: Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - `'weighted'`: Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `'macro'` to account for label imbalance. Note that it can result in an F-score that is not between precision and recall. - `'samples'`: Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). - **sample_weight** (`list` of `float`): Sample weights Defaults to `None`. - **zero_division** (): Sets the value to return when there is a zero division. Defaults to . - `'warn'`: If there is a zero division, the return value is `0`, but warnings are also raised. - `0`: If there is a zero division, the return value is `0`. - `1`: If there is a zero division, the return value is `1`. Returns: - **recall** (`float`, or `array` of `float`): Either the general recall score, or the recall scores for individual classes, depending on the values input to `labels` and `average`. Minimum possible value is 0. Maximum possible value is 1. A higher recall means that more of the positive examples have been labeled correctly. Therefore, a higher recall is generally considered better. Examples: Example 1-A simple example with some errors >>> recall_metric = datasets.load_metric('recall') >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1]) >>> print(results) {'recall': 0.6666666666666666} Example 2-The same example as Example 1, but with `pos_label=0` instead of the default `pos_label=1`. >>> recall_metric = datasets.load_metric('recall') >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], pos_label=0) >>> print(results) {'recall': 0.5} Example 3-The same example as Example 1, but with `sample_weight` included. >>> recall_metric = datasets.load_metric('recall') >>> sample_weight = [0.9, 0.2, 0.9, 0.3, 0.8] >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], sample_weight=sample_weight) >>> print(results) {'recall': 0.55} Example 4-A multiclass example, using different averages. >>> recall_metric = datasets.load_metric('recall') >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = recall_metric.compute(predictions=predictions, references=references, average='macro') >>> print(results) {'recall': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average='micro') >>> print(results) {'recall': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average='weighted') >>> print(results) {'recall': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {'recall': array([1., 0., 0.])} """ lowerCAmelCase : int = """ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase__ ( datasets.Metric ): """simple docstring""" def a ( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('int32' ) ), 'references': datasets.Sequence(datasets.Value('int32' ) ), } if self.config_name == 'multilabel' else { 'predictions': datasets.Value('int32' ), 'references': datasets.Value('int32' ), } ) , reference_urls=['https://scikit-learn.org/stable/modules/generated/sklearn.metrics.recall_score.html'] , ) def a ( self , snake_case__ , snake_case__ , snake_case__=None , snake_case__=1 , snake_case__="binary" , snake_case__=None , snake_case__="warn" , ): '''simple docstring''' _lowerCAmelCase : List[str] = recall_score( snake_case__ , snake_case__ , labels=snake_case__ , pos_label=snake_case__ , average=snake_case__ , sample_weight=snake_case__ , zero_division=snake_case__ , ) return {"recall": float(snake_case__ ) if score.size == 1 else score}
721
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase : str = logging.get_logger(__name__) lowerCAmelCase : int = { """facebook/data2vec-text-base""": """https://huggingface.co/data2vec/resolve/main/config.json""", } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "data2vec-text" def __init__( self , snake_case__=3_0522 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=2 , snake_case__=0.02 , snake_case__=1E-12 , snake_case__=1 , snake_case__=0 , snake_case__=2 , snake_case__="absolute" , snake_case__=True , snake_case__=None , **snake_case__ , ): '''simple docstring''' super().__init__(pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ , **snake_case__ ) _lowerCAmelCase : List[Any] = vocab_size _lowerCAmelCase : Tuple = hidden_size _lowerCAmelCase : Dict = num_hidden_layers _lowerCAmelCase : int = num_attention_heads _lowerCAmelCase : str = hidden_act _lowerCAmelCase : Any = intermediate_size _lowerCAmelCase : Any = hidden_dropout_prob _lowerCAmelCase : Optional[int] = attention_probs_dropout_prob _lowerCAmelCase : str = max_position_embeddings _lowerCAmelCase : Any = type_vocab_size _lowerCAmelCase : int = initializer_range _lowerCAmelCase : List[str] = layer_norm_eps _lowerCAmelCase : List[Any] = position_embedding_type _lowerCAmelCase : str = use_cache _lowerCAmelCase : Union[str, Any] = classifier_dropout class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" @property def a ( self ): '''simple docstring''' if self.task == "multiple-choice": _lowerCAmelCase : Union[str, Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _lowerCAmelCase : List[Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
630
0
'''simple docstring''' import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...image_processing_utils import ImageProcessingMixin from ...utils import CONFIG_NAME, IMAGE_PROCESSOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) lowerCAmelCase : Tuple = logging.get_logger(__name__) lowerCAmelCase : Any = OrderedDict( [ ("""align""", """EfficientNetImageProcessor"""), ("""beit""", """BeitImageProcessor"""), ("""bit""", """BitImageProcessor"""), ("""blip""", """BlipImageProcessor"""), ("""blip-2""", """BlipImageProcessor"""), ("""bridgetower""", """BridgeTowerImageProcessor"""), ("""chinese_clip""", """ChineseCLIPImageProcessor"""), ("""clip""", """CLIPImageProcessor"""), ("""clipseg""", """ViTImageProcessor"""), ("""conditional_detr""", """ConditionalDetrImageProcessor"""), ("""convnext""", """ConvNextImageProcessor"""), ("""convnextv2""", """ConvNextImageProcessor"""), ("""cvt""", """ConvNextImageProcessor"""), ("""data2vec-vision""", """BeitImageProcessor"""), ("""deformable_detr""", """DeformableDetrImageProcessor"""), ("""deit""", """DeiTImageProcessor"""), ("""deta""", """DetaImageProcessor"""), ("""detr""", """DetrImageProcessor"""), ("""dinat""", """ViTImageProcessor"""), ("""donut-swin""", """DonutImageProcessor"""), ("""dpt""", """DPTImageProcessor"""), ("""efficientformer""", """EfficientFormerImageProcessor"""), ("""efficientnet""", """EfficientNetImageProcessor"""), ("""flava""", """FlavaImageProcessor"""), ("""focalnet""", """BitImageProcessor"""), ("""git""", """CLIPImageProcessor"""), ("""glpn""", """GLPNImageProcessor"""), ("""groupvit""", """CLIPImageProcessor"""), ("""imagegpt""", """ImageGPTImageProcessor"""), ("""instructblip""", """BlipImageProcessor"""), ("""layoutlmv2""", """LayoutLMv2ImageProcessor"""), ("""layoutlmv3""", """LayoutLMv3ImageProcessor"""), ("""levit""", """LevitImageProcessor"""), ("""mask2former""", """Mask2FormerImageProcessor"""), ("""maskformer""", """MaskFormerImageProcessor"""), ("""mgp-str""", """ViTImageProcessor"""), ("""mobilenet_v1""", """MobileNetV1ImageProcessor"""), ("""mobilenet_v2""", """MobileNetV2ImageProcessor"""), ("""mobilevit""", """MobileViTImageProcessor"""), ("""mobilevit""", """MobileViTImageProcessor"""), ("""mobilevitv2""", """MobileViTImageProcessor"""), ("""nat""", """ViTImageProcessor"""), ("""oneformer""", """OneFormerImageProcessor"""), ("""owlvit""", """OwlViTImageProcessor"""), ("""perceiver""", """PerceiverImageProcessor"""), ("""pix2struct""", """Pix2StructImageProcessor"""), ("""poolformer""", """PoolFormerImageProcessor"""), ("""regnet""", """ConvNextImageProcessor"""), ("""resnet""", """ConvNextImageProcessor"""), ("""sam""", """SamImageProcessor"""), ("""segformer""", """SegformerImageProcessor"""), ("""swiftformer""", """ViTImageProcessor"""), ("""swin""", """ViTImageProcessor"""), ("""swin2sr""", """Swin2SRImageProcessor"""), ("""swinv2""", """ViTImageProcessor"""), ("""table-transformer""", """DetrImageProcessor"""), ("""timesformer""", """VideoMAEImageProcessor"""), ("""tvlt""", """TvltImageProcessor"""), ("""upernet""", """SegformerImageProcessor"""), ("""van""", """ConvNextImageProcessor"""), ("""videomae""", """VideoMAEImageProcessor"""), ("""vilt""", """ViltImageProcessor"""), ("""vit""", """ViTImageProcessor"""), ("""vit_hybrid""", """ViTHybridImageProcessor"""), ("""vit_mae""", """ViTImageProcessor"""), ("""vit_msn""", """ViTImageProcessor"""), ("""xclip""", """CLIPImageProcessor"""), ("""yolos""", """YolosImageProcessor"""), ] ) lowerCAmelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, IMAGE_PROCESSOR_MAPPING_NAMES) def lowercase (_A ) -> List[Any]: """simple docstring""" for module_name, extractors in IMAGE_PROCESSOR_MAPPING_NAMES.items(): if class_name in extractors: _lowerCAmelCase : int = model_type_to_module_name(_A ) _lowerCAmelCase : Tuple = importlib.import_module(f'.{module_name}' , 'transformers.models' ) try: return getattr(_A , _A ) except AttributeError: continue for _, extractor in IMAGE_PROCESSOR_MAPPING._extra_content.items(): if getattr(_A , '__name__' , _A ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. _lowerCAmelCase : List[Any] = importlib.import_module('transformers' ) if hasattr(_A , _A ): return getattr(_A , _A ) return None def lowercase (_A , _A = None , _A = False , _A = False , _A = None , _A = None , _A = None , _A = False , **_A , ) -> Any: """simple docstring""" _lowerCAmelCase : Tuple = get_file_from_repo( _A , _A , cache_dir=_A , force_download=_A , resume_download=_A , proxies=_A , use_auth_token=_A , revision=_A , local_files_only=_A , ) if resolved_config_file is None: logger.info( 'Could not locate the image processor configuration file, will try to use the model config instead.' ) return {} with open(_A , encoding='utf-8' ) as reader: return json.load(_A ) class UpperCamelCase__ : """simple docstring""" def __init__( self ): '''simple docstring''' raise EnvironmentError( 'AutoImageProcessor is designed to be instantiated ' 'using the `AutoImageProcessor.from_pretrained(pretrained_model_name_or_path)` method.' ) @classmethod @replace_list_option_in_docstrings(snake_case__ ) def a ( cls , snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = kwargs.pop('config' , snake_case__ ) _lowerCAmelCase : int = kwargs.pop('trust_remote_code' , snake_case__ ) _lowerCAmelCase : int = True _lowerCAmelCase : str = ImageProcessingMixin.get_image_processor_dict(snake_case__ , **snake_case__ ) _lowerCAmelCase : Dict = config_dict.get('image_processor_type' , snake_case__ ) _lowerCAmelCase : Optional[Any] = None if "AutoImageProcessor" in config_dict.get('auto_map' , {} ): _lowerCAmelCase : List[Any] = config_dict['auto_map']['AutoImageProcessor'] # If we still don't have the image processor class, check if we're loading from a previous feature extractor config # and if so, infer the image processor class from there. if image_processor_class is None and image_processor_auto_map is None: _lowerCAmelCase : int = config_dict.pop('feature_extractor_type' , snake_case__ ) if feature_extractor_class is not None: logger.warning( 'Could not find image processor class in the image processor config or the model config. Loading' ' based on pattern matching with the model\'s feature extractor configuration.' ) _lowerCAmelCase : List[str] = feature_extractor_class.replace('FeatureExtractor' , 'ImageProcessor' ) if "AutoFeatureExtractor" in config_dict.get('auto_map' , {} ): _lowerCAmelCase : Optional[int] = config_dict['auto_map']['AutoFeatureExtractor'] _lowerCAmelCase : Optional[int] = feature_extractor_auto_map.replace('FeatureExtractor' , 'ImageProcessor' ) logger.warning( 'Could not find image processor auto map in the image processor config or the model config.' ' Loading based on pattern matching with the model\'s feature extractor configuration.' ) # If we don't find the image processor class in the image processor config, let's try the model config. if image_processor_class is None and image_processor_auto_map is None: if not isinstance(snake_case__ , snake_case__ ): _lowerCAmelCase : List[Any] = AutoConfig.from_pretrained(snake_case__ , **snake_case__ ) # It could be in `config.image_processor_type`` _lowerCAmelCase : List[Any] = getattr(snake_case__ , 'image_processor_type' , snake_case__ ) if hasattr(snake_case__ , 'auto_map' ) and "AutoImageProcessor" in config.auto_map: _lowerCAmelCase : Tuple = config.auto_map['AutoImageProcessor'] if image_processor_class is not None: _lowerCAmelCase : Optional[int] = image_processor_class_from_name(snake_case__ ) _lowerCAmelCase : List[str] = image_processor_auto_map is not None _lowerCAmelCase : Optional[Any] = image_processor_class is not None or type(snake_case__ ) in IMAGE_PROCESSOR_MAPPING _lowerCAmelCase : List[str] = resolve_trust_remote_code( snake_case__ , snake_case__ , snake_case__ , snake_case__ ) if has_remote_code and trust_remote_code: _lowerCAmelCase : str = get_class_from_dynamic_module( snake_case__ , snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[int] = kwargs.pop('code_revision' , snake_case__ ) if os.path.isdir(snake_case__ ): image_processor_class.register_for_auto_class() return image_processor_class.from_dict(snake_case__ , **snake_case__ ) elif image_processor_class is not None: return image_processor_class.from_dict(snake_case__ , **snake_case__ ) # Last try: we use the IMAGE_PROCESSOR_MAPPING. elif type(snake_case__ ) in IMAGE_PROCESSOR_MAPPING: _lowerCAmelCase : List[Any] = IMAGE_PROCESSOR_MAPPING[type(snake_case__ )] return image_processor_class.from_dict(snake_case__ , **snake_case__ ) raise ValueError( F'Unrecognized image processor in {pretrained_model_name_or_path}. Should have a ' F'`image_processor_type` key in its {IMAGE_PROCESSOR_NAME} of {CONFIG_NAME}, or one of the following ' F'`model_type` keys in its {CONFIG_NAME}: {", ".join(c for c in IMAGE_PROCESSOR_MAPPING_NAMES.keys() )}' ) @staticmethod def a ( snake_case__ , snake_case__ ): '''simple docstring''' IMAGE_PROCESSOR_MAPPING.register(snake_case__ , snake_case__ )
700
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins lowerCAmelCase : List[str] = ["""tests.fixtures.files""", """tests.fixtures.hub""", """tests.fixtures.fsspec"""] def lowercase (_A , _A ): """simple docstring""" for item in items: if any(marker in item.keywords for marker in ['integration', 'unit'] ): continue item.add_marker(pytest.mark.unit ) def lowercase (_A ): """simple docstring""" config.addinivalue_line('markers' , 'torchaudio_latest: mark test to run with torchaudio>=0.12' ) @pytest.fixture(autouse=_A ) def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : str = tmp_path_factory.getbasetemp() / 'cache' _lowerCAmelCase : Dict = test_hf_cache_home / 'datasets' _lowerCAmelCase : List[Any] = test_hf_cache_home / 'metrics' _lowerCAmelCase : List[Any] = test_hf_cache_home / 'modules' monkeypatch.setattr('datasets.config.HF_DATASETS_CACHE' , str(_A ) ) monkeypatch.setattr('datasets.config.HF_METRICS_CACHE' , str(_A ) ) monkeypatch.setattr('datasets.config.HF_MODULES_CACHE' , str(_A ) ) _lowerCAmelCase : Dict = test_hf_datasets_cache / 'downloads' monkeypatch.setattr('datasets.config.DOWNLOADED_DATASETS_PATH' , str(_A ) ) _lowerCAmelCase : Union[str, Any] = test_hf_datasets_cache / 'downloads' / 'extracted' monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH' , str(_A ) ) @pytest.fixture(autouse=_A , scope='session' ) def lowercase (): """simple docstring""" datasets.disable_progress_bar() @pytest.fixture(autouse=_A ) def lowercase (_A ): """simple docstring""" monkeypatch.setattr('datasets.config.HF_UPDATE_DOWNLOAD_COUNTS' , _A ) @pytest.fixture def lowercase (_A ): """simple docstring""" monkeypatch.setattr('sqlalchemy.util.deprecations.SILENCE_UBER_WARNING' , _A )
630
0
import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = MobileBertTokenizer __magic_name__ = MobileBertTokenizerFast __magic_name__ = True __magic_name__ = True __magic_name__ = filter_non_english __magic_name__ = "google/mobilebert-uncased" def a ( self ): '''simple docstring''' super().setUp() _lowerCAmelCase : Optional[int] = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] _lowerCAmelCase : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) _lowerCAmelCase : Union[str, Any] = [ (tokenizer_def[0], self.pre_trained_model_path, tokenizer_def[2]) # else the 'google/' prefix is stripped for tokenizer_def in self.tokenizers_list ] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = 'UNwant\u00E9d,running' _lowerCAmelCase : List[Any] = 'unwanted, running' return input_text, output_text def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = self.tokenizer_class(self.vocab_file ) _lowerCAmelCase : int = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(snake_case__ , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case__ ) , [9, 6, 7, 12, 10, 11] ) def a ( self ): '''simple docstring''' if not self.test_rust_tokenizer: return _lowerCAmelCase : int = self.get_tokenizer() _lowerCAmelCase : List[Any] = self.get_rust_tokenizer() _lowerCAmelCase : Any = 'UNwant\u00E9d,running' _lowerCAmelCase : List[Any] = tokenizer.tokenize(snake_case__ ) _lowerCAmelCase : Tuple = rust_tokenizer.tokenize(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : Tuple = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) _lowerCAmelCase : List[str] = rust_tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : int = self.get_rust_tokenizer() _lowerCAmelCase : List[Any] = tokenizer.encode(snake_case__ ) _lowerCAmelCase : Optional[Any] = rust_tokenizer.encode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) # With lower casing _lowerCAmelCase : int = self.get_tokenizer(do_lower_case=snake_case__ ) _lowerCAmelCase : Tuple = self.get_rust_tokenizer(do_lower_case=snake_case__ ) _lowerCAmelCase : List[str] = 'UNwant\u00E9d,running' _lowerCAmelCase : Dict = tokenizer.tokenize(snake_case__ ) _lowerCAmelCase : int = rust_tokenizer.tokenize(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : Optional[int] = tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) _lowerCAmelCase : Tuple = rust_tokenizer.encode(snake_case__ , add_special_tokens=snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : Any = self.get_rust_tokenizer() _lowerCAmelCase : Union[str, Any] = tokenizer.encode(snake_case__ ) _lowerCAmelCase : Any = rust_tokenizer.encode(snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('ah\u535A\u63A8zz' ) , ['ah', '\u535A', '\u63A8', 'zz'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = BasicTokenizer(do_lower_case=snake_case__ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['hello', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = BasicTokenizer(do_lower_case=snake_case__ , strip_accents=snake_case__ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hällo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['h\u00E9llo'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = BasicTokenizer(do_lower_case=snake_case__ , strip_accents=snake_case__ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = BasicTokenizer(do_lower_case=snake_case__ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['hallo', '!', 'how', 'are', 'you', '?'] ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['hello'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = BasicTokenizer(do_lower_case=snake_case__ ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = BasicTokenizer(do_lower_case=snake_case__ , strip_accents=snake_case__ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HäLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = BasicTokenizer(do_lower_case=snake_case__ , strip_accents=snake_case__ ) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ' ) , ['HaLLo', '!', 'how', 'Are', 'yoU', '?'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = BasicTokenizer(do_lower_case=snake_case__ , never_split=['[UNK]'] ) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? [UNK]' ) , ['HeLLo', '!', 'how', 'Are', 'yoU', '?', '[UNK]'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = ['[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing'] _lowerCAmelCase : List[str] = {} for i, token in enumerate(snake_case__ ): _lowerCAmelCase : List[Any] = i _lowerCAmelCase : Tuple = WordpieceTokenizer(vocab=snake_case__ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('unwanted running' ) , ['un', '##want', '##ed', 'runn', '##ing'] ) self.assertListEqual(tokenizer.tokenize('unwantedX running' ) , ['[UNK]', 'runn', '##ing'] ) def a ( self ): '''simple docstring''' self.assertTrue(_is_whitespace(' ' ) ) self.assertTrue(_is_whitespace('\t' ) ) self.assertTrue(_is_whitespace('\r' ) ) self.assertTrue(_is_whitespace('\n' ) ) self.assertTrue(_is_whitespace('\u00A0' ) ) self.assertFalse(_is_whitespace('A' ) ) self.assertFalse(_is_whitespace('-' ) ) def a ( self ): '''simple docstring''' self.assertTrue(_is_control('\u0005' ) ) self.assertFalse(_is_control('A' ) ) self.assertFalse(_is_control(' ' ) ) self.assertFalse(_is_control('\t' ) ) self.assertFalse(_is_control('\r' ) ) def a ( self ): '''simple docstring''' self.assertTrue(_is_punctuation('-' ) ) self.assertTrue(_is_punctuation('$' ) ) self.assertTrue(_is_punctuation('`' ) ) self.assertTrue(_is_punctuation('.' ) ) self.assertFalse(_is_punctuation('A' ) ) self.assertFalse(_is_punctuation(' ' ) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : str = self.get_tokenizer() _lowerCAmelCase : Optional[Any] = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(snake_case__ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) self.assertListEqual( [rust_tokenizer.tokenize(snake_case__ ) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']] ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.tokenizer_class.from_pretrained('google/mobilebert-uncased' ) _lowerCAmelCase : List[str] = tokenizer.encode('sequence builders' , add_special_tokens=snake_case__ ) _lowerCAmelCase : List[str] = tokenizer.encode('multi-sequence build' , add_special_tokens=snake_case__ ) _lowerCAmelCase : List[Any] = tokenizer.build_inputs_with_special_tokens(snake_case__ ) _lowerCAmelCase : int = tokenizer.build_inputs_with_special_tokens(snake_case__ , snake_case__ ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def a ( self ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): _lowerCAmelCase : List[str] = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) _lowerCAmelCase : Union[str, Any] = F'A, naïve {tokenizer_r.mask_token} AllenNLP sentence.' _lowerCAmelCase : str = tokenizer_r.encode_plus( snake_case__ , return_attention_mask=snake_case__ , return_token_type_ids=snake_case__ , return_offsets_mapping=snake_case__ , add_special_tokens=snake_case__ , ) _lowerCAmelCase : Dict = tokenizer_r.do_lower_case if hasattr(snake_case__ , 'do_lower_case' ) else False _lowerCAmelCase : Union[str, Any] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'A'), ((1, 2), ','), ((3, 5), 'na'), ((5, 6), '##ï'), ((6, 8), '##ve'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'Allen'), ((21, 23), '##NL'), ((23, 24), '##P'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'a'), ((1, 2), ','), ((3, 8), 'naive'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'allen'), ((21, 23), '##nl'), ((23, 24), '##p'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['input_ids'] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['offset_mapping'] ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = ['的', '人', '有'] _lowerCAmelCase : str = ''.join(snake_case__ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): _lowerCAmelCase : Dict = True _lowerCAmelCase : Tuple = self.tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) _lowerCAmelCase : Any = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) _lowerCAmelCase : Union[str, Any] = tokenizer_p.encode(snake_case__ , add_special_tokens=snake_case__ ) _lowerCAmelCase : Optional[int] = tokenizer_r.encode(snake_case__ , add_special_tokens=snake_case__ ) _lowerCAmelCase : Optional[int] = tokenizer_r.convert_ids_to_tokens(snake_case__ ) _lowerCAmelCase : str = tokenizer_p.convert_ids_to_tokens(snake_case__ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(snake_case__ , snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ ) _lowerCAmelCase : Optional[int] = False _lowerCAmelCase : Optional[int] = self.rust_tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) _lowerCAmelCase : List[str] = self.tokenizer_class.from_pretrained(snake_case__ , **snake_case__ ) _lowerCAmelCase : List[Any] = tokenizer_r.encode(snake_case__ , add_special_tokens=snake_case__ ) _lowerCAmelCase : Any = tokenizer_p.encode(snake_case__ , add_special_tokens=snake_case__ ) _lowerCAmelCase : Tuple = tokenizer_r.convert_ids_to_tokens(snake_case__ ) _lowerCAmelCase : Optional[int] = tokenizer_p.convert_ids_to_tokens(snake_case__ ) # it is expected that only the first Chinese character is not preceded by "##". _lowerCAmelCase : Any = [ F'##{token}' if idx != 0 else token for idx, token in enumerate(snake_case__ ) ] self.assertListEqual(snake_case__ , snake_case__ ) self.assertListEqual(snake_case__ , snake_case__ )
701
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCAmelCase : str = logging.get_logger(__name__) # General docstring lowerCAmelCase : Optional[Any] = """RegNetConfig""" # Base docstring lowerCAmelCase : int = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = [1, 10_88, 7, 7] # Image classification docstring lowerCAmelCase : Any = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = """tabby, tabby cat""" lowerCAmelCase : Tuple = [ """facebook/regnet-y-040""", # See all regnet models at https://huggingface.co/models?filter=regnet ] class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 3 , snake_case__ = 1 , snake_case__ = 1 , snake_case__ = "relu" , **snake_case__ , ): '''simple docstring''' super().__init__(**snake_case__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _lowerCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _lowerCAmelCase : List[Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=snake_case__ , strides=snake_case__ , padding='VALID' , groups=snake_case__ , use_bias=snake_case__ , name='convolution' , ) _lowerCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) _lowerCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.convolution(self.padding(snake_case__ ) ) _lowerCAmelCase : Union[str, Any] = self.normalization(snake_case__ ) _lowerCAmelCase : int = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = config.num_channels _lowerCAmelCase : List[Any] = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name='embedder' , ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = shape_list(snake_case__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( 'Make sure that the channel dimension of the pixel values match with the one set in the configuration.' ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _lowerCAmelCase : List[Any] = tf.transpose(snake_case__ , perm=(0, 2, 3, 1) ) _lowerCAmelCase : Tuple = self.embedder(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=1 , strides=snake_case__ , use_bias=snake_case__ , name='convolution' ) _lowerCAmelCase : Tuple = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' return self.normalization(self.convolution(snake_case__ ) , training=snake_case__ ) class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Any = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) _lowerCAmelCase : str = [ tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='relu' , name='attention.0' ), tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='sigmoid' , name='attention.2' ), ] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.pooler(snake_case__ ) for layer_module in self.attention: _lowerCAmelCase : Tuple = layer_module(snake_case__ ) _lowerCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Optional[int] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _lowerCAmelCase : Any = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.2' ), ] _lowerCAmelCase : List[str] = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = hidden_state for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) _lowerCAmelCase : int = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : Tuple = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : List[str] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) _lowerCAmelCase : Tuple = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetSELayer(snake_case__ , reduced_channels=int(round(in_channels / 4 ) ) , name='layer.2' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.3' ), ] _lowerCAmelCase : Tuple = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = hidden_state for layer_module in self.layers: _lowerCAmelCase : List[Any] = layer_module(snake_case__ ) _lowerCAmelCase : Tuple = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : str = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 2 , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Dict = TFRegNetXLayer if config.layer_type == 'x' else TFRegNetYLayer _lowerCAmelCase : Optional[int] = [ # downsampling is done in the first layer with stride of 2 layer(snake_case__ , snake_case__ , snake_case__ , stride=snake_case__ , name='layers.0' ), *[layer(snake_case__ , snake_case__ , snake_case__ , name=F'layers.{i+1}' ) for i in range(depth - 1 )], ] def a ( self , snake_case__ ): '''simple docstring''' for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( snake_case__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name='stages.0' , ) ) _lowerCAmelCase : Union[str, Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(snake_case__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(snake_case__ , snake_case__ , snake_case__ , depth=snake_case__ , name=F'stages.{i+1}' ) ) def a ( self , snake_case__ , snake_case__ = False , snake_case__ = True ): '''simple docstring''' _lowerCAmelCase : List[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _lowerCAmelCase : str = hidden_states + (hidden_state,) _lowerCAmelCase : List[str] = stage_module(snake_case__ ) if output_hidden_states: _lowerCAmelCase : Dict = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=snake_case__ , hidden_states=snake_case__ ) @keras_serializable class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" __magic_name__ = RegNetConfig def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = config _lowerCAmelCase : Union[str, Any] = TFRegNetEmbeddings(snake_case__ , name='embedder' ) _lowerCAmelCase : Optional[int] = TFRegNetEncoder(snake_case__ , name='encoder' ) _lowerCAmelCase : Dict = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) @unpack_inputs def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' _lowerCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : int = self.embedder(snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[str] = self.encoder( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[Any] = encoder_outputs[0] _lowerCAmelCase : Tuple = self.pooler(snake_case__ ) # Change to NCHW output format have uniformity in the modules _lowerCAmelCase : Optional[int] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) _lowerCAmelCase : Optional[Any] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _lowerCAmelCase : Union[str, Any] = tuple([tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=snake_case__ , pooler_output=snake_case__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = RegNetConfig __magic_name__ = "regnet" __magic_name__ = "pixel_values" @property def a ( self ): '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 224, 224) , dtype=tf.floataa )} lowerCAmelCase : List[Any] = r""" Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. """ lowerCAmelCase : Dict = r""" Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : List[str] = TFRegNetMainLayer(snake_case__ , name='regnet' ) @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Any = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : str = self.regnet( pixel_values=snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[Any] = config.num_labels _lowerCAmelCase : Optional[Any] = TFRegNetMainLayer(snake_case__ , name='regnet' ) # classification head _lowerCAmelCase : Optional[int] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name='classifier.1' ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def a ( self , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : Dict = self.regnet( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.pooler_output if return_dict else outputs[1] _lowerCAmelCase : List[Any] = self.classifier[0](snake_case__ ) _lowerCAmelCase : Tuple = self.classifier[1](snake_case__ ) _lowerCAmelCase : int = None if labels is None else self.hf_compute_loss(labels=snake_case__ , logits=snake_case__ ) if not return_dict: _lowerCAmelCase : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=snake_case__ , logits=snake_case__ , hidden_states=outputs.hidden_states )
630
0
'''simple docstring''' import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = (DDPMScheduler,) def a ( self , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**snake_case__ ) return config def a ( self ): '''simple docstring''' for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=snake_case__ ) def a ( self ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=snake_case__ , beta_end=snake_case__ ) def a ( self ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=snake_case__ ) def a ( self ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=snake_case__ ) def a ( self ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=snake_case__ ) def a ( self ): '''simple docstring''' self.check_over_configs(thresholding=snake_case__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=snake_case__ , prediction_type=snake_case__ , sample_max_value=snake_case__ , ) def a ( self ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=snake_case__ ) def a ( self ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1E-5 def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : str = self.dummy_model() _lowerCAmelCase : str = self.dummy_sample_deter _lowerCAmelCase : Union[str, Any] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : List[Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Any = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Dict = pred_prev_sample _lowerCAmelCase : Dict = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : List[str] = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 258.9606 ) < 1E-2 assert abs(result_mean.item() - 0.3372 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.scheduler_classes[0] _lowerCAmelCase : Any = self.get_scheduler_config(prediction_type='v_prediction' ) _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : Any = self.dummy_model() _lowerCAmelCase : Tuple = self.dummy_sample_deter _lowerCAmelCase : Optional[int] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : Union[str, Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Dict = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Tuple = pred_prev_sample _lowerCAmelCase : Any = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : Dict = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 202.0296 ) < 1E-2 assert abs(result_mean.item() - 0.2631 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[int] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=snake_case__ ) _lowerCAmelCase : Union[str, Any] = scheduler.timesteps for i, timestep in enumerate(snake_case__ ): if i == len(snake_case__ ) - 1: _lowerCAmelCase : str = -1 else: _lowerCAmelCase : Optional[Any] = timesteps[i + 1] _lowerCAmelCase : int = scheduler.previous_timestep(snake_case__ ) _lowerCAmelCase : int = prev_t.item() self.assertEqual(snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : Tuple = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 51, 0] with self.assertRaises(snake_case__ , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : List[str] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 1, 0] _lowerCAmelCase : int = len(snake_case__ ) with self.assertRaises(snake_case__ , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=snake_case__ , timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : int = self.get_scheduler_config() _lowerCAmelCase : Any = scheduler_class(**snake_case__ ) _lowerCAmelCase : Any = [scheduler.config.num_train_timesteps] with self.assertRaises( snake_case__ , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=snake_case__ )
702
'''simple docstring''' from typing import Any def lowercase (_A ): """simple docstring""" if not input_list: return [] _lowerCAmelCase : Optional[int] = [input_list.count(_A ) for value in input_list] _lowerCAmelCase : int = max(_A ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(_A ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
630
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : List[str] = logging.get_logger(__name__) lowerCAmelCase : Optional[Any] = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "openai-gpt" __magic_name__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , snake_case__=4_0478 , snake_case__=512 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=1E-5 , snake_case__=0.02 , snake_case__="cls_index" , snake_case__=True , snake_case__=None , snake_case__=True , snake_case__=0.1 , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : List[Any] = vocab_size _lowerCAmelCase : Any = n_positions _lowerCAmelCase : Any = n_embd _lowerCAmelCase : Any = n_layer _lowerCAmelCase : Tuple = n_head _lowerCAmelCase : Optional[int] = afn _lowerCAmelCase : Any = resid_pdrop _lowerCAmelCase : Optional[int] = embd_pdrop _lowerCAmelCase : Tuple = attn_pdrop _lowerCAmelCase : Optional[Any] = layer_norm_epsilon _lowerCAmelCase : Any = initializer_range _lowerCAmelCase : int = summary_type _lowerCAmelCase : Dict = summary_use_proj _lowerCAmelCase : int = summary_activation _lowerCAmelCase : Union[str, Any] = summary_first_dropout _lowerCAmelCase : List[Any] = summary_proj_to_labels super().__init__(**snake_case__ )
703
'''simple docstring''' from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
630
0
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import TransformeraDModel, VQDiffusionPipeline, VQDiffusionScheduler, VQModel from diffusers.pipelines.vq_diffusion.pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings from diffusers.utils import load_numpy, slow, torch_device from diffusers.utils.testing_utils import require_torch_gpu lowerCAmelCase : str = False class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def a ( self ): '''simple docstring''' return 12 @property def a ( self ): '''simple docstring''' return 12 @property def a ( self ): '''simple docstring''' return 32 @property def a ( self ): '''simple docstring''' torch.manual_seed(0 ) _lowerCAmelCase : int = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=3 , num_vq_embeddings=self.num_embed , vq_embed_dim=3 , ) return model @property def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def a ( self ): '''simple docstring''' torch.manual_seed(0 ) _lowerCAmelCase : int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModel(snake_case__ ) @property def a ( self ): '''simple docstring''' torch.manual_seed(0 ) _lowerCAmelCase : Any = 12 _lowerCAmelCase : Any = 12 _lowerCAmelCase : Dict = { 'attention_bias': True, 'cross_attention_dim': 32, 'attention_head_dim': height * width, 'num_attention_heads': 1, 'num_vector_embeds': self.num_embed, 'num_embeds_ada_norm': self.num_embeds_ada_norm, 'norm_num_groups': 32, 'sample_size': width, 'activation_fn': 'geglu-approximate', } _lowerCAmelCase : str = TransformeraDModel(**snake_case__ ) return model def a ( self ): '''simple docstring''' _lowerCAmelCase : int = 'cpu' _lowerCAmelCase : int = self.dummy_vqvae _lowerCAmelCase : Tuple = self.dummy_text_encoder _lowerCAmelCase : List[str] = self.dummy_tokenizer _lowerCAmelCase : Tuple = self.dummy_transformer _lowerCAmelCase : Any = VQDiffusionScheduler(self.num_embed ) _lowerCAmelCase : str = LearnedClassifierFreeSamplingEmbeddings(learnable=snake_case__ ) _lowerCAmelCase : Tuple = VQDiffusionPipeline( vqvae=snake_case__ , text_encoder=snake_case__ , tokenizer=snake_case__ , transformer=snake_case__ , scheduler=snake_case__ , learned_classifier_free_sampling_embeddings=snake_case__ , ) _lowerCAmelCase : Dict = pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : List[str] = 'teddy bear playing in the pool' _lowerCAmelCase : List[str] = torch.Generator(device=snake_case__ ).manual_seed(0 ) _lowerCAmelCase : List[str] = pipe([prompt] , generator=snake_case__ , num_inference_steps=2 , output_type='np' ) _lowerCAmelCase : str = output.images _lowerCAmelCase : Union[str, Any] = torch.Generator(device=snake_case__ ).manual_seed(0 ) _lowerCAmelCase : Union[str, Any] = pipe( [prompt] , generator=snake_case__ , output_type='np' , return_dict=snake_case__ , num_inference_steps=2 )[0] _lowerCAmelCase : int = image[0, -3:, -3:, -1] _lowerCAmelCase : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) _lowerCAmelCase : Tuple = np.array([0.6551, 0.6168, 0.5008, 0.5676, 0.5659, 0.4295, 0.6073, 0.5599, 0.4992] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = 'cpu' _lowerCAmelCase : Tuple = self.dummy_vqvae _lowerCAmelCase : List[Any] = self.dummy_text_encoder _lowerCAmelCase : Optional[int] = self.dummy_tokenizer _lowerCAmelCase : Any = self.dummy_transformer _lowerCAmelCase : Dict = VQDiffusionScheduler(self.num_embed ) _lowerCAmelCase : Tuple = LearnedClassifierFreeSamplingEmbeddings( learnable=snake_case__ , hidden_size=self.text_embedder_hidden_size , length=tokenizer.model_max_length ) _lowerCAmelCase : Tuple = VQDiffusionPipeline( vqvae=snake_case__ , text_encoder=snake_case__ , tokenizer=snake_case__ , transformer=snake_case__ , scheduler=snake_case__ , learned_classifier_free_sampling_embeddings=snake_case__ , ) _lowerCAmelCase : Optional[Any] = pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : Union[str, Any] = 'teddy bear playing in the pool' _lowerCAmelCase : Union[str, Any] = torch.Generator(device=snake_case__ ).manual_seed(0 ) _lowerCAmelCase : str = pipe([prompt] , generator=snake_case__ , num_inference_steps=2 , output_type='np' ) _lowerCAmelCase : int = output.images _lowerCAmelCase : Dict = torch.Generator(device=snake_case__ ).manual_seed(0 ) _lowerCAmelCase : Optional[Any] = pipe( [prompt] , generator=snake_case__ , output_type='np' , return_dict=snake_case__ , num_inference_steps=2 )[0] _lowerCAmelCase : str = image[0, -3:, -3:, -1] _lowerCAmelCase : Optional[int] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) _lowerCAmelCase : List[str] = np.array([0.6693, 0.6075, 0.4959, 0.5701, 0.5583, 0.4333, 0.6171, 0.5684, 0.4988] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2.0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/vq_diffusion/teddy_bear_pool_classifier_free_sampling.npy' ) _lowerCAmelCase : Tuple = VQDiffusionPipeline.from_pretrained('microsoft/vq-diffusion-ithq' ) _lowerCAmelCase : Any = pipeline.to(snake_case__ ) pipeline.set_progress_bar_config(disable=snake_case__ ) # requires GPU generator for gumbel softmax # don't use GPU generator in tests though _lowerCAmelCase : Optional[int] = torch.Generator(device=snake_case__ ).manual_seed(0 ) _lowerCAmelCase : str = pipeline( 'teddy bear playing in the pool' , num_images_per_prompt=1 , generator=snake_case__ , output_type='np' , ) _lowerCAmelCase : int = output.images[0] assert image.shape == (256, 256, 3) assert np.abs(expected_image - image ).max() < 2.0
704
'''simple docstring''' import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase : int = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = {"""vocab_file""": """spiece.model"""} lowerCAmelCase : Optional[int] = { """vocab_file""": { """AI-Sweden/gpt-sw3-126m""": """https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-350m""": """https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-1.6b""": """https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-6.7b""": """https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-20b""": """https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model""", } } lowerCAmelCase : Union[str, Any] = { """AI-Sweden/gpt-sw3-126m""": 20_48, """AI-Sweden/gpt-sw3-350m""": 20_48, """AI-Sweden/gpt-sw3-1.6b""": 20_48, """AI-Sweden/gpt-sw3-6.7b""": 20_48, """AI-Sweden/gpt-sw3-20b""": 20_48, } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = ["input_ids", "attention_mask"] def __init__( self , snake_case__ , snake_case__=False , snake_case__=False , snake_case__=False , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Tuple = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase : List[Any] = kwargs.get('name_or_path' ) if name_or_path is None: logger.warning( 'name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,' ' you are testing the model, this can safely be ignored' ) _lowerCAmelCase : Any = 'None' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing _lowerCAmelCase : str = '<|endoftext|>' if eos_token is None else eos_token _lowerCAmelCase : Tuple = '<unk>' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: _lowerCAmelCase : List[str] = unk_token if pad_token is None else pad_token _lowerCAmelCase : Optional[int] = eos_token if bos_token is None else bos_token else: _lowerCAmelCase : Tuple = '<pad>' if pad_token is None else pad_token _lowerCAmelCase : Union[str, Any] = '<s>' if bos_token is None else bos_token super().__init__( do_lower_case=snake_case__ , remove_space=snake_case__ , keep_accents=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , pad_token=snake_case__ , sp_model_kwargs=self.sp_model_kwargs , **snake_case__ , ) _lowerCAmelCase : Union[str, Any] = do_lower_case _lowerCAmelCase : Optional[int] = remove_space _lowerCAmelCase : Any = keep_accents _lowerCAmelCase : Optional[int] = vocab_file _lowerCAmelCase : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case__ ) # Used for whitespace normalization in input texts # fmt : off _lowerCAmelCase : Optional[Any] = {' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', '', '„'} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing _lowerCAmelCase : Optional[Any] = re.compile( F'[{"".join(map(snake_case__ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]' ) def __getstate__( self ): '''simple docstring''' _lowerCAmelCase : List[str] = self.__dict__.copy() _lowerCAmelCase : int = None return state def __setstate__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowerCAmelCase : int = {} _lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def a ( self ): '''simple docstring''' return len(self.sp_model ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.non_printing_characters_re.sub('' , snake_case__ ) # Normalize whitespaces _lowerCAmelCase : Tuple = ''.join([char if char not in self.whitespaces else ' ' for char in text] ) # NFC Unicode normalization _lowerCAmelCase : Union[str, Any] = unicodedata.normalize('NFC' , snake_case__ ) return text def a ( self , snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : str = self.preprocess_text(snake_case__ ) return self.sp_model.encode(snake_case__ , out_type=snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.PieceToId(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case__ ) @staticmethod def a ( snake_case__ ): '''simple docstring''' return out_string def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = [] _lowerCAmelCase : Optional[Any] = '' _lowerCAmelCase : Tuple = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(snake_case__ ) + token _lowerCAmelCase : Union[str, Any] = True _lowerCAmelCase : List[Any] = [] else: current_sub_tokens.append(snake_case__ ) _lowerCAmelCase : List[Any] = False out_string += self.sp_model.decode(snake_case__ ) return out_string def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = {self.convert_ids_to_tokens(snake_case__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCAmelCase : int = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case__ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case__ , 'wb' ) as fi: _lowerCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (out_vocab_file,) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' if isinstance(snake_case__ , snake_case__ ): _lowerCAmelCase : Optional[Any] = self.preprocess_text(snake_case__ ) _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) else: _lowerCAmelCase : Tuple = [self.preprocess_text(snake_case__ ) for t in text] _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) if return_tensors is True or return_tensors == "pt": _lowerCAmelCase : int = torch.tensor(snake_case__ ) return token_ids def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.decode(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = [F'User: {text}' if is_user else F'Bot: {text}' for is_user, text in conversation.iter_texts()] _lowerCAmelCase : str = ( F'{self.eos_token}{self.bos_token}' + F'{self.bos_token}'.join(snake_case__ ) + F'{self.bos_token}Bot:' ) return self.encode(text=snake_case__ )
630
0
from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL lowerCAmelCase : Optional[int] = logging.get_logger(__name__) def lowercase (_A ): """simple docstring""" if isinstance(_A , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(_A , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(_A ): return [[videos]] raise ValueError(f'Could not make batched video from {videos}' ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = ["pixel_values"] def __init__( self , snake_case__ = True , snake_case__ = None , snake_case__ = PILImageResampling.BILINEAR , snake_case__ = True , snake_case__ = None , snake_case__ = True , snake_case__ = 1 / 255 , snake_case__ = True , snake_case__ = True , snake_case__ = None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Optional[Any] = size if size is not None else {'shortest_edge': 256} _lowerCAmelCase : str = get_size_dict(snake_case__ , default_to_square=snake_case__ ) _lowerCAmelCase : Union[str, Any] = crop_size if crop_size is not None else {'height': 224, 'width': 224} _lowerCAmelCase : Any = get_size_dict(snake_case__ , param_name='crop_size' ) _lowerCAmelCase : Tuple = do_resize _lowerCAmelCase : int = size _lowerCAmelCase : Optional[Any] = do_center_crop _lowerCAmelCase : Tuple = crop_size _lowerCAmelCase : Optional[int] = resample _lowerCAmelCase : Optional[Any] = do_rescale _lowerCAmelCase : Optional[Any] = rescale_factor _lowerCAmelCase : str = offset _lowerCAmelCase : Tuple = do_normalize _lowerCAmelCase : List[str] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _lowerCAmelCase : Optional[int] = image_std if image_std is not None else IMAGENET_STANDARD_STD def a ( self , snake_case__ , snake_case__ , snake_case__ = PILImageResampling.BILINEAR , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Any = get_size_dict(snake_case__ , default_to_square=snake_case__ ) if "shortest_edge" in size: _lowerCAmelCase : List[Any] = get_resize_output_image_size(snake_case__ , size['shortest_edge'] , default_to_square=snake_case__ ) elif "height" in size and "width" in size: _lowerCAmelCase : Union[str, Any] = (size['height'], size['width']) else: raise ValueError(F'Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}' ) return resize(snake_case__ , size=snake_case__ , resample=snake_case__ , data_format=snake_case__ , **snake_case__ ) def a ( self , snake_case__ , snake_case__ , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Any = get_size_dict(snake_case__ ) if "height" not in size or "width" not in size: raise ValueError(F'Size must have \'height\' and \'width\' as keys. Got {size.keys()}' ) return center_crop(snake_case__ , size=(size['height'], size['width']) , data_format=snake_case__ , **snake_case__ ) def a ( self , snake_case__ , snake_case__ , snake_case__ = True , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : int = image.astype(np.floataa ) if offset: _lowerCAmelCase : Union[str, Any] = image - (scale / 2) return rescale(snake_case__ , scale=snake_case__ , data_format=snake_case__ , **snake_case__ ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = None , **snake_case__ , ): '''simple docstring''' return normalize(snake_case__ , mean=snake_case__ , std=snake_case__ , data_format=snake_case__ , **snake_case__ ) def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = ChannelDimension.FIRST , ): '''simple docstring''' if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) if offset and not do_rescale: raise ValueError('For offset, do_rescale must also be set to True.' ) # All transformations expect numpy arrays. _lowerCAmelCase : Dict = to_numpy_array(snake_case__ ) if do_resize: _lowerCAmelCase : Dict = self.resize(image=snake_case__ , size=snake_case__ , resample=snake_case__ ) if do_center_crop: _lowerCAmelCase : Union[str, Any] = self.center_crop(snake_case__ , size=snake_case__ ) if do_rescale: _lowerCAmelCase : Union[str, Any] = self.rescale(image=snake_case__ , scale=snake_case__ , offset=snake_case__ ) if do_normalize: _lowerCAmelCase : int = self.normalize(image=snake_case__ , mean=snake_case__ , std=snake_case__ ) _lowerCAmelCase : List[str] = to_channel_dimension_format(snake_case__ , snake_case__ ) return image def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = ChannelDimension.FIRST , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Any = do_resize if do_resize is not None else self.do_resize _lowerCAmelCase : Any = resample if resample is not None else self.resample _lowerCAmelCase : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop _lowerCAmelCase : str = do_rescale if do_rescale is not None else self.do_rescale _lowerCAmelCase : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor _lowerCAmelCase : str = offset if offset is not None else self.offset _lowerCAmelCase : Any = do_normalize if do_normalize is not None else self.do_normalize _lowerCAmelCase : Union[str, Any] = image_mean if image_mean is not None else self.image_mean _lowerCAmelCase : List[str] = image_std if image_std is not None else self.image_std _lowerCAmelCase : Tuple = size if size is not None else self.size _lowerCAmelCase : Union[str, Any] = get_size_dict(snake_case__ , default_to_square=snake_case__ ) _lowerCAmelCase : str = crop_size if crop_size is not None else self.crop_size _lowerCAmelCase : List[str] = get_size_dict(snake_case__ , param_name='crop_size' ) if not valid_images(snake_case__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) _lowerCAmelCase : int = make_batched(snake_case__ ) _lowerCAmelCase : Optional[int] = [ [ self._preprocess_image( image=snake_case__ , do_resize=snake_case__ , size=snake_case__ , resample=snake_case__ , do_center_crop=snake_case__ , crop_size=snake_case__ , do_rescale=snake_case__ , rescale_factor=snake_case__ , offset=snake_case__ , do_normalize=snake_case__ , image_mean=snake_case__ , image_std=snake_case__ , data_format=snake_case__ , ) for img in video ] for video in videos ] _lowerCAmelCase : Any = {'pixel_values': videos} return BatchFeature(data=snake_case__ , tensor_type=snake_case__ )
705
'''simple docstring''' import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = (DDPMScheduler,) def a ( self , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**snake_case__ ) return config def a ( self ): '''simple docstring''' for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=snake_case__ ) def a ( self ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=snake_case__ , beta_end=snake_case__ ) def a ( self ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=snake_case__ ) def a ( self ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=snake_case__ ) def a ( self ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=snake_case__ ) def a ( self ): '''simple docstring''' self.check_over_configs(thresholding=snake_case__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=snake_case__ , prediction_type=snake_case__ , sample_max_value=snake_case__ , ) def a ( self ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=snake_case__ ) def a ( self ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1E-5 def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : str = self.dummy_model() _lowerCAmelCase : str = self.dummy_sample_deter _lowerCAmelCase : Union[str, Any] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : List[Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Any = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Dict = pred_prev_sample _lowerCAmelCase : Dict = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : List[str] = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 258.9606 ) < 1E-2 assert abs(result_mean.item() - 0.3372 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.scheduler_classes[0] _lowerCAmelCase : Any = self.get_scheduler_config(prediction_type='v_prediction' ) _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : Any = self.dummy_model() _lowerCAmelCase : Tuple = self.dummy_sample_deter _lowerCAmelCase : Optional[int] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : Union[str, Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Dict = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Tuple = pred_prev_sample _lowerCAmelCase : Any = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : Dict = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 202.0296 ) < 1E-2 assert abs(result_mean.item() - 0.2631 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[int] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=snake_case__ ) _lowerCAmelCase : Union[str, Any] = scheduler.timesteps for i, timestep in enumerate(snake_case__ ): if i == len(snake_case__ ) - 1: _lowerCAmelCase : str = -1 else: _lowerCAmelCase : Optional[Any] = timesteps[i + 1] _lowerCAmelCase : int = scheduler.previous_timestep(snake_case__ ) _lowerCAmelCase : int = prev_t.item() self.assertEqual(snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : Tuple = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 51, 0] with self.assertRaises(snake_case__ , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : List[str] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 1, 0] _lowerCAmelCase : int = len(snake_case__ ) with self.assertRaises(snake_case__ , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=snake_case__ , timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : int = self.get_scheduler_config() _lowerCAmelCase : Any = scheduler_class(**snake_case__ ) _lowerCAmelCase : Any = [scheduler.config.num_train_timesteps] with self.assertRaises( snake_case__ , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=snake_case__ )
630
0
'''simple docstring''' from __future__ import annotations from typing import TypedDict class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = 4_2 __magic_name__ = 4_2 def lowercase (_A ): """simple docstring""" if not isinstance(_A , _A ): raise TypeError('The parameter s type must be str.' ) return [s[i:] + s[:i] for i in range(len(_A ) )] def lowercase (_A ): """simple docstring""" if not isinstance(_A , _A ): raise TypeError('The parameter s type must be str.' ) if not s: raise ValueError('The parameter s must not be empty.' ) _lowerCAmelCase : int = all_rotations(_A ) rotations.sort() # sort the list of rotations in alphabetically order # make a string composed of the last char of each rotation _lowerCAmelCase : BWTTransformDict = { "bwt_string": "".join([word[-1] for word in rotations] ), "idx_original_string": rotations.index(_A ), } return response def lowercase (_A , _A ): """simple docstring""" if not isinstance(_A , _A ): raise TypeError('The parameter bwt_string type must be str.' ) if not bwt_string: raise ValueError('The parameter bwt_string must not be empty.' ) try: _lowerCAmelCase : Any = int(_A ) except ValueError: raise TypeError( 'The parameter idx_original_string type must be int or passive' ' of cast to int.' ) if idx_original_string < 0: raise ValueError('The parameter idx_original_string must not be lower than 0.' ) if idx_original_string >= len(_A ): raise ValueError( 'The parameter idx_original_string must be lower than' ' len(bwt_string).' ) _lowerCAmelCase : Optional[int] = [''] * len(_A ) for _ in range(len(_A ) ): for i in range(len(_A ) ): _lowerCAmelCase : Optional[Any] = bwt_string[i] + ordered_rotations[i] ordered_rotations.sort() return ordered_rotations[idx_original_string] if __name__ == "__main__": lowerCAmelCase : Optional[int] = """Provide a string that I will generate its BWT transform: """ lowerCAmelCase : str = input(entry_msg).strip() lowerCAmelCase : Tuple = bwt_transform(s) print( F'''Burrows Wheeler transform for string \'{s}\' results ''' F'''in \'{result["bwt_string"]}\'''' ) lowerCAmelCase : Optional[Any] = reverse_bwt(result["""bwt_string"""], result["""idx_original_string"""]) print( F'''Reversing Burrows Wheeler transform for entry \'{result["bwt_string"]}\' ''' F'''we get original string \'{original_string}\'''' )
706
'''simple docstring''' import socket def lowercase (): """simple docstring""" _lowerCAmelCase : Tuple = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCAmelCase : Optional[int] = socket.gethostname() _lowerCAmelCase : Any = 1_2_3_1_2 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCAmelCase : Union[str, Any] = sock.recv(1_0_2_4 ) if not data: break out_file.write(_A ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
630
0
'''simple docstring''' import fire from utils import calculate_rouge, save_json def lowercase (_A , _A , _A=None , **_A ): """simple docstring""" _lowerCAmelCase : Tuple = [x.strip() for x in open(_A ).readlines()] _lowerCAmelCase : List[Any] = [x.strip() for x in open(_A ).readlines()][: len(_A )] _lowerCAmelCase : Optional[Any] = calculate_rouge(_A , _A , **_A ) if save_path is not None: save_json(_A , _A , indent=_A ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
707
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase : Tuple = False lowerCAmelCase : str = True lowerCAmelCase : List[Any] = False if __name__ == "__main__": lowerCAmelCase : Any = argparse.ArgumentParser() parser.add_argument( """--repo_path""", default=None, type=str, required=True, help="""The config json file corresponding to the architecture.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") lowerCAmelCase : Optional[int] = parser.parse_args() lowerCAmelCase : int = { """image_size""": """sample_size""", """num_res_blocks""": """layers_per_block""", """block_channels""": """block_out_channels""", """down_blocks""": """down_block_types""", """up_blocks""": """up_block_types""", """downscale_freq_shift""": """freq_shift""", """resnet_num_groups""": """norm_num_groups""", """resnet_act_fn""": """act_fn""", """resnet_eps""": """norm_eps""", """num_head_channels""": """attention_head_dim""", } lowerCAmelCase : int = { """time_steps""": """time_proj""", """mid""": """mid_block""", """downsample_blocks""": """down_blocks""", """upsample_blocks""": """up_blocks""", } lowerCAmelCase : Optional[Any] = """""" if has_file(args.repo_path, """config.json""") else """unet""" with open(os.path.join(args.repo_path, subfolder, """config.json"""), """r""", encoding="""utf-8""") as reader: lowerCAmelCase : int = reader.read() lowerCAmelCase : List[str] = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, """config.json"""): lowerCAmelCase : str = UNetaDModel(**config) else: lowerCAmelCase : Union[str, Any] = UNetaDConditionModel if """ldm-text2im-large-256""" in args.repo_path else UNetaDModel lowerCAmelCase : Dict = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase : Union[str, Any] = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase : str = config[key] del config[key] lowerCAmelCase : Optional[int] = [k.replace("""UNetRes""", """""") for k in config["""down_block_types"""]] lowerCAmelCase : Dict = [k.replace("""UNetRes""", """""") for k in config["""up_block_types"""]] if do_only_weights: lowerCAmelCase : Tuple = torch.load(os.path.join(args.repo_path, subfolder, """diffusion_pytorch_model.bin""")) lowerCAmelCase : str = {} for param_key, param_value in state_dict.items(): if param_key.endswith(""".op.bias""") or param_key.endswith(""".op.weight"""): continue lowerCAmelCase : str = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split(""".""")[0] == key: lowerCAmelCase : Dict = param_value lowerCAmelCase : Tuple = True if not has_changed: lowerCAmelCase : Tuple = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
630
0
'''simple docstring''' import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = (UnCLIPScheduler,) def a ( self , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = { 'num_train_timesteps': 1000, 'variance_type': 'fixed_small_log', 'clip_sample': True, 'clip_sample_range': 1.0, 'prediction_type': 'epsilon', } config.update(**snake_case__ ) return config def a ( self ): '''simple docstring''' for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=snake_case__ ) def a ( self ): '''simple docstring''' for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=snake_case__ ) def a ( self ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=snake_case__ ) def a ( self ): '''simple docstring''' for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=snake_case__ ) def a ( self ): '''simple docstring''' for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=snake_case__ ) def a ( self ): '''simple docstring''' for time_step in [0, 500, 999]: for prev_timestep in [None, 5, 100, 250, 500, 750]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=snake_case__ , prev_timestep=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = self.scheduler_classes[0] _lowerCAmelCase : List[Any] = self.get_scheduler_config(variance_type='fixed_small_log' ) _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0000E-10 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.054_9625 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.999_4987 ) ) < 1E-5 def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[int] = self.get_scheduler_config(variance_type='learned_range' ) _lowerCAmelCase : int = scheduler_class(**snake_case__ ) _lowerCAmelCase : int = 0.5 assert scheduler._get_variance(1 , predicted_variance=snake_case__ ) - -10.171_2790 < 1E-5 assert scheduler._get_variance(487 , predicted_variance=snake_case__ ) - -5.799_8052 < 1E-5 assert scheduler._get_variance(999 , predicted_variance=snake_case__ ) - -0.001_0011 < 1E-5 def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.scheduler_classes[0] _lowerCAmelCase : int = self.get_scheduler_config() _lowerCAmelCase : str = scheduler_class(**snake_case__ ) _lowerCAmelCase : List[Any] = scheduler.timesteps _lowerCAmelCase : List[str] = self.dummy_model() _lowerCAmelCase : int = self.dummy_sample_deter _lowerCAmelCase : Dict = torch.manual_seed(0 ) for i, t in enumerate(snake_case__ ): # 1. predict noise residual _lowerCAmelCase : List[str] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : List[str] = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample _lowerCAmelCase : int = pred_prev_sample _lowerCAmelCase : List[str] = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : Dict = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 252.268_2495 ) < 1E-2 assert abs(result_mean.item() - 0.328_4743 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[str] = scheduler_class(**snake_case__ ) scheduler.set_timesteps(25 ) _lowerCAmelCase : List[str] = scheduler.timesteps _lowerCAmelCase : List[str] = self.dummy_model() _lowerCAmelCase : List[Any] = self.dummy_sample_deter _lowerCAmelCase : Dict = torch.manual_seed(0 ) for i, t in enumerate(snake_case__ ): # 1. predict noise residual _lowerCAmelCase : Optional[Any] = model(snake_case__ , snake_case__ ) if i + 1 == timesteps.shape[0]: _lowerCAmelCase : List[Any] = None else: _lowerCAmelCase : str = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Optional[int] = scheduler.step( snake_case__ , snake_case__ , snake_case__ , prev_timestep=snake_case__ , generator=snake_case__ ).prev_sample _lowerCAmelCase : Any = pred_prev_sample _lowerCAmelCase : Union[str, Any] = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : str = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 258.204_4983 ) < 1E-2 assert abs(result_mean.item() - 0.336_2038 ) < 1E-3 def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' pass
708
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = None ): '''simple docstring''' super().__init__() _lowerCAmelCase : Union[str, Any] = pad_token_id _lowerCAmelCase : List[Any] = max_length _lowerCAmelCase : Tuple = vocab _lowerCAmelCase : str = merges _lowerCAmelCase : List[str] = BytePairTokenizer(snake_case__ , snake_case__ , sequence_length=snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = [' '.join(snake_case__ ) for m in tokenizer.bpe_ranks.keys()] _lowerCAmelCase : Any = tokenizer.get_vocab() return cls(snake_case__ , snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = GPTaTokenizer.from_pretrained(snake_case__ , *snake_case__ , **snake_case__ ) return cls.from_tokenizer(snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ ): '''simple docstring''' return cls(**snake_case__ ) def a ( self ): '''simple docstring''' return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : str = self.tf_tokenizer(snake_case__ ) _lowerCAmelCase : str = tf.ones_like(snake_case__ ) if self.pad_token_id is not None: # pad the tokens up to max length _lowerCAmelCase : Optional[int] = max_length if max_length is not None else self.max_length if max_length is not None: _lowerCAmelCase , _lowerCAmelCase : str = pad_model_inputs( snake_case__ , max_seq_length=snake_case__ , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
630
0
'''simple docstring''' import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase : Optional[int] = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = {"""vocab_file""": """vocab.txt""", """emoji_file""": """emoji.json"""} lowerCAmelCase : Tuple = { """vocab_file""": { """abeja/gpt-neox-japanese-2.7b""": """https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt""", }, """emoji_file""": { """abeja/gpt-neox-japanese-2.7b""": """https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json""", }, } lowerCAmelCase : Union[str, Any] = { """abeja/gpt-neox-japanese-2.7b""": 20_48, } def lowercase (_A , _A ): """simple docstring""" with open(_A , 'r' , encoding='utf-8' ) as f: _lowerCAmelCase : int = json.loads(f.read() ) _lowerCAmelCase : List[str] = collections.OrderedDict() _lowerCAmelCase : List[str] = collections.OrderedDict() _lowerCAmelCase : Dict = collections.OrderedDict() with open(_A , 'r' , encoding='utf-8' ) as f: _lowerCAmelCase : Any = f.readlines() _lowerCAmelCase : List[str] = [[t.rstrip('\n' )] if (t == ',' or ',' not in t) else t.rstrip('\n' ).split(',' ) for t in token] for idx, b in enumerate(_A ): _lowerCAmelCase : int = b _lowerCAmelCase : Union[str, Any] = idx for wd in b: _lowerCAmelCase : Any = idx return vocab, raw_vocab, ids_to_tokens, emoji class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = ["input_ids", "attention_mask"] def __init__( self , snake_case__ , snake_case__ , snake_case__="<|endoftext|>" , snake_case__="<|endoftext|>" , snake_case__="<|startoftext|>" , snake_case__="<|endoftext|>" , snake_case__=False , **snake_case__ , ): '''simple docstring''' super().__init__( unk_token=snake_case__ , pad_token=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , do_clean_text=snake_case__ , **snake_case__ , ) if not os.path.isfile(snake_case__ ): raise ValueError( F'Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained' ' model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) if not os.path.isfile(snake_case__ ): raise ValueError( F'Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google' ' pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) _lowerCAmelCase : Optional[Any] = do_clean_text _lowerCAmelCase : List[Any] = load_vocab_and_emoji(snake_case__ , snake_case__ ) _lowerCAmelCase : Optional[int] = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def a ( self ): '''simple docstring''' return len(self.raw_vocab ) def a ( self ): '''simple docstring''' return dict(self.raw_vocab , **self.added_tokens_encoder ) def a ( self , snake_case__ ): '''simple docstring''' return self.subword_tokenizer.tokenize(snake_case__ , clean=self.do_clean_text ) def a ( self , snake_case__ ): '''simple docstring''' return self.vocab.get(snake_case__ , self.vocab.get(self.unk_token ) ) def a ( self , snake_case__ ): '''simple docstring''' return self.subword_tokenizer.convert_id_to_token(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = ''.join(snake_case__ ).strip() return out_string def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Tuple = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(snake_case__ , add_special_tokens=snake_case__ ) + [self.eos_token_id] ) if len(snake_case__ ) > self.model_max_length: _lowerCAmelCase : List[str] = input_ids[-self.model_max_length :] return input_ids def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : Tuple = 0 if os.path.isdir(snake_case__ ): _lowerCAmelCase : Optional[Any] = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _lowerCAmelCase : Tuple = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['emoji_file'] ) else: _lowerCAmelCase : str = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['vocab_file'] ) _lowerCAmelCase : Dict = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['emoji_file'] ) with open(snake_case__ , 'w' , encoding='utf-8' ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( F'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' ' Please check that the vocabulary is not corrupted!' ) _lowerCAmelCase : Optional[Any] = token_index writer.write(','.join(snake_case__ ) + '\n' ) index += 1 with open(snake_case__ , 'w' , encoding='utf-8' ) as writer: json.dump(self.emoji , snake_case__ ) return vocab_file, emoji_file class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = vocab # same as swe _lowerCAmelCase : Tuple = ids_to_tokens # same as bpe _lowerCAmelCase : List[str] = emoji _lowerCAmelCase : Union[str, Any] = np.max([len(snake_case__ ) for w in self.vocab.keys()] ) _lowerCAmelCase : int = re.compile(R'(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)' ) _lowerCAmelCase : Optional[int] = re.compile(R'[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*' ) _lowerCAmelCase : Any = re.compile(R'[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}' ) _lowerCAmelCase : Optional[Any] = re.compile( R'([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) _lowerCAmelCase : List[Any] = re.compile( R'(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) _lowerCAmelCase : Optional[Any] = re.compile( R'((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*' ) _lowerCAmelCase : Dict = '─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿' _lowerCAmelCase : List[Any] = '▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟' _lowerCAmelCase : Optional[Any] = str.maketrans({k: '<BLOCK>' for k in keisen + blocks} ) def __len__( self ): '''simple docstring''' return len(self.ids_to_tokens ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = self.content_repattera.sub('<URL>' , snake_case__ ) _lowerCAmelCase : str = self.content_repattera.sub('<EMAIL>' , snake_case__ ) _lowerCAmelCase : int = self.content_repattera.sub('<TEL>' , snake_case__ ) _lowerCAmelCase : str = self.content_repattera.sub('<DATE>' , snake_case__ ) _lowerCAmelCase : int = self.content_repattera.sub('<DATE>' , snake_case__ ) _lowerCAmelCase : List[Any] = self.content_repattera.sub('<PRICE>' , snake_case__ ) _lowerCAmelCase : Dict = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: _lowerCAmelCase : Dict = content.replace('<BLOCK><BLOCK>' , '<BLOCK>' ) return content def a ( self , snake_case__ , snake_case__=False ): '''simple docstring''' _lowerCAmelCase : Any = text.replace(' ' , '<SP>' ) _lowerCAmelCase : Any = text.replace(' ' , '<SP>' ) _lowerCAmelCase : Dict = text.replace('\r\n' , '<BR>' ) _lowerCAmelCase : int = text.replace('\n' , '<BR>' ) _lowerCAmelCase : Tuple = text.replace('\r' , '<BR>' ) _lowerCAmelCase : int = text.replace('\t' , '<TAB>' ) _lowerCAmelCase : List[Any] = text.replace('—' , 'ー' ) _lowerCAmelCase : List[Any] = text.replace('−' , 'ー' ) for k, v in self.emoji["emoji"].items(): if k in text: _lowerCAmelCase : Union[str, Any] = text.replace(snake_case__ , snake_case__ ) if clean: _lowerCAmelCase : Optional[int] = self.clean_text(snake_case__ ) def check_simbol(snake_case__ ): _lowerCAmelCase : int = x.encode() if len(snake_case__ ) == 1 and len(snake_case__ ) == 2: _lowerCAmelCase : str = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0xC2_A1 and c <= 0xC2_BF) or (c >= 0xC7_80 and c <= 0xC7_83) or (c >= 0xCA_B9 and c <= 0xCB_BF) or (c >= 0xCC_80 and c <= 0xCD_A2) ): return True return False def checkuae(snake_case__ ): _lowerCAmelCase : Union[str, Any] = x.encode() if len(snake_case__ ) == 1 and len(snake_case__ ) == 3: _lowerCAmelCase : Dict = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0xE2_80_80 and c <= 0xE2_B0_7F: return True return False _lowerCAmelCase : Optional[Any] = 0 _lowerCAmelCase : int = [] while pos < len(snake_case__ ): _lowerCAmelCase : List[Any] = min(len(snake_case__ ) , pos + self.maxlen + 1 ) if text[pos] == '<' else pos + 3 _lowerCAmelCase : str = [] # (token_id, token, pos) for e in range(snake_case__ , snake_case__ , -1 ): _lowerCAmelCase : Tuple = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(snake_case__ ) > 2: _lowerCAmelCase : List[Any] = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(snake_case__ ) > 0: # the smallest token_id is adopted _lowerCAmelCase : List[str] = sorted(snake_case__ , key=lambda snake_case__ : x[0] )[0] result.append(snake_case__ ) _lowerCAmelCase : int = e else: _lowerCAmelCase : List[Any] = pos + 1 _lowerCAmelCase : List[str] = text[pos:end] if check_simbol(snake_case__ ): result.append('<KIGOU>' ) elif checkuae(snake_case__ ): result.append('<U2000U2BFF>' ) else: for i in wd.encode('utf-8' ): result.append('<|byte%d|>' % i ) _lowerCAmelCase : List[Any] = end return result def a ( self , snake_case__ , snake_case__="\n" ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [] _lowerCAmelCase : List[Any] = [] _lowerCAmelCase : int = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(snake_case__ ) > 0: words.append(bytearray(snake_case__ ).decode('utf-8' , errors='replace' ) ) _lowerCAmelCase : Tuple = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji['emoji_inv'][word] ) elif word == "<SP>": words.append(' ' ) elif word == "<BR>": words.append(snake_case__ ) elif word == "<TAB>": words.append('\t' ) elif word == "<BLOCK>": words.append('▀' ) elif word == "<KIGOU>": words.append('ǀ' ) elif word == "<U2000U2BFF>": words.append('‖' ) else: words.append(snake_case__ ) if len(snake_case__ ) > 0: words.append(bytearray(snake_case__ ).decode('utf-8' , errors='replace' ) ) _lowerCAmelCase : Dict = ''.join(snake_case__ ) return text
709
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Optional[int] = { """configuration_mctct""": ["""MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MCTCTConfig"""], """feature_extraction_mctct""": ["""MCTCTFeatureExtractor"""], """processing_mctct""": ["""MCTCTProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : List[str] = [ """MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MCTCTForCTC""", """MCTCTModel""", """MCTCTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys lowerCAmelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCAmelCase : Union[str, Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = ["""BartphoTokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys lowerCAmelCase : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
710
'''simple docstring''' lowerCAmelCase : Optional[int] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_00_00)] def lowercase (_A ): """simple docstring""" _lowerCAmelCase : str = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 1_0_0_0_0_0] number //= 1_0_0_0_0_0 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution lowerCAmelCase : list[bool | None] = [None] * 10_00_00_00 lowerCAmelCase : List[str] = True lowerCAmelCase : Union[str, Any] = False def lowercase (_A ): """simple docstring""" if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _lowerCAmelCase : Any = chain(next_number(_A ) ) _lowerCAmelCase : List[str] = number_chain while number < 1_0_0_0_0_0_0_0: _lowerCAmelCase : Tuple = number_chain number *= 1_0 return number_chain def lowercase (_A = 1_0_0_0_0_0_0_0 ): """simple docstring""" for i in range(1 , _A ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(_A ) if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution() = }''')
630
0
'''simple docstring''' import qiskit def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : Optional[Any] = qiskit.Aer.get_backend('aer_simulator' ) # Create a Quantum Circuit acting on the q register _lowerCAmelCase : Optional[int] = qiskit.QuantumCircuit(_A , _A ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator _lowerCAmelCase : List[str] = qiskit.execute(_A , _A , shots=1_0_0_0 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_A ) if __name__ == "__main__": print(F'''Total count for various states are: {single_qubit_measure(1, 1)}''')
711
'''simple docstring''' import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(snake_case__ , 'width_multiplier' ) ) class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=64 , snake_case__=2 , snake_case__=3 , snake_case__="swish" , snake_case__=3 , snake_case__=32 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=True , snake_case__=True , snake_case__=10 , snake_case__=None , snake_case__=0.25 , snake_case__=0.0 , snake_case__=0.0 , ): '''simple docstring''' _lowerCAmelCase : Tuple = parent _lowerCAmelCase : Optional[int] = batch_size _lowerCAmelCase : List[Any] = image_size _lowerCAmelCase : List[Any] = patch_size _lowerCAmelCase : List[str] = num_channels _lowerCAmelCase : Optional[Any] = make_divisible(512 * width_multiplier , divisor=8 ) _lowerCAmelCase : Optional[Any] = hidden_act _lowerCAmelCase : List[Any] = conv_kernel_size _lowerCAmelCase : Optional[Any] = output_stride _lowerCAmelCase : List[Any] = classifier_dropout_prob _lowerCAmelCase : str = use_labels _lowerCAmelCase : List[str] = is_training _lowerCAmelCase : Optional[int] = num_labels _lowerCAmelCase : List[str] = initializer_range _lowerCAmelCase : str = scope _lowerCAmelCase : Any = width_multiplier _lowerCAmelCase : Union[str, Any] = ffn_dropout _lowerCAmelCase : Optional[int] = attn_dropout def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase : Optional[Any] = None _lowerCAmelCase : Dict = None if self.use_labels: _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _lowerCAmelCase : Optional[int] = self.get_config() return config, pixel_values, labels, pixel_labels def a ( self ): '''simple docstring''' return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : str = model(snake_case__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.num_labels _lowerCAmelCase : List[Any] = MobileViTVaForImageClassification(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : int = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.num_labels _lowerCAmelCase : Optional[int] = MobileViTVaForSemanticSegmentation(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Dict = model(snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) _lowerCAmelCase : Any = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Optional[int] = config_and_inputs _lowerCAmelCase : Tuple = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = ( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) __magic_name__ = ( { "feature-extraction": MobileViTVaModel, "image-classification": MobileViTVaForImageClassification, "image-segmentation": MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False def a ( self ): '''simple docstring''' _lowerCAmelCase : int = MobileViTVaModelTester(self ) _lowerCAmelCase : Dict = MobileViTVaConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ ) def a ( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='MobileViTV2 does not use inputs_embeds' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not support input and output embeddings' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not output attentions' ) def a ( self ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason='Got `CUDA error: misaligned address` for tests after this one being run.' ) def a ( self ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : str = model_class(snake_case__ ) _lowerCAmelCase : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCAmelCase : int = [*signature.parameters.keys()] _lowerCAmelCase : Any = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def a ( self ): '''simple docstring''' def check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ): _lowerCAmelCase : Dict = model_class(snake_case__ ) model.to(snake_case__ ) model.eval() with torch.no_grad(): _lowerCAmelCase : Optional[Any] = model(**self._prepare_for_class(snake_case__ , snake_case__ ) ) _lowerCAmelCase : List[str] = outputs.hidden_states _lowerCAmelCase : List[str] = 5 self.assertEqual(len(snake_case__ ) , snake_case__ ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. _lowerCAmelCase : List[Any] = 2 for i in range(len(snake_case__ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : Optional[int] = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCAmelCase : Any = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*snake_case__ ) @slow def a ( self ): '''simple docstring''' for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase : Dict = MobileViTVaModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) def lowercase (): """simple docstring""" _lowerCAmelCase : Any = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def a ( self ): '''simple docstring''' return ( MobileViTImageProcessor.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ) if is_vision_available() else None ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = MobileViTVaForImageClassification.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ).to( snake_case__ ) _lowerCAmelCase : str = self.default_image_processor _lowerCAmelCase : Any = prepare_img() _lowerCAmelCase : Optional[int] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Tuple = model(**snake_case__ ) # verify the logits _lowerCAmelCase : Optional[Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , snake_case__ ) _lowerCAmelCase : Tuple = torch.tensor([-1.6336E00, -7.3204E-02, -5.1883E-01] ).to(snake_case__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Any = model.to(snake_case__ ) _lowerCAmelCase : int = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Optional[int] = prepare_img() _lowerCAmelCase : Dict = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : int = model(**snake_case__ ) _lowerCAmelCase : Dict = outputs.logits # verify the logits _lowerCAmelCase : str = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , snake_case__ ) _lowerCAmelCase : Any = torch.tensor( [ [[7.0863, 7.1525, 6.8201], [6.6931, 6.8770, 6.8933], [6.2978, 7.0366, 6.9636]], [[-3.7134, -3.6712, -3.6675], [-3.5825, -3.3549, -3.4777], [-3.3435, -3.3979, -3.2857]], [[-2.9329, -2.8003, -2.7369], [-3.0564, -2.4780, -2.0207], [-2.6889, -1.9298, -1.7640]], ] , device=snake_case__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : List[Any] = model.to(snake_case__ ) _lowerCAmelCase : str = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Tuple = prepare_img() _lowerCAmelCase : List[str] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Any = model(**snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.logits.detach().cpu() _lowerCAmelCase : Any = image_processor.post_process_semantic_segmentation(outputs=snake_case__ , target_sizes=[(50, 60)] ) _lowerCAmelCase : List[Any] = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , snake_case__ ) _lowerCAmelCase : List[str] = image_processor.post_process_semantic_segmentation(outputs=snake_case__ ) _lowerCAmelCase : Tuple = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , snake_case__ )
630
0
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging lowerCAmelCase : Any = """\ """ lowerCAmelCase : Union[str, Any] = """ Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity """ lowerCAmelCase : str = """ Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to 'cuda' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"] >>> results = perplexity.compute(model_id='gpt2', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 78.22 >>> print(round(results[\"perplexities\"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = datasets.load_dataset(\"wikitext\", ... \"wikitext-2-raw-v1\", ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=''] >>> results = perplexity.compute(model_id='gpt2', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 60.35 >>> print(round(results[\"perplexities\"][0], 2)) 81.12 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase__ ( datasets.Metric ): """simple docstring""" def a ( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'input_texts': datasets.Value('string' ), } ) , reference_urls=['https://huggingface.co/docs/transformers/perplexity'] , ) def a ( self , snake_case__ , snake_case__ , snake_case__ = 16 , snake_case__ = True , snake_case__=None ): '''simple docstring''' if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": _lowerCAmelCase : Tuple = 'cuda' else: _lowerCAmelCase : Optional[Any] = 'cuda' if torch.cuda.is_available() else 'cpu' _lowerCAmelCase : Any = AutoModelForCausalLM.from_pretrained(snake_case__ ) _lowerCAmelCase : Union[str, Any] = model.to(snake_case__ ) _lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained(snake_case__ ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: _lowerCAmelCase : List[str] = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(snake_case__ ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'pad_token': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" _lowerCAmelCase : List[str] = model.config.max_length - 1 else: _lowerCAmelCase : Optional[int] = model.config.max_length _lowerCAmelCase : Optional[int] = tokenizer( snake_case__ , add_special_tokens=snake_case__ , padding=snake_case__ , truncation=snake_case__ , max_length=snake_case__ , return_tensors='pt' , return_attention_mask=snake_case__ , ).to(snake_case__ ) _lowerCAmelCase : int = encodings['input_ids'] _lowerCAmelCase : Optional[Any] = encodings['attention_mask'] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." _lowerCAmelCase : List[str] = [] _lowerCAmelCase : Dict = CrossEntropyLoss(reduction='none' ) for start_index in logging.tqdm(range(0 , len(snake_case__ ) , snake_case__ ) ): _lowerCAmelCase : Union[str, Any] = min(start_index + batch_size , len(snake_case__ ) ) _lowerCAmelCase : Tuple = encoded_texts[start_index:end_index] _lowerCAmelCase : Union[str, Any] = attn_masks[start_index:end_index] if add_start_token: _lowerCAmelCase : Any = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(snake_case__ ) _lowerCAmelCase : int = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) _lowerCAmelCase : Optional[Any] = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(snake_case__ ), attn_mask] , dim=1 ) _lowerCAmelCase : Optional[Any] = encoded_batch with torch.no_grad(): _lowerCAmelCase : Any = model(snake_case__ , attention_mask=snake_case__ ).logits _lowerCAmelCase : List[str] = out_logits[..., :-1, :].contiguous() _lowerCAmelCase : Optional[int] = labels[..., 1:].contiguous() _lowerCAmelCase : Any = attn_mask[..., 1:].contiguous() _lowerCAmelCase : Any = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , snake_case__ ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(snake_case__ )}
712
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = FlaxXLMRobertaModel.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Dict = 'The dog is cute and lives in the garden house' _lowerCAmelCase : List[str] = jnp.array([tokenizer.encode(snake_case__ )] ) _lowerCAmelCase : Optional[int] = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim _lowerCAmelCase : Tuple = jnp.array( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) _lowerCAmelCase : Union[str, Any] = model(snake_case__ )['last_hidden_state'] self.assertEqual(output.shape , snake_case__ ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , snake_case__ , atol=1E-3 ) )
630
0
'''simple docstring''' from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class UpperCamelCase__ : """simple docstring""" pass
713
'''simple docstring''' def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Dict = len(_A ) while cur > 1: # Find the maximum number in arr _lowerCAmelCase : int = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi _lowerCAmelCase : Dict = arr[mi::-1] + arr[mi + 1 : len(_A )] # Reverse whole list _lowerCAmelCase : Optional[int] = arr[cur - 1 :: -1] + arr[cur : len(_A )] cur -= 1 return arr if __name__ == "__main__": lowerCAmelCase : Tuple = input("""Enter numbers separated by a comma:\n""").strip() lowerCAmelCase : Tuple = [int(item) for item in user_input.split(""",""")] print(pancake_sort(unsorted))
630
0
from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
714
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) lowerCAmelCase : str = { """EleutherAI/gpt-j-6B""": """https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json""", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "gptj" __magic_name__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , snake_case__=5_0400 , snake_case__=2048 , snake_case__=4096 , snake_case__=28 , snake_case__=16 , snake_case__=64 , snake_case__=None , snake_case__="gelu_new" , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=1E-5 , snake_case__=0.02 , snake_case__=True , snake_case__=5_0256 , snake_case__=5_0256 , snake_case__=False , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : int = vocab_size _lowerCAmelCase : Optional[int] = n_positions _lowerCAmelCase : Optional[int] = n_embd _lowerCAmelCase : Optional[int] = n_layer _lowerCAmelCase : str = n_head _lowerCAmelCase : Tuple = n_inner _lowerCAmelCase : Tuple = rotary_dim _lowerCAmelCase : Optional[int] = activation_function _lowerCAmelCase : Any = resid_pdrop _lowerCAmelCase : List[str] = embd_pdrop _lowerCAmelCase : int = attn_pdrop _lowerCAmelCase : Any = layer_norm_epsilon _lowerCAmelCase : Optional[int] = initializer_range _lowerCAmelCase : List[str] = use_cache _lowerCAmelCase : Dict = bos_token_id _lowerCAmelCase : Any = eos_token_id super().__init__( bos_token_id=snake_case__ , eos_token_id=snake_case__ , tie_word_embeddings=snake_case__ , **snake_case__ ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = "default" , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' super().__init__(snake_case__ , task=snake_case__ , patching_specs=snake_case__ , use_past=snake_case__ ) if not getattr(self._config , 'pad_token_id' , snake_case__ ): # TODO: how to do that better? _lowerCAmelCase : Any = 0 @property def a ( self ): '''simple docstring''' _lowerCAmelCase : str = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(snake_case__ , direction='inputs' ) _lowerCAmelCase : int = {0: 'batch', 1: 'past_sequence + sequence'} else: _lowerCAmelCase : int = {0: 'batch', 1: 'sequence'} return common_inputs @property def a ( self ): '''simple docstring''' return self._config.n_layer @property def a ( self ): '''simple docstring''' return self._config.n_head def a ( self , snake_case__ , snake_case__ = -1 , snake_case__ = -1 , snake_case__ = False , snake_case__ = None , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = super(snake_case__ , self ).generate_dummy_inputs( snake_case__ , batch_size=snake_case__ , seq_length=snake_case__ , is_pair=snake_case__ , framework=snake_case__ ) # We need to order the input in the way they appears in the forward() _lowerCAmelCase : Any = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = common_inputs['input_ids'].shape # Not using the same length for past_key_values _lowerCAmelCase : Any = seqlen + 2 _lowerCAmelCase : Optional[int] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _lowerCAmelCase : Tuple = [ (torch.zeros(snake_case__ ), torch.zeros(snake_case__ )) for _ in range(self.num_layers ) ] _lowerCAmelCase : Tuple = common_inputs['attention_mask'] if self.use_past: _lowerCAmelCase : Any = ordered_inputs['attention_mask'].dtype _lowerCAmelCase : Union[str, Any] = torch.cat( [ordered_inputs['attention_mask'], torch.ones(snake_case__ , snake_case__ , dtype=snake_case__ )] , dim=1 ) return ordered_inputs @property def a ( self ): '''simple docstring''' return 13
630
0
'''simple docstring''' import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def lowercase (): """simple docstring""" _lowerCAmelCase : Dict = 'https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png' _lowerCAmelCase : Any = Image.open(requests.get(_A , stream=_A ).raw ).convert('RGB' ) return image def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Any = [] # fmt: off # vision encoder rename_keys.append(('visual_encoder.cls_token', 'vision_model.embeddings.class_embedding') ) rename_keys.append(('visual_encoder.pos_embed', 'vision_model.embeddings.position_embedding') ) rename_keys.append(('visual_encoder.patch_embed.proj.weight', 'vision_model.embeddings.patch_embedding.weight') ) rename_keys.append(('visual_encoder.patch_embed.proj.bias', 'vision_model.embeddings.patch_embedding.bias') ) rename_keys.append(('ln_vision.weight', 'vision_model.post_layernorm.weight') ) rename_keys.append(('ln_vision.bias', 'vision_model.post_layernorm.bias') ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((f'visual_encoder.blocks.{i}.norm1.weight', f'vision_model.encoder.layers.{i}.layer_norm1.weight') ) rename_keys.append((f'visual_encoder.blocks.{i}.norm1.bias', f'vision_model.encoder.layers.{i}.layer_norm1.bias') ) rename_keys.append((f'visual_encoder.blocks.{i}.norm2.weight', f'vision_model.encoder.layers.{i}.layer_norm2.weight') ) rename_keys.append((f'visual_encoder.blocks.{i}.norm2.bias', f'vision_model.encoder.layers.{i}.layer_norm2.bias') ) rename_keys.append((f'visual_encoder.blocks.{i}.attn.qkv.weight', f'vision_model.encoder.layers.{i}.self_attn.qkv.weight') ) rename_keys.append((f'visual_encoder.blocks.{i}.attn.proj.weight', f'vision_model.encoder.layers.{i}.self_attn.projection.weight',) ) rename_keys.append((f'visual_encoder.blocks.{i}.attn.proj.bias', f'vision_model.encoder.layers.{i}.self_attn.projection.bias') ) rename_keys.append((f'visual_encoder.blocks.{i}.mlp.fc1.weight', f'vision_model.encoder.layers.{i}.mlp.fc1.weight') ) rename_keys.append((f'visual_encoder.blocks.{i}.mlp.fc1.bias', f'vision_model.encoder.layers.{i}.mlp.fc1.bias') ) rename_keys.append((f'visual_encoder.blocks.{i}.mlp.fc2.weight', f'vision_model.encoder.layers.{i}.mlp.fc2.weight') ) rename_keys.append((f'visual_encoder.blocks.{i}.mlp.fc2.bias', f'vision_model.encoder.layers.{i}.mlp.fc2.bias') ) # QFormer rename_keys.append(('Qformer.bert.embeddings.LayerNorm.weight', 'qformer.layernorm.weight') ) rename_keys.append(('Qformer.bert.embeddings.LayerNorm.bias', 'qformer.layernorm.bias') ) # fmt: on return rename_keys def lowercase (_A , _A , _A ): """simple docstring""" _lowerCAmelCase : Optional[int] = dct.pop(_A ) _lowerCAmelCase : Optional[int] = val def lowercase (_A , _A ): """simple docstring""" for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases _lowerCAmelCase : List[str] = state_dict.pop(f'visual_encoder.blocks.{i}.attn.q_bias' ) _lowerCAmelCase : Dict = state_dict.pop(f'visual_encoder.blocks.{i}.attn.v_bias' ) # next, set bias in the state dict _lowerCAmelCase : Dict = torch.cat((q_bias, torch.zeros_like(_A , requires_grad=_A ), v_bias) ) _lowerCAmelCase : Optional[int] = qkv_bias def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : Optional[int] = 3_6_4 if 'coco' in model_name else 2_2_4 _lowerCAmelCase : List[str] = BlipaVisionConfig(image_size=_A ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: _lowerCAmelCase : int = OPTConfig.from_pretrained('facebook/opt-2.7b' , eos_token_id=_A ).to_dict() elif "opt-6.7b" in model_name: _lowerCAmelCase : List[Any] = OPTConfig.from_pretrained('facebook/opt-6.7b' , eos_token_id=_A ).to_dict() elif "t5-xl" in model_name: _lowerCAmelCase : Optional[Any] = TaConfig.from_pretrained('google/flan-t5-xl' , dense_act_fn='gelu' , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: _lowerCAmelCase : int = TaConfig.from_pretrained('google/flan-t5-xxl' , dense_act_fn='gelu' , bos_token_id=1 ).to_dict() _lowerCAmelCase : List[str] = BlipaConfig(vision_config=_A , text_config=_A ) return config, image_size @torch.no_grad() def lowercase (_A , _A=None , _A=False ): """simple docstring""" _lowerCAmelCase : Union[str, Any] = ( AutoTokenizer.from_pretrained('facebook/opt-2.7b' ) if 'opt' in model_name else AutoTokenizer.from_pretrained('google/flan-t5-xl' ) ) _lowerCAmelCase : str = tokenizer('\n' , add_special_tokens=_A ).input_ids[0] _lowerCAmelCase : Any = get_blipa_config(_A , eos_token_id=_A ) _lowerCAmelCase : Any = BlipaForConditionalGeneration(_A ).eval() _lowerCAmelCase : Any = { 'blip2-opt-2.7b': ('blip2_opt', 'pretrain_opt2.7b'), 'blip2-opt-6.7b': ('blip2_opt', 'pretrain_opt6.7b'), 'blip2-opt-2.7b-coco': ('blip2_opt', 'caption_coco_opt2.7b'), 'blip2-opt-6.7b-coco': ('blip2_opt', 'caption_coco_opt6.7b'), 'blip2-flan-t5-xl': ('blip2_t5', 'pretrain_flant5xl'), 'blip2-flan-t5-xl-coco': ('blip2_t5', 'caption_coco_flant5xl'), 'blip2-flan-t5-xxl': ('blip2_t5', 'pretrain_flant5xxl'), } _lowerCAmelCase : Optional[Any] = model_name_to_original[model_name] # load original model print('Loading original model...' ) _lowerCAmelCase : Optional[Any] = 'cuda' if torch.cuda.is_available() else 'cpu' _lowerCAmelCase : str = load_model_and_preprocess( name=_A , model_type=_A , is_eval=_A , device=_A ) original_model.eval() print('Done!' ) # update state dict keys _lowerCAmelCase : str = original_model.state_dict() _lowerCAmelCase : Union[str, Any] = create_rename_keys(_A ) for src, dest in rename_keys: rename_key(_A , _A , _A ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): _lowerCAmelCase : Optional[int] = state_dict.pop(_A ) if key.startswith('Qformer.bert' ): _lowerCAmelCase : List[Any] = key.replace('Qformer.bert' , 'qformer' ) if "attention.self" in key: _lowerCAmelCase : List[str] = key.replace('self' , 'attention' ) if "opt_proj" in key: _lowerCAmelCase : Any = key.replace('opt_proj' , 'language_projection' ) if "t5_proj" in key: _lowerCAmelCase : List[Any] = key.replace('t5_proj' , 'language_projection' ) if key.startswith('opt' ): _lowerCAmelCase : List[Any] = key.replace('opt' , 'language' ) if key.startswith('t5' ): _lowerCAmelCase : Tuple = key.replace('t5' , 'language' ) _lowerCAmelCase : Optional[int] = val # read in qv biases read_in_q_v_bias(_A , _A ) _lowerCAmelCase : Any = hf_model.load_state_dict(_A , strict=_A ) assert len(_A ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] _lowerCAmelCase : Tuple = load_demo_image() _lowerCAmelCase : Optional[int] = vis_processors['eval'](_A ).unsqueeze(0 ).to(_A ) _lowerCAmelCase : Union[str, Any] = tokenizer(['\n'] , return_tensors='pt' ).input_ids.to(_A ) # create processor _lowerCAmelCase : Tuple = BlipImageProcessor( size={'height': image_size, 'width': image_size} , image_mean=_A , image_std=_A ) _lowerCAmelCase : List[str] = BlipaProcessor(image_processor=_A , tokenizer=_A ) _lowerCAmelCase : Union[str, Any] = processor(images=_A , return_tensors='pt' ).pixel_values.to(_A ) # make sure processor creates exact same pixel values assert torch.allclose(_A , _A ) original_model.to(_A ) hf_model.to(_A ) with torch.no_grad(): if "opt" in model_name: _lowerCAmelCase : Optional[int] = original_model({'image': original_pixel_values, 'text_input': ['']} ).logits _lowerCAmelCase : Tuple = hf_model(_A , _A ).logits else: _lowerCAmelCase : Optional[Any] = original_model( {'image': original_pixel_values, 'text_input': ['\n'], 'text_output': ['\n']} ).logits _lowerCAmelCase : List[str] = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -1_0_0 ) _lowerCAmelCase : Optional[int] = hf_model(_A , _A , labels=_A ).logits assert original_logits.shape == logits.shape print('First values of original logits:' , original_logits[0, :3, :3] ) print('First values of HF logits:' , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": _lowerCAmelCase : int = torch.tensor( [[-41.5_850, -4.4_440, -8.9_922], [-47.4_322, -5.9_143, -1.7_340]] , device=_A ) assert torch.allclose(logits[0, :3, :3] , _A , atol=1E-4 ) elif model_name == "blip2-flan-t5-xl-coco": _lowerCAmelCase : Tuple = torch.tensor( [[-57.0_109, -9.8_967, -12.6_280], [-68.6_578, -12.7_191, -10.5_065]] , device=_A ) else: # cast to same type _lowerCAmelCase : List[str] = logits.dtype assert torch.allclose(original_logits.to(_A ) , _A , atol=1E-2 ) print('Looks ok!' ) print('Generating a caption...' ) _lowerCAmelCase : List[str] = '' _lowerCAmelCase : List[str] = tokenizer(_A , return_tensors='pt' ).input_ids.to(_A ) _lowerCAmelCase : Optional[Any] = original_model.generate({'image': original_pixel_values} ) _lowerCAmelCase : Dict = hf_model.generate( _A , _A , do_sample=_A , num_beams=5 , max_length=3_0 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print('Original generation:' , _A ) _lowerCAmelCase : int = input_ids.shape[1] _lowerCAmelCase : str = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=_A ) _lowerCAmelCase : int = [text.strip() for text in output_text] print('HF generation:' , _A ) if pytorch_dump_folder_path is not None: processor.save_pretrained(_A ) hf_model.save_pretrained(_A ) if push_to_hub: processor.push_to_hub(f'nielsr/{model_name}' ) hf_model.push_to_hub(f'nielsr/{model_name}' ) if __name__ == "__main__": lowerCAmelCase : Tuple = argparse.ArgumentParser() lowerCAmelCase : str = [ """blip2-opt-2.7b""", """blip2-opt-6.7b""", """blip2-opt-2.7b-coco""", """blip2-opt-6.7b-coco""", """blip2-flan-t5-xl""", """blip2-flan-t5-xl-coco""", """blip2-flan-t5-xxl""", ] parser.add_argument( """--model_name""", default="""blip2-opt-2.7b""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub after converting""", ) lowerCAmelCase : Any = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
715
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Any = { """configuration_x_clip""": [ """XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XCLIPConfig""", """XCLIPTextConfig""", """XCLIPVisionConfig""", ], """processing_x_clip""": ["""XCLIPProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """XCLIPModel""", """XCLIPPreTrainedModel""", """XCLIPTextModel""", """XCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys lowerCAmelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = len(snake_case__ ) _lowerCAmelCase : List[str] = [0] * len_array if len_array > 0: _lowerCAmelCase : List[str] = array[0] for i in range(1 , snake_case__ ): _lowerCAmelCase : Optional[Any] = self.prefix_sum[i - 1] + array[i] def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Tuple = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(snake_case__ ) return False if __name__ == "__main__": import doctest doctest.testmod()
716
'''simple docstring''' import math from datetime import datetime, timedelta def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Optional[Any] = year % 1_9 _lowerCAmelCase : Any = year % 4 _lowerCAmelCase : Optional[int] = year % 7 _lowerCAmelCase : int = math.floor(year / 1_0_0 ) _lowerCAmelCase : Dict = math.floor((1_3 + 8 * leap_day_inhibits) / 2_5 ) _lowerCAmelCase : Optional[Any] = leap_day_inhibits / 4 _lowerCAmelCase : Dict = ( 1_5 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 3_0 _lowerCAmelCase : List[Any] = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 _lowerCAmelCase : Dict = (1_9 * metonic_cycle + secular_moon_shift) % 3_0 # PHM -> Paschal Full Moon _lowerCAmelCase : Union[str, Any] = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 2_9 and days_from_phm_to_sunday == 6: return datetime(_A , 4 , 1_9 ) elif days_to_add == 2_8 and days_from_phm_to_sunday == 6: return datetime(_A , 4 , 1_8 ) else: return datetime(_A , 3 , 2_2 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (19_94, 20_00, 20_10, 20_21, 20_23): lowerCAmelCase : List[str] = """will be""" if year > datetime.now().year else """was""" print(F'''Easter in {year} {tense} {gauss_easter(year)}''')
630
0
'''simple docstring''' import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel lowerCAmelCase : str = logging.getLogger(__name__) def lowercase (_A , _A ): """simple docstring""" if os.path.exists(_A ): if os.path.exists(os.path.join(_A , 'config.json' ) ) and os.path.isfile( os.path.join(_A , 'config.json' ) ): os.remove(os.path.join(_A , 'config.json' ) ) if os.path.exists(os.path.join(_A , 'pytorch_model.bin' ) ) and os.path.isfile( os.path.join(_A , 'pytorch_model.bin' ) ): os.remove(os.path.join(_A , 'pytorch_model.bin' ) ) else: os.makedirs(_A ) model.save_pretrained(_A ) def lowercase (_A , _A=False ): """simple docstring""" _lowerCAmelCase : Tuple = 2 if unlogit: _lowerCAmelCase : Union[str, Any] = torch.pow(_A , _A ) _lowerCAmelCase : List[Any] = p * torch.log(_A ) _lowerCAmelCase : List[str] = 0 return -plogp.sum(dim=-1 ) def lowercase (_A ): """simple docstring""" logger.info('lv, h >\t' + '\t'.join(f'{x + 1}' for x in range(len(_A ) ) ) ) for row in range(len(_A ) ): if tensor.dtype != torch.long: logger.info(f'layer {row + 1}:\t' + '\t'.join(f'{x:.5f}' for x in tensor[row].cpu().data ) ) else: logger.info(f'layer {row + 1}:\t' + '\t'.join(f'{x:d}' for x in tensor[row].cpu().data ) ) def lowercase (_A , _A , _A , _A=True , _A=True , _A=None , _A=False ): """simple docstring""" _lowerCAmelCase : List[str] = model.config.num_hidden_layers, model.config.num_attention_heads _lowerCAmelCase : Union[str, Any] = torch.zeros(_A , _A ).to(args.device ) _lowerCAmelCase : List[str] = torch.zeros(_A , _A ).to(args.device ) if head_mask is None: _lowerCAmelCase : Optional[Any] = torch.ones(_A , _A ).to(args.device ) head_mask.requires_grad_(requires_grad=_A ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: _lowerCAmelCase : Any = None _lowerCAmelCase : Any = 0.0 _lowerCAmelCase : List[str] = 0.0 for step, inputs in enumerate(tqdm(_A , desc='Iteration' , disable=args.local_rank not in [-1, 0] ) ): _lowerCAmelCase : Optional[int] = tuple(t.to(args.device ) for t in inputs ) (_lowerCAmelCase ) : Optional[int] = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) _lowerCAmelCase : str = model(_A , labels=_A , head_mask=_A ) # (loss), lm_logits, presents, (all hidden_states), (attentions) _lowerCAmelCase : str = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(_A ): _lowerCAmelCase : Optional[int] = entropy(attn.detach() , _A ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(_A ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: _lowerCAmelCase : Tuple = 2 _lowerCAmelCase : Tuple = torch.pow(torch.pow(_A , _A ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1E-20 if not args.dont_normalize_global_importance: _lowerCAmelCase : Dict = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info('Attention entropies' ) print_ad_tensor(_A ) if compute_importance: logger.info('Head importance scores' ) print_ad_tensor(_A ) logger.info('Head ranked by importance scores' ) _lowerCAmelCase : Optional[int] = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) _lowerCAmelCase : Union[str, Any] = torch.arange( head_importance.numel() , device=args.device ) _lowerCAmelCase : Tuple = head_ranks.view_as(_A ) print_ad_tensor(_A ) return attn_entropy, head_importance, total_loss def lowercase (_A , _A , _A ): """simple docstring""" _lowerCAmelCase : Union[str, Any] = compute_heads_importance(_A , _A , _A , compute_entropy=_A ) _lowerCAmelCase : int = 1 / loss # instead of downsteam score use the LM loss logger.info('Pruning: original score: %f, threshold: %f' , _A , original_score * args.masking_threshold ) _lowerCAmelCase : List[Any] = torch.ones_like(_A ) _lowerCAmelCase : Dict = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) _lowerCAmelCase : Optional[Any] = original_score while current_score >= original_score * args.masking_threshold: _lowerCAmelCase : Dict = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads _lowerCAmelCase : int = float('Inf' ) _lowerCAmelCase : int = head_importance.view(-1 ).sort()[1] if len(_A ) <= num_to_mask: print('BREAK BY num_to_mask' ) break # mask heads _lowerCAmelCase : Optional[int] = current_heads_to_mask[:num_to_mask] logger.info('Heads to mask: %s' , str(current_heads_to_mask.tolist() ) ) _lowerCAmelCase : Dict = new_head_mask.view(-1 ) _lowerCAmelCase : List[Any] = 0.0 _lowerCAmelCase : Optional[Any] = new_head_mask.view_as(_A ) _lowerCAmelCase : Union[str, Any] = new_head_mask.clone().detach() print_ad_tensor(_A ) # Compute metric and head importance again _lowerCAmelCase : int = compute_heads_importance( _A , _A , _A , compute_entropy=_A , head_mask=_A ) _lowerCAmelCase : List[str] = 1 / loss logger.info( 'Masking: current score: %f, remaining heads %d (%.1f percents)' , _A , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 1_0_0 , ) logger.info('Final head mask' ) print_ad_tensor(_A ) np.save(os.path.join(args.output_dir , 'head_mask.npy' ) , head_mask.detach().cpu().numpy() ) return head_mask def lowercase (_A , _A , _A , _A ): """simple docstring""" _lowerCAmelCase : Tuple = datetime.now() _lowerCAmelCase : Optional[int] = compute_heads_importance( _A , _A , _A , compute_entropy=_A , compute_importance=_A , head_mask=_A ) _lowerCAmelCase : Dict = 1 / loss _lowerCAmelCase : Optional[int] = datetime.now() - before_time _lowerCAmelCase : Dict = sum(p.numel() for p in model.parameters() ) _lowerCAmelCase : int = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(_A ) ) } for k, v in heads_to_prune.items(): if isinstance(_A , _A ): _lowerCAmelCase : Dict = [ v, ] assert sum(len(_A ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(_A ) _lowerCAmelCase : Dict = sum(p.numel() for p in model.parameters() ) _lowerCAmelCase : Union[str, Any] = datetime.now() _lowerCAmelCase : Optional[int] = compute_heads_importance( _A , _A , _A , compute_entropy=_A , compute_importance=_A , head_mask=_A , actually_pruned=_A , ) _lowerCAmelCase : Dict = 1 / loss _lowerCAmelCase : Tuple = datetime.now() - before_time logger.info( 'Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)' , _A , _A , pruned_num_params / original_num_params * 1_0_0 , ) logger.info('Pruning: score with masking: %f score with pruning: %f' , _A , _A ) logger.info('Pruning: speed ratio (original timing / new timing): %f percents' , original_time / new_time * 1_0_0 ) save_model(_A , args.output_dir ) def lowercase (): """simple docstring""" _lowerCAmelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--data_dir' , default=_A , type=_A , required=_A , help='The input data dir. Should contain the .tsv files (or other data files) for the task.' , ) parser.add_argument( '--model_name_or_path' , default=_A , type=_A , required=_A , help='Path to pretrained model or model identifier from huggingface.co/models' , ) parser.add_argument( '--output_dir' , default=_A , type=_A , required=_A , help='The output directory where the model predictions and checkpoints will be written.' , ) # Other parameters parser.add_argument( '--config_name' , default='' , type=_A , help='Pretrained config name or path if not the same as model_name_or_path' , ) parser.add_argument( '--tokenizer_name' , default='' , type=_A , help='Pretrained tokenizer name or path if not the same as model_name_or_path' , ) parser.add_argument( '--cache_dir' , default=_A , type=_A , help='Where do you want to store the pre-trained models downloaded from s3' , ) parser.add_argument( '--data_subset' , type=_A , default=-1 , help='If > 0: limit the data to a subset of data_subset instances.' ) parser.add_argument( '--overwrite_output_dir' , action='store_true' , help='Whether to overwrite data in output directory' ) parser.add_argument( '--overwrite_cache' , action='store_true' , help='Overwrite the cached training and evaluation sets' ) parser.add_argument( '--dont_normalize_importance_by_layer' , action='store_true' , help='Don\'t normalize importance score by layers' ) parser.add_argument( '--dont_normalize_global_importance' , action='store_true' , help='Don\'t normalize all importance scores between 0 and 1' , ) parser.add_argument( '--try_masking' , action='store_true' , help='Whether to try to mask head until a threshold of accuracy.' ) parser.add_argument( '--masking_threshold' , default=0.9 , type=_A , help='masking threshold in term of metrics (stop masking when metric < threshold * original metric value).' , ) parser.add_argument( '--masking_amount' , default=0.1 , type=_A , help='Amount to heads to masking at each masking step.' ) parser.add_argument('--metric_name' , default='acc' , type=_A , help='Metric to use for head masking.' ) parser.add_argument( '--max_seq_length' , default=1_2_8 , type=_A , help=( 'The maximum total input sequence length after WordPiece tokenization. \n' 'Sequences longer than this will be truncated, sequences shorter padded.' ) , ) parser.add_argument('--batch_size' , default=1 , type=_A , help='Batch size.' ) parser.add_argument('--seed' , type=_A , default=4_2 ) parser.add_argument('--local_rank' , type=_A , default=-1 , help='local_rank for distributed training on gpus' ) parser.add_argument('--no_cuda' , action='store_true' , help='Whether not to use CUDA when available' ) parser.add_argument('--server_ip' , type=_A , default='' , help='Can be used for distant debugging.' ) parser.add_argument('--server_port' , type=_A , default='' , help='Can be used for distant debugging.' ) _lowerCAmelCase : Union[str, Any] = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('Waiting for debugger attach' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=_A ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: _lowerCAmelCase : Dict = torch.device('cuda' if torch.cuda.is_available() and not args.no_cuda else 'cpu' ) _lowerCAmelCase : Union[str, Any] = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) _lowerCAmelCase : List[str] = torch.device('cuda' , args.local_rank ) _lowerCAmelCase : Any = 1 torch.distributed.init_process_group(backend='nccl' ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info('device: {} n_gpu: {}, distributed: {}'.format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) _lowerCAmelCase : str = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: _lowerCAmelCase : Any = nn.parallel.DistributedDataParallel( _A , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=_A ) elif args.n_gpu > 1: _lowerCAmelCase : int = nn.DataParallel(_A ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=_A ) torch.save(_A , os.path.join(args.output_dir , 'run_args.bin' ) ) logger.info('Training/evaluation parameters %s' , _A ) # Prepare dataset _lowerCAmelCase : List[str] = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) _lowerCAmelCase : Dict = (torch.from_numpy(_A ),) _lowerCAmelCase : str = TensorDataset(*_A ) _lowerCAmelCase : Dict = RandomSampler(_A ) _lowerCAmelCase : Any = DataLoader(_A , sampler=_A , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(_A , _A , _A ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: _lowerCAmelCase : Union[str, Any] = mask_heads(_A , _A , _A ) prune_heads(_A , _A , _A , _A ) if __name__ == "__main__": main()
717
'''simple docstring''' import unittest from knapsack import greedy_knapsack as kp class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [10, 20, 30, 40, 50, 60] _lowerCAmelCase : Union[str, Any] = [2, 4, 6, 8, 10, 12] _lowerCAmelCase : Dict = 100 self.assertEqual(kp.calc_profit(snake_case__ , snake_case__ , snake_case__ ) , 210 ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'max_weight must greater than zero.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'Weight can not be negative.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'Profit can not be negative.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'max_weight must greater than zero.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex( snake_case__ , 'The length of profit and weight must be same.' ) if __name__ == "__main__": unittest.main()
630
0
'''simple docstring''' import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase : Any = logging.get_logger(__name__) lowerCAmelCase : Dict = """▁""" lowerCAmelCase : Any = { """vocab_file""": """vocab.json""", """spm_file""": """sentencepiece.bpe.model""", } lowerCAmelCase : Tuple = { """vocab_file""": { """facebook/s2t-small-librispeech-asr""": ( """https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/vocab.json""" ), }, """spm_file""": { """facebook/s2t-small-librispeech-asr""": ( """https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/sentencepiece.bpe.model""" ) }, } lowerCAmelCase : str = { """facebook/s2t-small-librispeech-asr""": 10_24, } lowerCAmelCase : Any = ["""pt""", """fr""", """ru""", """nl""", """ro""", """it""", """es""", """de"""] lowerCAmelCase : List[str] = {"""mustc""": MUSTC_LANGS} class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = MAX_MODEL_INPUT_SIZES __magic_name__ = ["input_ids", "attention_mask"] __magic_name__ = [] def __init__( self , snake_case__ , snake_case__ , snake_case__="<s>" , snake_case__="</s>" , snake_case__="<pad>" , snake_case__="<unk>" , snake_case__=False , snake_case__=False , snake_case__=None , snake_case__=None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , pad_token=snake_case__ , do_upper_case=snake_case__ , do_lower_case=snake_case__ , tgt_lang=snake_case__ , lang_codes=snake_case__ , sp_model_kwargs=self.sp_model_kwargs , **snake_case__ , ) _lowerCAmelCase : Optional[Any] = do_upper_case _lowerCAmelCase : List[Any] = do_lower_case _lowerCAmelCase : List[Any] = load_json(snake_case__ ) _lowerCAmelCase : Optional[Any] = {v: k for k, v in self.encoder.items()} _lowerCAmelCase : str = spm_file _lowerCAmelCase : List[str] = load_spm(snake_case__ , self.sp_model_kwargs ) if lang_codes is not None: _lowerCAmelCase : Dict = lang_codes _lowerCAmelCase : List[Any] = LANGUAGES[lang_codes] _lowerCAmelCase : Optional[int] = [F'<lang:{lang}>' for lang in self.langs] _lowerCAmelCase : List[str] = {lang: self.sp_model.PieceToId(F'<lang:{lang}>' ) for lang in self.langs} _lowerCAmelCase : str = self.lang_tokens _lowerCAmelCase : Union[str, Any] = tgt_lang if tgt_lang is not None else self.langs[0] self.set_tgt_lang_special_tokens(self._tgt_lang ) else: _lowerCAmelCase : Optional[int] = {} @property def a ( self ): '''simple docstring''' return len(self.encoder ) @property def a ( self ): '''simple docstring''' return self._tgt_lang @tgt_lang.setter def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = new_tgt_lang self.set_tgt_lang_special_tokens(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.lang_code_to_id[tgt_lang] _lowerCAmelCase : Optional[Any] = [lang_code_id] def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.encode(snake_case__ , out_type=snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.encoder.get(snake_case__ , self.encoder[self.unk_token] ) def a ( self , snake_case__ ): '''simple docstring''' return self.decoder.get(snake_case__ , self.unk_token ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = [] _lowerCAmelCase : Dict = '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: _lowerCAmelCase : Any = self.sp_model.decode(snake_case__ ) out_string += (decoded.upper() if self.do_upper_case else decoded) + token + " " _lowerCAmelCase : List[str] = [] else: current_sub_tokens.append(snake_case__ ) _lowerCAmelCase : Tuple = self.sp_model.decode(snake_case__ ) out_string += decoded.upper() if self.do_upper_case else decoded return out_string.strip() def a ( self , snake_case__ , snake_case__=None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + [self.eos_token_id] def a ( self , snake_case__ , snake_case__ = None , snake_case__ = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case__ , token_ids_a=snake_case__ , already_has_special_tokens=snake_case__ ) _lowerCAmelCase : List[Any] = [1] * len(self.prefix_tokens ) _lowerCAmelCase : Optional[Any] = [1] if token_ids_a is None: return prefix_ones + ([0] * len(snake_case__ )) + suffix_ones return prefix_ones + ([0] * len(snake_case__ )) + ([0] * len(snake_case__ )) + suffix_ones def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.encoder.copy() vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = self.__dict__.copy() _lowerCAmelCase : str = None return state def __setstate__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowerCAmelCase : List[str] = {} _lowerCAmelCase : List[str] = load_spm(self.spm_file , self.sp_model_kwargs ) def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = Path(snake_case__ ) assert save_dir.is_dir(), F'{save_directory} should be a directory' _lowerCAmelCase : Union[str, Any] = save_dir / ( (filename_prefix + '-' if filename_prefix else '') + self.vocab_files_names['vocab_file'] ) _lowerCAmelCase : List[str] = save_dir / ( (filename_prefix + '-' if filename_prefix else '') + self.vocab_files_names['spm_file'] ) save_json(self.encoder , snake_case__ ) if os.path.abspath(self.spm_file ) != os.path.abspath(snake_case__ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , snake_case__ ) elif not os.path.isfile(self.spm_file ): with open(snake_case__ , 'wb' ) as fi: _lowerCAmelCase : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (str(snake_case__ ), str(snake_case__ )) def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : Union[str, Any] = sentencepiece.SentencePieceProcessor(**_A ) spm.Load(str(_A ) ) return spm def lowercase (_A ): """simple docstring""" with open(_A , 'r' ) as f: return json.load(_A ) def lowercase (_A , _A ): """simple docstring""" with open(_A , 'w' ) as f: json.dump(_A , _A , indent=2 )
718
'''simple docstring''' def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : Optional[int] = (boundary[1] - boundary[0]) / steps _lowerCAmelCase : Any = boundary[0] _lowerCAmelCase : List[str] = boundary[1] _lowerCAmelCase : Tuple = make_points(_A , _A , _A ) _lowerCAmelCase : Tuple = 0.0 y += (h / 2.0) * f(_A ) for i in x_i: # print(i) y += h * f(_A ) y += (h / 2.0) * f(_A ) return y def lowercase (_A , _A , _A ): """simple docstring""" _lowerCAmelCase : Tuple = a + h while x < (b - h): yield x _lowerCAmelCase : Any = x + h def lowercase (_A ): # enter your function here """simple docstring""" _lowerCAmelCase : int = (x - 0) * (x - 0) return y def lowercase (): """simple docstring""" _lowerCAmelCase : Optional[Any] = 0.0 # Lower bound of integration _lowerCAmelCase : Dict = 1.0 # Upper bound of integration _lowerCAmelCase : Optional[Any] = 10.0 # define number of steps or resolution _lowerCAmelCase : Optional[int] = [a, b] # define boundary of integration _lowerCAmelCase : List[Any] = method_a(_A , _A ) print(f'y = {y}' ) if __name__ == "__main__": main()
630
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Dict = { """configuration_nllb_moe""": [ """NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NllbMoeConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Union[str, Any] = [ """NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST""", """NllbMoeForConditionalGeneration""", """NllbMoeModel""", """NllbMoePreTrainedModel""", """NllbMoeTop2Router""", """NllbMoeSparseMLP""", ] if TYPE_CHECKING: from .configuration_nllb_moe import ( NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP, NllbMoeConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nllb_moe import ( NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST, NllbMoeForConditionalGeneration, NllbMoeModel, NllbMoePreTrainedModel, NllbMoeSparseMLP, NllbMoeTopaRouter, ) else: import sys lowerCAmelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
719
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) lowerCAmelCase : int = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys lowerCAmelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' def lowercase (_A = 5_0 ): """simple docstring""" _lowerCAmelCase : List[str] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(F'''{solution() = }''')
720
'''simple docstring''' from collections import Counter from timeit import timeit def lowercase (_A = "" , ): """simple docstring""" return sum(c % 2 for c in Counter(input_str.replace(' ' , '' ).lower() ).values() ) < 2 def lowercase (_A = "" ): """simple docstring""" if len(_A ) == 0: return True _lowerCAmelCase : Union[str, Any] = input_str.replace(' ' , '' ).lower() # character_freq_dict: Stores the frequency of every character in the input string _lowerCAmelCase : dict[str, int] = {} for character in lower_case_input_str: _lowerCAmelCase : Union[str, Any] = character_freq_dict.get(_A , 0 ) + 1 _lowerCAmelCase : List[Any] = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def lowercase (_A = "" ): """simple docstring""" print('\nFor string = ' , _A , ':' ) print( '> can_string_be_rearranged_as_palindrome_counter()' , '\tans =' , can_string_be_rearranged_as_palindrome_counter(_A ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome_counter(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) print( '> can_string_be_rearranged_as_palindrome()' , '\tans =' , can_string_be_rearranged_as_palindrome(_A ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) if __name__ == "__main__": lowerCAmelCase : Tuple = input( """Enter string to determine if it can be rearranged as a palindrome or not: """ ).strip() benchmark(check_str) lowerCAmelCase : Optional[Any] = can_string_be_rearranged_as_palindrome_counter(check_str) print(F'''{check_str} can {"" if status else "not "}be rearranged as a palindrome''')
630
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: lowerCAmelCase : Optional[int] = None lowerCAmelCase : List[Any] = logging.get_logger(__name__) lowerCAmelCase : str = """▁""" lowerCAmelCase : str = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCAmelCase : List[str] = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } lowerCAmelCase : Tuple = { """google/pegasus-xsum""": 5_12, } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = PegasusTokenizer __magic_name__ = ["input_ids", "attention_mask"] def __init__( self , snake_case__=None , snake_case__=None , snake_case__="<pad>" , snake_case__="</s>" , snake_case__="<unk>" , snake_case__="<mask_2>" , snake_case__="<mask_1>" , snake_case__=None , snake_case__=103 , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = offset if additional_special_tokens is not None: if not isinstance(snake_case__ , snake_case__ ): raise TypeError( F'additional_special_tokens should be of type {type(snake_case__ )}, but is' F' {type(snake_case__ )}' ) _lowerCAmelCase : Tuple = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F'<unk_{i}>' for i in range(len(snake_case__ ) , self.offset - 1 ) ] if len(set(snake_case__ ) ) != len(snake_case__ ): raise ValueError( 'Please make sure that the provided additional_special_tokens do not contain an incorrectly' F' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) _lowerCAmelCase : Any = additional_special_tokens_extended else: _lowerCAmelCase : int = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F'<unk_{i}>' for i in range(2 , self.offset )] super().__init__( snake_case__ , tokenizer_file=snake_case__ , pad_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , mask_token=snake_case__ , mask_token_sent=snake_case__ , offset=snake_case__ , additional_special_tokens=snake_case__ , **snake_case__ , ) _lowerCAmelCase : Union[str, Any] = vocab_file _lowerCAmelCase : str = False if not self.vocab_file else True def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( 'There should be 3 special tokens: mask_token, pad_token, and eos_token +' F' {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}' ) return [1 if x in all_special_ids else 0 for x in seq] def a ( self , snake_case__ , snake_case__ = None , snake_case__ = False ): '''simple docstring''' if already_has_special_tokens: return self._special_token_mask(snake_case__ ) elif token_ids_a is None: return self._special_token_mask(snake_case__ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def a ( self , snake_case__ , snake_case__=None ): '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCAmelCase : Union[str, Any] = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ): copyfile(self.vocab_file , snake_case__ ) return (out_vocab_file,)
721
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase : str = logging.get_logger(__name__) lowerCAmelCase : int = { """facebook/data2vec-text-base""": """https://huggingface.co/data2vec/resolve/main/config.json""", } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "data2vec-text" def __init__( self , snake_case__=3_0522 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=2 , snake_case__=0.02 , snake_case__=1E-12 , snake_case__=1 , snake_case__=0 , snake_case__=2 , snake_case__="absolute" , snake_case__=True , snake_case__=None , **snake_case__ , ): '''simple docstring''' super().__init__(pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ , **snake_case__ ) _lowerCAmelCase : List[Any] = vocab_size _lowerCAmelCase : Tuple = hidden_size _lowerCAmelCase : Dict = num_hidden_layers _lowerCAmelCase : int = num_attention_heads _lowerCAmelCase : str = hidden_act _lowerCAmelCase : Any = intermediate_size _lowerCAmelCase : Any = hidden_dropout_prob _lowerCAmelCase : Optional[int] = attention_probs_dropout_prob _lowerCAmelCase : str = max_position_embeddings _lowerCAmelCase : Any = type_vocab_size _lowerCAmelCase : int = initializer_range _lowerCAmelCase : List[str] = layer_norm_eps _lowerCAmelCase : List[Any] = position_embedding_type _lowerCAmelCase : str = use_cache _lowerCAmelCase : Union[str, Any] = classifier_dropout class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" @property def a ( self ): '''simple docstring''' if self.task == "multiple-choice": _lowerCAmelCase : Union[str, Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _lowerCAmelCase : List[Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
630
0
'''simple docstring''' def lowercase () -> List[Any]: """simple docstring""" return [list(range(1_0_0_0 - i , -1_0_0_0 - i , -1 ) ) for i in range(1_0_0_0 )] lowerCAmelCase : Union[str, Any] = generate_large_matrix() lowerCAmelCase : Any = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def lowercase (_A ) -> Dict: """simple docstring""" assert all(row == sorted(_A , reverse=_A ) for row in grid ) assert all(list(_A ) == sorted(_A , reverse=_A ) for col in zip(*_A ) ) def lowercase (_A ) -> Union[str, Any]: """simple docstring""" _lowerCAmelCase : str = 0 _lowerCAmelCase : Dict = len(_A ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _lowerCAmelCase : int = (left + right) // 2 _lowerCAmelCase : Any = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _lowerCAmelCase : int = mid + 1 else: _lowerCAmelCase : str = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(_A ) def lowercase (_A ) -> Optional[Any]: """simple docstring""" _lowerCAmelCase : Optional[Any] = 0 _lowerCAmelCase : List[str] = len(grid[0] ) for i in range(len(_A ) ): _lowerCAmelCase : Union[str, Any] = find_negative_index(grid[i][:bound] ) total += bound return (len(_A ) * len(grid[0] )) - total def lowercase (_A ) -> Dict: """simple docstring""" return len([number for row in grid for number in row if number < 0] ) def lowercase (_A ) -> Dict: """simple docstring""" _lowerCAmelCase : int = 0 for row in grid: for i, number in enumerate(_A ): if number < 0: total += len(_A ) - i break return total def lowercase () -> Dict: """simple docstring""" from timeit import timeit print('Running benchmarks' ) _lowerCAmelCase : Optional[Any] = ( 'from __main__ import count_negatives_binary_search, ' 'count_negatives_brute_force, count_negatives_brute_force_with_break, grid' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _lowerCAmelCase : Any = timeit(f'{func}(grid=grid)' , setup=_A , number=5_0_0 ) print(f'{func}() took {time:0.4f} seconds' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
700
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins lowerCAmelCase : List[str] = ["""tests.fixtures.files""", """tests.fixtures.hub""", """tests.fixtures.fsspec"""] def lowercase (_A , _A ): """simple docstring""" for item in items: if any(marker in item.keywords for marker in ['integration', 'unit'] ): continue item.add_marker(pytest.mark.unit ) def lowercase (_A ): """simple docstring""" config.addinivalue_line('markers' , 'torchaudio_latest: mark test to run with torchaudio>=0.12' ) @pytest.fixture(autouse=_A ) def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : str = tmp_path_factory.getbasetemp() / 'cache' _lowerCAmelCase : Dict = test_hf_cache_home / 'datasets' _lowerCAmelCase : List[Any] = test_hf_cache_home / 'metrics' _lowerCAmelCase : List[Any] = test_hf_cache_home / 'modules' monkeypatch.setattr('datasets.config.HF_DATASETS_CACHE' , str(_A ) ) monkeypatch.setattr('datasets.config.HF_METRICS_CACHE' , str(_A ) ) monkeypatch.setattr('datasets.config.HF_MODULES_CACHE' , str(_A ) ) _lowerCAmelCase : Dict = test_hf_datasets_cache / 'downloads' monkeypatch.setattr('datasets.config.DOWNLOADED_DATASETS_PATH' , str(_A ) ) _lowerCAmelCase : Union[str, Any] = test_hf_datasets_cache / 'downloads' / 'extracted' monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH' , str(_A ) ) @pytest.fixture(autouse=_A , scope='session' ) def lowercase (): """simple docstring""" datasets.disable_progress_bar() @pytest.fixture(autouse=_A ) def lowercase (_A ): """simple docstring""" monkeypatch.setattr('datasets.config.HF_UPDATE_DOWNLOAD_COUNTS' , _A ) @pytest.fixture def lowercase (_A ): """simple docstring""" monkeypatch.setattr('sqlalchemy.util.deprecations.SILENCE_UBER_WARNING' , _A )
630
0
import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed lowerCAmelCase : Any = { """distilbert""": (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), """roberta""": (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), """bert""": (BertConfig, BertForMaskedLM, BertTokenizer), """gpt2""": (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def lowercase (_A ): """simple docstring""" assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def lowercase (_A , _A ): """simple docstring""" if args.student_type == "roberta": _lowerCAmelCase : Tuple = False elif args.student_type == "gpt2": _lowerCAmelCase : Tuple = False def lowercase (_A , _A ): """simple docstring""" if args.student_type == "roberta": _lowerCAmelCase : Tuple = False def lowercase (): """simple docstring""" _lowerCAmelCase : Optional[int] = argparse.ArgumentParser(description='Training' ) parser.add_argument('--force' , action='store_true' , help='Overwrite dump_path if it already exists.' ) parser.add_argument( '--dump_path' , type=_A , required=_A , help='The output directory (log, checkpoints, parameters, etc.)' ) parser.add_argument( '--data_file' , type=_A , required=_A , help='The binarized file (tokenized + tokens_to_ids) and grouped by sequence.' , ) parser.add_argument( '--student_type' , type=_A , choices=['distilbert', 'roberta', 'gpt2'] , required=_A , help='The student type (DistilBERT, RoBERTa).' , ) parser.add_argument('--student_config' , type=_A , required=_A , help='Path to the student configuration.' ) parser.add_argument( '--student_pretrained_weights' , default=_A , type=_A , help='Load student initialization checkpoint.' ) parser.add_argument( '--teacher_type' , choices=['bert', 'roberta', 'gpt2'] , required=_A , help='Teacher type (BERT, RoBERTa).' ) parser.add_argument('--teacher_name' , type=_A , required=_A , help='The teacher model.' ) parser.add_argument('--temperature' , default=2.0 , type=_A , help='Temperature for the softmax temperature.' ) parser.add_argument( '--alpha_ce' , default=0.5 , type=_A , help='Linear weight for the distillation loss. Must be >=0.' ) parser.add_argument( '--alpha_mlm' , default=0.0 , type=_A , help='Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.' , ) parser.add_argument('--alpha_clm' , default=0.5 , type=_A , help='Linear weight for the CLM loss. Must be >=0.' ) parser.add_argument('--alpha_mse' , default=0.0 , type=_A , help='Linear weight of the MSE loss. Must be >=0.' ) parser.add_argument( '--alpha_cos' , default=0.0 , type=_A , help='Linear weight of the cosine embedding loss. Must be >=0.' ) parser.add_argument( '--mlm' , action='store_true' , help='The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.' ) parser.add_argument( '--mlm_mask_prop' , default=0.15 , type=_A , help='Proportion of tokens for which we need to make a prediction.' , ) parser.add_argument('--word_mask' , default=0.8 , type=_A , help='Proportion of tokens to mask out.' ) parser.add_argument('--word_keep' , default=0.1 , type=_A , help='Proportion of tokens to keep.' ) parser.add_argument('--word_rand' , default=0.1 , type=_A , help='Proportion of tokens to randomly replace.' ) parser.add_argument( '--mlm_smoothing' , default=0.7 , type=_A , help='Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).' , ) parser.add_argument('--token_counts' , type=_A , help='The token counts in the data_file for MLM.' ) parser.add_argument( '--restrict_ce_to_mask' , action='store_true' , help='If true, compute the distillation loss only the [MLM] prediction distribution.' , ) parser.add_argument( '--freeze_pos_embs' , action='store_true' , help='Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only.' , ) parser.add_argument( '--freeze_token_type_embds' , action='store_true' , help='Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only.' , ) parser.add_argument('--n_epoch' , type=_A , default=3 , help='Number of pass on the whole dataset.' ) parser.add_argument('--batch_size' , type=_A , default=5 , help='Batch size (for each process).' ) parser.add_argument( '--group_by_size' , action='store_false' , help='If true, group sequences that have similar length into the same batch. Default is true.' , ) parser.add_argument( '--gradient_accumulation_steps' , type=_A , default=5_0 , help='Gradient accumulation for larger training batches.' , ) parser.add_argument('--warmup_prop' , default=0.05 , type=_A , help='Linear warmup proportion.' ) parser.add_argument('--weight_decay' , default=0.0 , type=_A , help='Weight decay if we apply some.' ) parser.add_argument('--learning_rate' , default=5E-4 , type=_A , help='The initial learning rate for Adam.' ) parser.add_argument('--adam_epsilon' , default=1E-6 , type=_A , help='Epsilon for Adam optimizer.' ) parser.add_argument('--max_grad_norm' , default=5.0 , type=_A , help='Max gradient norm.' ) parser.add_argument('--initializer_range' , default=0.02 , type=_A , help='Random initialization range.' ) parser.add_argument( '--fp16' , action='store_true' , help='Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit' , ) parser.add_argument( '--fp16_opt_level' , type=_A , default='O1' , help=( 'For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].' 'See details at https://nvidia.github.io/apex/amp.html' ) , ) parser.add_argument('--n_gpu' , type=_A , default=1 , help='Number of GPUs in the node.' ) parser.add_argument('--local_rank' , type=_A , default=-1 , help='Distributed training - Local rank' ) parser.add_argument('--seed' , type=_A , default=5_6 , help='Random seed' ) parser.add_argument('--log_interval' , type=_A , default=5_0_0 , help='Tensorboard logging interval.' ) parser.add_argument('--checkpoint_interval' , type=_A , default=4_0_0_0 , help='Checkpoint interval.' ) _lowerCAmelCase : Tuple = parser.parse_args() sanity_checks(_A ) # ARGS # init_gpu_params(_A ) set_seed(_A ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( f'Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite' ' itUse `--force` if you want to overwrite it' ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(f'Experiment will be dumped and logged in {args.dump_path}' ) # SAVE PARAMS # logger.info(f'Param: {args}' ) with open(os.path.join(args.dump_path , 'parameters.json' ) , 'w' ) as f: json.dump(vars(_A ) , _A , indent=4 ) git_log(args.dump_path ) _lowerCAmelCase : List[Any] = MODEL_CLASSES[args.student_type] _lowerCAmelCase : Tuple = MODEL_CLASSES[args.teacher_type] # TOKENIZER # _lowerCAmelCase : int = teacher_tokenizer_class.from_pretrained(args.teacher_name ) _lowerCAmelCase : List[Any] = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): _lowerCAmelCase : Optional[Any] = tokenizer.all_special_tokens.index(_A ) _lowerCAmelCase : str = tokenizer.all_special_ids[idx] logger.info(f'Special tokens {special_tok_ids}' ) _lowerCAmelCase : Tuple = special_tok_ids _lowerCAmelCase : List[Any] = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f'Loading data from {args.data_file}' ) with open(args.data_file , 'rb' ) as fp: _lowerCAmelCase : Dict = pickle.load(_A ) if args.mlm: logger.info(f'Loading token counts from {args.token_counts} (already pre-computed)' ) with open(args.token_counts , 'rb' ) as fp: _lowerCAmelCase : int = pickle.load(_A ) _lowerCAmelCase : List[Any] = np.maximum(_A , 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): _lowerCAmelCase : Tuple = 0.0 # do not predict special tokens _lowerCAmelCase : Optional[Any] = torch.from_numpy(_A ) else: _lowerCAmelCase : Dict = None _lowerCAmelCase : List[str] = LmSeqsDataset(params=_A , data=_A ) logger.info('Data loader created.' ) # STUDENT # logger.info(f'Loading student config from {args.student_config}' ) _lowerCAmelCase : int = student_config_class.from_pretrained(args.student_config ) _lowerCAmelCase : List[Any] = True if args.student_pretrained_weights is not None: logger.info(f'Loading pretrained weights from {args.student_pretrained_weights}' ) _lowerCAmelCase : Optional[int] = student_model_class.from_pretrained(args.student_pretrained_weights , config=_A ) else: _lowerCAmelCase : Optional[Any] = student_model_class(_A ) if args.n_gpu > 0: student.to(f'cuda:{args.local_rank}' ) logger.info('Student loaded.' ) # TEACHER # _lowerCAmelCase : int = teacher_model_class.from_pretrained(args.teacher_name , output_hidden_states=_A ) if args.n_gpu > 0: teacher.to(f'cuda:{args.local_rank}' ) logger.info(f'Teacher loaded from {args.teacher_name}.' ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(_A , _A ) if args.freeze_token_type_embds: freeze_token_type_embeddings(_A , _A ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() _lowerCAmelCase : int = Distiller( params=_A , dataset=_A , token_probs=_A , student=_A , teacher=_A ) distiller.train() logger.info('Let\'s go get some drinks.' ) if __name__ == "__main__": main()
701
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCAmelCase : str = logging.get_logger(__name__) # General docstring lowerCAmelCase : Optional[Any] = """RegNetConfig""" # Base docstring lowerCAmelCase : int = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = [1, 10_88, 7, 7] # Image classification docstring lowerCAmelCase : Any = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = """tabby, tabby cat""" lowerCAmelCase : Tuple = [ """facebook/regnet-y-040""", # See all regnet models at https://huggingface.co/models?filter=regnet ] class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 3 , snake_case__ = 1 , snake_case__ = 1 , snake_case__ = "relu" , **snake_case__ , ): '''simple docstring''' super().__init__(**snake_case__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _lowerCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _lowerCAmelCase : List[Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=snake_case__ , strides=snake_case__ , padding='VALID' , groups=snake_case__ , use_bias=snake_case__ , name='convolution' , ) _lowerCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) _lowerCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.convolution(self.padding(snake_case__ ) ) _lowerCAmelCase : Union[str, Any] = self.normalization(snake_case__ ) _lowerCAmelCase : int = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = config.num_channels _lowerCAmelCase : List[Any] = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name='embedder' , ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = shape_list(snake_case__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( 'Make sure that the channel dimension of the pixel values match with the one set in the configuration.' ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _lowerCAmelCase : List[Any] = tf.transpose(snake_case__ , perm=(0, 2, 3, 1) ) _lowerCAmelCase : Tuple = self.embedder(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=1 , strides=snake_case__ , use_bias=snake_case__ , name='convolution' ) _lowerCAmelCase : Tuple = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' return self.normalization(self.convolution(snake_case__ ) , training=snake_case__ ) class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Any = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) _lowerCAmelCase : str = [ tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='relu' , name='attention.0' ), tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='sigmoid' , name='attention.2' ), ] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.pooler(snake_case__ ) for layer_module in self.attention: _lowerCAmelCase : Tuple = layer_module(snake_case__ ) _lowerCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Optional[int] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _lowerCAmelCase : Any = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.2' ), ] _lowerCAmelCase : List[str] = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = hidden_state for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) _lowerCAmelCase : int = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : Tuple = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : List[str] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) _lowerCAmelCase : Tuple = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetSELayer(snake_case__ , reduced_channels=int(round(in_channels / 4 ) ) , name='layer.2' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.3' ), ] _lowerCAmelCase : Tuple = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = hidden_state for layer_module in self.layers: _lowerCAmelCase : List[Any] = layer_module(snake_case__ ) _lowerCAmelCase : Tuple = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : str = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 2 , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Dict = TFRegNetXLayer if config.layer_type == 'x' else TFRegNetYLayer _lowerCAmelCase : Optional[int] = [ # downsampling is done in the first layer with stride of 2 layer(snake_case__ , snake_case__ , snake_case__ , stride=snake_case__ , name='layers.0' ), *[layer(snake_case__ , snake_case__ , snake_case__ , name=F'layers.{i+1}' ) for i in range(depth - 1 )], ] def a ( self , snake_case__ ): '''simple docstring''' for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( snake_case__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name='stages.0' , ) ) _lowerCAmelCase : Union[str, Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(snake_case__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(snake_case__ , snake_case__ , snake_case__ , depth=snake_case__ , name=F'stages.{i+1}' ) ) def a ( self , snake_case__ , snake_case__ = False , snake_case__ = True ): '''simple docstring''' _lowerCAmelCase : List[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _lowerCAmelCase : str = hidden_states + (hidden_state,) _lowerCAmelCase : List[str] = stage_module(snake_case__ ) if output_hidden_states: _lowerCAmelCase : Dict = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=snake_case__ , hidden_states=snake_case__ ) @keras_serializable class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" __magic_name__ = RegNetConfig def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = config _lowerCAmelCase : Union[str, Any] = TFRegNetEmbeddings(snake_case__ , name='embedder' ) _lowerCAmelCase : Optional[int] = TFRegNetEncoder(snake_case__ , name='encoder' ) _lowerCAmelCase : Dict = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) @unpack_inputs def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' _lowerCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : int = self.embedder(snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[str] = self.encoder( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[Any] = encoder_outputs[0] _lowerCAmelCase : Tuple = self.pooler(snake_case__ ) # Change to NCHW output format have uniformity in the modules _lowerCAmelCase : Optional[int] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) _lowerCAmelCase : Optional[Any] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _lowerCAmelCase : Union[str, Any] = tuple([tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=snake_case__ , pooler_output=snake_case__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = RegNetConfig __magic_name__ = "regnet" __magic_name__ = "pixel_values" @property def a ( self ): '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 224, 224) , dtype=tf.floataa )} lowerCAmelCase : List[Any] = r""" Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. """ lowerCAmelCase : Dict = r""" Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : List[str] = TFRegNetMainLayer(snake_case__ , name='regnet' ) @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Any = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : str = self.regnet( pixel_values=snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[Any] = config.num_labels _lowerCAmelCase : Optional[Any] = TFRegNetMainLayer(snake_case__ , name='regnet' ) # classification head _lowerCAmelCase : Optional[int] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name='classifier.1' ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def a ( self , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : Dict = self.regnet( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.pooler_output if return_dict else outputs[1] _lowerCAmelCase : List[Any] = self.classifier[0](snake_case__ ) _lowerCAmelCase : Tuple = self.classifier[1](snake_case__ ) _lowerCAmelCase : int = None if labels is None else self.hf_compute_loss(labels=snake_case__ , logits=snake_case__ ) if not return_dict: _lowerCAmelCase : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=snake_case__ , logits=snake_case__ , hidden_states=outputs.hidden_states )
630
0
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowerCAmelCase : Optional[int] = False class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : Union[str, Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) _lowerCAmelCase : List[Any] = torch.manual_seed(0 ) _lowerCAmelCase : List[str] = pipe.dual_guided( prompt='first prompt' , image=snake_case__ , text_to_image_strength=0.75 , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(snake_case__ ) _lowerCAmelCase : Optional[Any] = VersatileDiffusionPipeline.from_pretrained(snake_case__ , torch_dtype=torch.floataa ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : Tuple = generator.manual_seed(0 ) _lowerCAmelCase : List[Any] = pipe.dual_guided( prompt='first prompt' , image=snake_case__ , text_to_image_strength=0.75 , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = VersatileDiffusionPipeline.from_pretrained('shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : str = 'cyberpunk 2077' _lowerCAmelCase : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) _lowerCAmelCase : Optional[int] = torch.manual_seed(0 ) _lowerCAmelCase : Optional[int] = pipe.dual_guided( prompt=snake_case__ , image=snake_case__ , text_to_image_strength=0.75 , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' , ).images _lowerCAmelCase : Union[str, Any] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _lowerCAmelCase : int = np.array([0.1448, 0.1619, 0.1741, 0.1086, 0.1147, 0.1128, 0.1199, 0.1165, 0.1001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _lowerCAmelCase : Union[str, Any] = 'A painting of a squirrel eating a burger ' _lowerCAmelCase : Union[str, Any] = torch.manual_seed(0 ) _lowerCAmelCase : Optional[int] = pipe.text_to_image( prompt=snake_case__ , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _lowerCAmelCase : Optional[Any] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _lowerCAmelCase : Dict = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _lowerCAmelCase : int = pipe.image_variation(snake_case__ , generator=snake_case__ , output_type='numpy' ).images _lowerCAmelCase : List[Any] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _lowerCAmelCase : Dict = np.array([0.3076, 0.3123, 0.3284, 0.3782, 0.3770, 0.3894, 0.4297, 0.4331, 0.4456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
702
'''simple docstring''' from typing import Any def lowercase (_A ): """simple docstring""" if not input_list: return [] _lowerCAmelCase : Optional[int] = [input_list.count(_A ) for value in input_list] _lowerCAmelCase : int = max(_A ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(_A ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
630
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: lowerCAmelCase : Any = None lowerCAmelCase : Union[str, Any] = logging.get_logger(__name__) lowerCAmelCase : Optional[int] = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} lowerCAmelCase : Dict = { """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json""" ), }, } lowerCAmelCase : Optional[Any] = { """facebook/nllb-large-en-ro""": 10_24, """facebook/nllb-200-distilled-600M""": 10_24, } # fmt: off lowerCAmelCase : str = ["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = ["input_ids", "attention_mask"] __magic_name__ = NllbTokenizer __magic_name__ = [] __magic_name__ = [] def __init__( self , snake_case__=None , snake_case__=None , snake_case__="<s>" , snake_case__="</s>" , snake_case__="</s>" , snake_case__="<s>" , snake_case__="<unk>" , snake_case__="<pad>" , snake_case__="<mask>" , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=False , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Dict = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token _lowerCAmelCase : Any = legacy_behaviour super().__init__( vocab_file=snake_case__ , tokenizer_file=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , sep_token=snake_case__ , cls_token=snake_case__ , unk_token=snake_case__ , pad_token=snake_case__ , mask_token=snake_case__ , src_lang=snake_case__ , tgt_lang=snake_case__ , additional_special_tokens=snake_case__ , legacy_behaviour=snake_case__ , **snake_case__ , ) _lowerCAmelCase : Optional[Any] = vocab_file _lowerCAmelCase : Tuple = False if not self.vocab_file else True _lowerCAmelCase : List[str] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens} ) _lowerCAmelCase : Dict = { lang_code: self.convert_tokens_to_ids(snake_case__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } _lowerCAmelCase : str = src_lang if src_lang is not None else 'eng_Latn' _lowerCAmelCase : int = self.convert_tokens_to_ids(self._src_lang ) _lowerCAmelCase : int = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def a ( self ): '''simple docstring''' return self._src_lang @src_lang.setter def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : str = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : Dict = [self.sep_token_id] _lowerCAmelCase : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , **snake_case__ ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) _lowerCAmelCase : Union[str, Any] = src_lang _lowerCAmelCase : List[Any] = self(snake_case__ , add_special_tokens=snake_case__ , return_tensors=snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[int] = self.convert_tokens_to_ids(snake_case__ ) _lowerCAmelCase : List[Any] = tgt_lang_id return inputs def a ( self , snake_case__ , snake_case__ = "eng_Latn" , snake_case__ = None , snake_case__ = "fra_Latn" , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Dict = src_lang _lowerCAmelCase : Tuple = tgt_lang return super().prepare_seqaseq_batch(snake_case__ , snake_case__ , **snake_case__ ) def a ( self ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def a ( self ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.convert_tokens_to_ids(snake_case__ ) if self.legacy_behaviour: _lowerCAmelCase : List[Any] = [] _lowerCAmelCase : Dict = [self.eos_token_id, self.cur_lang_code] else: _lowerCAmelCase : Optional[int] = [self.cur_lang_code] _lowerCAmelCase : int = [self.eos_token_id] _lowerCAmelCase : Dict = self.convert_ids_to_tokens(self.prefix_tokens ) _lowerCAmelCase : Optional[int] = self.convert_ids_to_tokens(self.suffix_tokens ) _lowerCAmelCase : int = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = self.convert_tokens_to_ids(snake_case__ ) if self.legacy_behaviour: _lowerCAmelCase : List[str] = [] _lowerCAmelCase : Tuple = [self.eos_token_id, self.cur_lang_code] else: _lowerCAmelCase : int = [self.cur_lang_code] _lowerCAmelCase : List[Any] = [self.eos_token_id] _lowerCAmelCase : Optional[int] = self.convert_ids_to_tokens(self.prefix_tokens ) _lowerCAmelCase : Optional[int] = self.convert_ids_to_tokens(self.suffix_tokens ) _lowerCAmelCase : List[Any] = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory.' ) return _lowerCAmelCase : int = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ): copyfile(self.vocab_file , snake_case__ ) return (out_vocab_file,)
703
'''simple docstring''' from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
630
0
'''simple docstring''' def lowercase (_A ): """simple docstring""" if edge <= 0 or not isinstance(_A , _A ): raise ValueError('Length must be a positive.' ) return 3 * ((2_5 + 1_0 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def lowercase (_A ): """simple docstring""" if edge <= 0 or not isinstance(_A , _A ): raise ValueError('Length must be a positive.' ) return ((1_5 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
704
'''simple docstring''' import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase : int = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = {"""vocab_file""": """spiece.model"""} lowerCAmelCase : Optional[int] = { """vocab_file""": { """AI-Sweden/gpt-sw3-126m""": """https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-350m""": """https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-1.6b""": """https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-6.7b""": """https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-20b""": """https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model""", } } lowerCAmelCase : Union[str, Any] = { """AI-Sweden/gpt-sw3-126m""": 20_48, """AI-Sweden/gpt-sw3-350m""": 20_48, """AI-Sweden/gpt-sw3-1.6b""": 20_48, """AI-Sweden/gpt-sw3-6.7b""": 20_48, """AI-Sweden/gpt-sw3-20b""": 20_48, } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = ["input_ids", "attention_mask"] def __init__( self , snake_case__ , snake_case__=False , snake_case__=False , snake_case__=False , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Tuple = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase : List[Any] = kwargs.get('name_or_path' ) if name_or_path is None: logger.warning( 'name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,' ' you are testing the model, this can safely be ignored' ) _lowerCAmelCase : Any = 'None' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing _lowerCAmelCase : str = '<|endoftext|>' if eos_token is None else eos_token _lowerCAmelCase : Tuple = '<unk>' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: _lowerCAmelCase : List[str] = unk_token if pad_token is None else pad_token _lowerCAmelCase : Optional[int] = eos_token if bos_token is None else bos_token else: _lowerCAmelCase : Tuple = '<pad>' if pad_token is None else pad_token _lowerCAmelCase : Union[str, Any] = '<s>' if bos_token is None else bos_token super().__init__( do_lower_case=snake_case__ , remove_space=snake_case__ , keep_accents=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , pad_token=snake_case__ , sp_model_kwargs=self.sp_model_kwargs , **snake_case__ , ) _lowerCAmelCase : Union[str, Any] = do_lower_case _lowerCAmelCase : Optional[int] = remove_space _lowerCAmelCase : Any = keep_accents _lowerCAmelCase : Optional[int] = vocab_file _lowerCAmelCase : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case__ ) # Used for whitespace normalization in input texts # fmt : off _lowerCAmelCase : Optional[Any] = {' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', '', '„'} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing _lowerCAmelCase : Optional[Any] = re.compile( F'[{"".join(map(snake_case__ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]' ) def __getstate__( self ): '''simple docstring''' _lowerCAmelCase : List[str] = self.__dict__.copy() _lowerCAmelCase : int = None return state def __setstate__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowerCAmelCase : int = {} _lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def a ( self ): '''simple docstring''' return len(self.sp_model ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.non_printing_characters_re.sub('' , snake_case__ ) # Normalize whitespaces _lowerCAmelCase : Tuple = ''.join([char if char not in self.whitespaces else ' ' for char in text] ) # NFC Unicode normalization _lowerCAmelCase : Union[str, Any] = unicodedata.normalize('NFC' , snake_case__ ) return text def a ( self , snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : str = self.preprocess_text(snake_case__ ) return self.sp_model.encode(snake_case__ , out_type=snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.PieceToId(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case__ ) @staticmethod def a ( snake_case__ ): '''simple docstring''' return out_string def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = [] _lowerCAmelCase : Optional[Any] = '' _lowerCAmelCase : Tuple = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(snake_case__ ) + token _lowerCAmelCase : Union[str, Any] = True _lowerCAmelCase : List[Any] = [] else: current_sub_tokens.append(snake_case__ ) _lowerCAmelCase : List[Any] = False out_string += self.sp_model.decode(snake_case__ ) return out_string def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = {self.convert_ids_to_tokens(snake_case__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCAmelCase : int = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case__ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case__ , 'wb' ) as fi: _lowerCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (out_vocab_file,) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' if isinstance(snake_case__ , snake_case__ ): _lowerCAmelCase : Optional[Any] = self.preprocess_text(snake_case__ ) _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) else: _lowerCAmelCase : Tuple = [self.preprocess_text(snake_case__ ) for t in text] _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) if return_tensors is True or return_tensors == "pt": _lowerCAmelCase : int = torch.tensor(snake_case__ ) return token_ids def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.decode(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = [F'User: {text}' if is_user else F'Bot: {text}' for is_user, text in conversation.iter_texts()] _lowerCAmelCase : str = ( F'{self.eos_token}{self.bos_token}' + F'{self.bos_token}'.join(snake_case__ ) + F'{self.bos_token}Bot:' ) return self.encode(text=snake_case__ )
630
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetTokenizer else: lowerCAmelCase : Union[str, Any] = None lowerCAmelCase : Union[str, Any] = logging.get_logger(__name__) lowerCAmelCase : Tuple = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCAmelCase : Tuple = { """vocab_file""": { """xlnet-base-cased""": """https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model""", """xlnet-large-cased""": """https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model""", }, """tokenizer_file""": { """xlnet-base-cased""": """https://huggingface.co/xlnet-base-cased/resolve/main/tokenizer.json""", """xlnet-large-cased""": """https://huggingface.co/xlnet-large-cased/resolve/main/tokenizer.json""", }, } lowerCAmelCase : str = { """xlnet-base-cased""": None, """xlnet-large-cased""": None, } lowerCAmelCase : int = """▁""" # Segments (not really needed) lowerCAmelCase : Union[str, Any] = 0 lowerCAmelCase : Union[str, Any] = 1 lowerCAmelCase : Union[str, Any] = 2 lowerCAmelCase : Optional[Any] = 3 lowerCAmelCase : Dict = 4 class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = "left" __magic_name__ = XLNetTokenizer def __init__( self , snake_case__=None , snake_case__=None , snake_case__=False , snake_case__=True , snake_case__=False , snake_case__="<s>" , snake_case__="</s>" , snake_case__="<unk>" , snake_case__="<sep>" , snake_case__="<pad>" , snake_case__="<cls>" , snake_case__="<mask>" , snake_case__=["<eop>", "<eod>"] , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : List[str] = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token super().__init__( vocab_file=snake_case__ , tokenizer_file=snake_case__ , do_lower_case=snake_case__ , remove_space=snake_case__ , keep_accents=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , pad_token=snake_case__ , cls_token=snake_case__ , mask_token=snake_case__ , additional_special_tokens=snake_case__ , **snake_case__ , ) _lowerCAmelCase : Optional[int] = 3 _lowerCAmelCase : List[Any] = do_lower_case _lowerCAmelCase : Any = remove_space _lowerCAmelCase : str = keep_accents _lowerCAmelCase : Tuple = vocab_file _lowerCAmelCase : int = False if not self.vocab_file else True def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : List[Any] = [self.sep_token_id] _lowerCAmelCase : int = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : Any = [self.sep_token_id] _lowerCAmelCase : List[str] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCAmelCase : List[Any] = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ): copyfile(self.vocab_file , snake_case__ ) return (out_vocab_file,)
705
'''simple docstring''' import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = (DDPMScheduler,) def a ( self , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**snake_case__ ) return config def a ( self ): '''simple docstring''' for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=snake_case__ ) def a ( self ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=snake_case__ , beta_end=snake_case__ ) def a ( self ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=snake_case__ ) def a ( self ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=snake_case__ ) def a ( self ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=snake_case__ ) def a ( self ): '''simple docstring''' self.check_over_configs(thresholding=snake_case__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=snake_case__ , prediction_type=snake_case__ , sample_max_value=snake_case__ , ) def a ( self ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=snake_case__ ) def a ( self ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1E-5 def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : str = self.dummy_model() _lowerCAmelCase : str = self.dummy_sample_deter _lowerCAmelCase : Union[str, Any] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : List[Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Any = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Dict = pred_prev_sample _lowerCAmelCase : Dict = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : List[str] = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 258.9606 ) < 1E-2 assert abs(result_mean.item() - 0.3372 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.scheduler_classes[0] _lowerCAmelCase : Any = self.get_scheduler_config(prediction_type='v_prediction' ) _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : Any = self.dummy_model() _lowerCAmelCase : Tuple = self.dummy_sample_deter _lowerCAmelCase : Optional[int] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : Union[str, Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Dict = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Tuple = pred_prev_sample _lowerCAmelCase : Any = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : Dict = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 202.0296 ) < 1E-2 assert abs(result_mean.item() - 0.2631 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[int] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=snake_case__ ) _lowerCAmelCase : Union[str, Any] = scheduler.timesteps for i, timestep in enumerate(snake_case__ ): if i == len(snake_case__ ) - 1: _lowerCAmelCase : str = -1 else: _lowerCAmelCase : Optional[Any] = timesteps[i + 1] _lowerCAmelCase : int = scheduler.previous_timestep(snake_case__ ) _lowerCAmelCase : int = prev_t.item() self.assertEqual(snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : Tuple = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 51, 0] with self.assertRaises(snake_case__ , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : List[str] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 1, 0] _lowerCAmelCase : int = len(snake_case__ ) with self.assertRaises(snake_case__ , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=snake_case__ , timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : int = self.get_scheduler_config() _lowerCAmelCase : Any = scheduler_class(**snake_case__ ) _lowerCAmelCase : Any = [scheduler.config.num_train_timesteps] with self.assertRaises( snake_case__ , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=snake_case__ )
630
0
'''simple docstring''' from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = 5_0257 , snake_case__ = 1024 , snake_case__ = 768 , snake_case__ = 12 , snake_case__ = 12 , snake_case__ = None , snake_case__ = "gelu_new" , snake_case__ = 0.1 , snake_case__ = 0.1 , snake_case__ = 0.1 , snake_case__ = 1E-5 , snake_case__ = 0.02 , snake_case__ = True , snake_case__ = True , snake_case__ = False , snake_case__ = False , ): '''simple docstring''' super().__init__() _lowerCAmelCase : Any = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F'`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and' F' `n_embd`: {n_embd} are not equal.' ) _lowerCAmelCase : Any = prefix_inner_dim _lowerCAmelCase : Any = prefix_hidden_dim _lowerCAmelCase : str = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCAmelCase : Optional[Any] = ( nn.Linear(self.prefix_hidden_dim , snake_case__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCAmelCase : Union[str, Any] = GPTaConfig( vocab_size=snake_case__ , n_positions=snake_case__ , n_embd=snake_case__ , n_layer=snake_case__ , n_head=snake_case__ , n_inner=snake_case__ , activation_function=snake_case__ , resid_pdrop=snake_case__ , embd_pdrop=snake_case__ , attn_pdrop=snake_case__ , layer_norm_epsilon=snake_case__ , initializer_range=snake_case__ , scale_attn_weights=snake_case__ , use_cache=snake_case__ , scale_attn_by_inverse_layer_idx=snake_case__ , reorder_and_upcast_attn=snake_case__ , ) _lowerCAmelCase : List[str] = GPTaLMHeadModel(snake_case__ ) def a ( self , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = None , ): '''simple docstring''' _lowerCAmelCase : str = self.transformer.transformer.wte(snake_case__ ) _lowerCAmelCase : str = self.encode_prefix(snake_case__ ) _lowerCAmelCase : Tuple = self.decode_prefix(snake_case__ ) _lowerCAmelCase : Optional[int] = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: _lowerCAmelCase : List[str] = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) _lowerCAmelCase : Union[str, Any] = torch.cat((dummy_token, input_ids) , dim=1 ) _lowerCAmelCase : int = self.transformer(inputs_embeds=snake_case__ , labels=snake_case__ , attention_mask=snake_case__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' return torch.zeros(snake_case__ , self.prefix_length , dtype=torch.intaa , device=snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.encode_prefix(snake_case__ ) @torch.no_grad() def a ( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = torch.split(snake_case__ , 1 , dim=0 ) _lowerCAmelCase : Dict = [] _lowerCAmelCase : str = [] for feature in features: _lowerCAmelCase : List[Any] = self.decode_prefix(feature.to(snake_case__ ) ) # back to the clip feature # Only support beam search for now _lowerCAmelCase : List[Any] = self.generate_beam( input_embeds=snake_case__ , device=snake_case__ , eos_token_id=snake_case__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) _lowerCAmelCase : Tuple = torch.stack(snake_case__ ) _lowerCAmelCase : List[Any] = torch.stack(snake_case__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def a ( self , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__ = 5 , snake_case__ = 67 , snake_case__ = 1.0 , snake_case__ = None , ): '''simple docstring''' _lowerCAmelCase : List[Any] = eos_token_id _lowerCAmelCase : List[Any] = None _lowerCAmelCase : Optional[int] = None _lowerCAmelCase : Optional[Any] = torch.ones(snake_case__ , device=snake_case__ , dtype=torch.int ) _lowerCAmelCase : Optional[Any] = torch.zeros(snake_case__ , device=snake_case__ , dtype=torch.bool ) if input_embeds is not None: _lowerCAmelCase : int = input_embeds else: _lowerCAmelCase : List[Any] = self.transformer.transformer.wte(snake_case__ ) for i in range(snake_case__ ): _lowerCAmelCase : str = self.transformer(inputs_embeds=snake_case__ ) _lowerCAmelCase : Tuple = outputs.logits _lowerCAmelCase : Optional[Any] = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) _lowerCAmelCase : Any = logits.softmax(-1 ).log() if scores is None: _lowerCAmelCase : Dict = logits.topk(snake_case__ , -1 ) _lowerCAmelCase : Optional[Any] = generated.expand(snake_case__ , *generated.shape[1:] ) _lowerCAmelCase : Dict = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: _lowerCAmelCase : Dict = next_tokens else: _lowerCAmelCase : str = tokens.expand(snake_case__ , *tokens.shape[1:] ) _lowerCAmelCase : Optional[int] = torch.cat((tokens, next_tokens) , dim=1 ) else: _lowerCAmelCase : Any = -float(np.inf ) _lowerCAmelCase : Union[str, Any] = 0 _lowerCAmelCase : int = scores[:, None] + logits seq_lengths[~is_stopped] += 1 _lowerCAmelCase : Any = scores_sum / seq_lengths[:, None] _lowerCAmelCase : List[Any] = scores_sum_average.view(-1 ).topk(snake_case__ , -1 ) _lowerCAmelCase : List[str] = next_tokens // scores_sum.shape[1] _lowerCAmelCase : Tuple = seq_lengths[next_tokens_source] _lowerCAmelCase : Optional[int] = next_tokens % scores_sum.shape[1] _lowerCAmelCase : Dict = next_tokens.unsqueeze(1 ) _lowerCAmelCase : Dict = tokens[next_tokens_source] _lowerCAmelCase : Dict = torch.cat((tokens, next_tokens) , dim=1 ) _lowerCAmelCase : Optional[int] = generated[next_tokens_source] _lowerCAmelCase : Optional[int] = scores_sum_average * seq_lengths _lowerCAmelCase : List[Any] = is_stopped[next_tokens_source] _lowerCAmelCase : Any = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) _lowerCAmelCase : Dict = torch.cat((generated, next_token_embed) , dim=1 ) _lowerCAmelCase : Optional[Any] = is_stopped + next_tokens.eq(snake_case__ ).squeeze() if is_stopped.all(): break _lowerCAmelCase : Dict = scores / seq_lengths _lowerCAmelCase : List[str] = scores.argsort(descending=snake_case__ ) # tokens tensors are already padded to max_seq_length _lowerCAmelCase : Optional[int] = [tokens[i] for i in order] _lowerCAmelCase : Optional[int] = torch.stack(snake_case__ , dim=0 ) _lowerCAmelCase : str = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
706
'''simple docstring''' import socket def lowercase (): """simple docstring""" _lowerCAmelCase : Tuple = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCAmelCase : Optional[int] = socket.gethostname() _lowerCAmelCase : Any = 1_2_3_1_2 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCAmelCase : Union[str, Any] = sock.recv(1_0_2_4 ) if not data: break out_file.write(_A ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
630
0
'''simple docstring''' from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowerCAmelCase : Any = input("""Enter image url: """).strip() print(F'''Downloading image from {url} ...''') lowerCAmelCase : Union[str, Any] = BeautifulSoup(requests.get(url).content, """html.parser""") # The image URL is in the content field of the first meta tag with property og:image lowerCAmelCase : Optional[int] = soup.find("""meta""", {"""property""": """og:image"""})["""content"""] lowerCAmelCase : Dict = requests.get(image_url).content lowerCAmelCase : int = F'''{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg''' with open(file_name, """wb""") as fp: fp.write(image_data) print(F'''Done. Image saved to disk as {file_name}.''')
707
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase : Tuple = False lowerCAmelCase : str = True lowerCAmelCase : List[Any] = False if __name__ == "__main__": lowerCAmelCase : Any = argparse.ArgumentParser() parser.add_argument( """--repo_path""", default=None, type=str, required=True, help="""The config json file corresponding to the architecture.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") lowerCAmelCase : Optional[int] = parser.parse_args() lowerCAmelCase : int = { """image_size""": """sample_size""", """num_res_blocks""": """layers_per_block""", """block_channels""": """block_out_channels""", """down_blocks""": """down_block_types""", """up_blocks""": """up_block_types""", """downscale_freq_shift""": """freq_shift""", """resnet_num_groups""": """norm_num_groups""", """resnet_act_fn""": """act_fn""", """resnet_eps""": """norm_eps""", """num_head_channels""": """attention_head_dim""", } lowerCAmelCase : int = { """time_steps""": """time_proj""", """mid""": """mid_block""", """downsample_blocks""": """down_blocks""", """upsample_blocks""": """up_blocks""", } lowerCAmelCase : Optional[Any] = """""" if has_file(args.repo_path, """config.json""") else """unet""" with open(os.path.join(args.repo_path, subfolder, """config.json"""), """r""", encoding="""utf-8""") as reader: lowerCAmelCase : int = reader.read() lowerCAmelCase : List[str] = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, """config.json"""): lowerCAmelCase : str = UNetaDModel(**config) else: lowerCAmelCase : Union[str, Any] = UNetaDConditionModel if """ldm-text2im-large-256""" in args.repo_path else UNetaDModel lowerCAmelCase : Dict = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase : Union[str, Any] = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase : str = config[key] del config[key] lowerCAmelCase : Optional[int] = [k.replace("""UNetRes""", """""") for k in config["""down_block_types"""]] lowerCAmelCase : Dict = [k.replace("""UNetRes""", """""") for k in config["""up_block_types"""]] if do_only_weights: lowerCAmelCase : Tuple = torch.load(os.path.join(args.repo_path, subfolder, """diffusion_pytorch_model.bin""")) lowerCAmelCase : str = {} for param_key, param_value in state_dict.items(): if param_key.endswith(""".op.bias""") or param_key.endswith(""".op.weight"""): continue lowerCAmelCase : str = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split(""".""")[0] == key: lowerCAmelCase : Dict = param_value lowerCAmelCase : Tuple = True if not has_changed: lowerCAmelCase : Tuple = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
630
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) lowerCAmelCase : Dict = { """configuration_swiftformer""": [ """SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwiftFormerConfig""", """SwiftFormerOnnxConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[int] = [ """SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwiftFormerForImageClassification""", """SwiftFormerModel""", """SwiftFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys lowerCAmelCase : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
708
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = None ): '''simple docstring''' super().__init__() _lowerCAmelCase : Union[str, Any] = pad_token_id _lowerCAmelCase : List[Any] = max_length _lowerCAmelCase : Tuple = vocab _lowerCAmelCase : str = merges _lowerCAmelCase : List[str] = BytePairTokenizer(snake_case__ , snake_case__ , sequence_length=snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = [' '.join(snake_case__ ) for m in tokenizer.bpe_ranks.keys()] _lowerCAmelCase : Any = tokenizer.get_vocab() return cls(snake_case__ , snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = GPTaTokenizer.from_pretrained(snake_case__ , *snake_case__ , **snake_case__ ) return cls.from_tokenizer(snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ ): '''simple docstring''' return cls(**snake_case__ ) def a ( self ): '''simple docstring''' return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : str = self.tf_tokenizer(snake_case__ ) _lowerCAmelCase : str = tf.ones_like(snake_case__ ) if self.pad_token_id is not None: # pad the tokens up to max length _lowerCAmelCase : Optional[int] = max_length if max_length is not None else self.max_length if max_length is not None: _lowerCAmelCase , _lowerCAmelCase : str = pad_model_inputs( snake_case__ , max_seq_length=snake_case__ , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
630
0
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_5_0, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "pytorch", "script": "run_ddp.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_0_0, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "tensorflow", "script": "run_tf_dist.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_0_0, "eval_accuracy": 0.6, "eval_loss": 0.7}, }, ] ) class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' if self.framework == "pytorch": subprocess.run( F'cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'.split() , encoding='utf-8' , check=snake_case__ , ) assert hasattr(self , 'env' ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[str] = F'{self.env.base_job_name}-{instance_count}-{"ddp" if "ddp" in self.script else "smd"}' # distributed data settings _lowerCAmelCase : str = {'smdistributed': {'dataparallel': {'enabled': True}}} if self.script != 'run_ddp.py' else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=snake_case__ , instance_count=snake_case__ , instance_type=self.instance_type , debugger_hook_config=snake_case__ , hyperparameters={**self.env.distributed_hyperparameters, 'model_name_or_path': self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=snake_case__ , py_version='py36' , ) def a ( self , snake_case__ ): '''simple docstring''' TrainingJobAnalytics(snake_case__ ).export_csv(F'{self.env.test_path}/{job_name}_metrics.csv' ) @parameterized.expand([(2,)] ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = self.create_estimator(snake_case__ ) # run training estimator.fit() # result dataframe _lowerCAmelCase : Any = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _lowerCAmelCase : int = list(result_metrics_df[result_metrics_df.metric_name == 'eval_accuracy']['value'] ) _lowerCAmelCase : Any = list(result_metrics_df[result_metrics_df.metric_name == 'eval_loss']['value'] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _lowerCAmelCase : Dict = ( Session().describe_training_job(estimator.latest_training_job.name ).get('TrainingTimeInSeconds' , 99_9999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['eval_accuracy'] for t in eval_accuracy ) assert all(t <= self.results['eval_loss'] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'{estimator.latest_training_job.name}.json' , 'w' ) as outfile: json.dump({'train_time': train_runtime, 'eval_accuracy': eval_accuracy, 'eval_loss': eval_loss} , snake_case__ )
709
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Optional[int] = { """configuration_mctct""": ["""MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MCTCTConfig"""], """feature_extraction_mctct""": ["""MCTCTFeatureExtractor"""], """processing_mctct""": ["""MCTCTProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : List[str] = [ """MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MCTCTForCTC""", """MCTCTModel""", """MCTCTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys lowerCAmelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : List[Any] = logging.get_logger(__name__) lowerCAmelCase : int = { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json""" ), """facebook/dpr-question_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json""" ), """facebook/dpr-reader-single-nq-base""": ( """https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json""" ), """facebook/dpr-ctx_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json""" ), """facebook/dpr-question_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json""" ), """facebook/dpr-reader-multiset-base""": ( """https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json""" ), } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "dpr" def __init__( self , snake_case__=3_0522 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=2 , snake_case__=0.02 , snake_case__=1E-12 , snake_case__=0 , snake_case__="absolute" , snake_case__ = 0 , **snake_case__ , ): '''simple docstring''' super().__init__(pad_token_id=snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[Any] = vocab_size _lowerCAmelCase : Union[str, Any] = hidden_size _lowerCAmelCase : Tuple = num_hidden_layers _lowerCAmelCase : Union[str, Any] = num_attention_heads _lowerCAmelCase : Union[str, Any] = hidden_act _lowerCAmelCase : str = intermediate_size _lowerCAmelCase : Optional[Any] = hidden_dropout_prob _lowerCAmelCase : Any = attention_probs_dropout_prob _lowerCAmelCase : Any = max_position_embeddings _lowerCAmelCase : Optional[int] = type_vocab_size _lowerCAmelCase : str = initializer_range _lowerCAmelCase : Tuple = layer_norm_eps _lowerCAmelCase : str = projection_dim _lowerCAmelCase : int = position_embedding_type
710
'''simple docstring''' lowerCAmelCase : Optional[int] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_00_00)] def lowercase (_A ): """simple docstring""" _lowerCAmelCase : str = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 1_0_0_0_0_0] number //= 1_0_0_0_0_0 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution lowerCAmelCase : list[bool | None] = [None] * 10_00_00_00 lowerCAmelCase : List[str] = True lowerCAmelCase : Union[str, Any] = False def lowercase (_A ): """simple docstring""" if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _lowerCAmelCase : Any = chain(next_number(_A ) ) _lowerCAmelCase : List[str] = number_chain while number < 1_0_0_0_0_0_0_0: _lowerCAmelCase : Tuple = number_chain number *= 1_0 return number_chain def lowercase (_A = 1_0_0_0_0_0_0_0 ): """simple docstring""" for i in range(1 , _A ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(_A ) if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution() = }''')
630
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : List[str] = logging.get_logger(__name__) lowerCAmelCase : Optional[Any] = { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/config.json""", """funnel-transformer/small-base""": """https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json""", """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/config.json""", """funnel-transformer/medium-base""": """https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json""", """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/config.json""", """funnel-transformer/large-base""": """https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json""", """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json""", """funnel-transformer/xlarge-base""": """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json""", } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "funnel" __magic_name__ = { "hidden_size": "d_model", "num_attention_heads": "n_head", } def __init__( self , snake_case__=3_0522 , snake_case__=[4, 4, 4] , snake_case__=None , snake_case__=2 , snake_case__=768 , snake_case__=12 , snake_case__=64 , snake_case__=3072 , snake_case__="gelu_new" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.0 , snake_case__=0.1 , snake_case__=None , snake_case__=1E-9 , snake_case__="mean" , snake_case__="relative_shift" , snake_case__=True , snake_case__=True , snake_case__=True , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = vocab_size _lowerCAmelCase : List[Any] = block_sizes _lowerCAmelCase : int = [1] * len(snake_case__ ) if block_repeats is None else block_repeats assert len(snake_case__ ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." _lowerCAmelCase : Tuple = num_decoder_layers _lowerCAmelCase : List[Any] = d_model _lowerCAmelCase : Optional[int] = n_head _lowerCAmelCase : List[str] = d_head _lowerCAmelCase : Union[str, Any] = d_inner _lowerCAmelCase : List[str] = hidden_act _lowerCAmelCase : List[Any] = hidden_dropout _lowerCAmelCase : Union[str, Any] = attention_dropout _lowerCAmelCase : Optional[int] = activation_dropout _lowerCAmelCase : str = initializer_range _lowerCAmelCase : int = initializer_std _lowerCAmelCase : Dict = layer_norm_eps assert pooling_type in [ "mean", "max", ], F'Got {pooling_type} for `pooling_type` but only \'mean\' and \'max\' are supported.' _lowerCAmelCase : str = pooling_type assert attention_type in [ "relative_shift", "factorized", ], F'Got {attention_type} for `attention_type` but only \'relative_shift\' and \'factorized\' are supported.' _lowerCAmelCase : List[Any] = attention_type _lowerCAmelCase : int = separate_cls _lowerCAmelCase : Union[str, Any] = truncate_seq _lowerCAmelCase : List[Any] = pool_q_only super().__init__(**snake_case__ ) @property def a ( self ): '''simple docstring''' return sum(self.block_sizes ) @num_hidden_layers.setter def a ( self , snake_case__ ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.' ) @property def a ( self ): '''simple docstring''' return len(self.block_sizes ) @num_blocks.setter def a ( self , snake_case__ ): '''simple docstring''' raise NotImplementedError('This model does not support the setting of `num_blocks`. Please set `block_sizes`.' )
711
'''simple docstring''' import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(snake_case__ , 'width_multiplier' ) ) class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=64 , snake_case__=2 , snake_case__=3 , snake_case__="swish" , snake_case__=3 , snake_case__=32 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=True , snake_case__=True , snake_case__=10 , snake_case__=None , snake_case__=0.25 , snake_case__=0.0 , snake_case__=0.0 , ): '''simple docstring''' _lowerCAmelCase : Tuple = parent _lowerCAmelCase : Optional[int] = batch_size _lowerCAmelCase : List[Any] = image_size _lowerCAmelCase : List[Any] = patch_size _lowerCAmelCase : List[str] = num_channels _lowerCAmelCase : Optional[Any] = make_divisible(512 * width_multiplier , divisor=8 ) _lowerCAmelCase : Optional[Any] = hidden_act _lowerCAmelCase : List[Any] = conv_kernel_size _lowerCAmelCase : Optional[Any] = output_stride _lowerCAmelCase : List[Any] = classifier_dropout_prob _lowerCAmelCase : str = use_labels _lowerCAmelCase : List[str] = is_training _lowerCAmelCase : Optional[int] = num_labels _lowerCAmelCase : List[str] = initializer_range _lowerCAmelCase : str = scope _lowerCAmelCase : Any = width_multiplier _lowerCAmelCase : Union[str, Any] = ffn_dropout _lowerCAmelCase : Optional[int] = attn_dropout def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase : Optional[Any] = None _lowerCAmelCase : Dict = None if self.use_labels: _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _lowerCAmelCase : Optional[int] = self.get_config() return config, pixel_values, labels, pixel_labels def a ( self ): '''simple docstring''' return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : str = model(snake_case__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.num_labels _lowerCAmelCase : List[Any] = MobileViTVaForImageClassification(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : int = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.num_labels _lowerCAmelCase : Optional[int] = MobileViTVaForSemanticSegmentation(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Dict = model(snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) _lowerCAmelCase : Any = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Optional[int] = config_and_inputs _lowerCAmelCase : Tuple = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = ( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) __magic_name__ = ( { "feature-extraction": MobileViTVaModel, "image-classification": MobileViTVaForImageClassification, "image-segmentation": MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False def a ( self ): '''simple docstring''' _lowerCAmelCase : int = MobileViTVaModelTester(self ) _lowerCAmelCase : Dict = MobileViTVaConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ ) def a ( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='MobileViTV2 does not use inputs_embeds' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not support input and output embeddings' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not output attentions' ) def a ( self ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason='Got `CUDA error: misaligned address` for tests after this one being run.' ) def a ( self ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : str = model_class(snake_case__ ) _lowerCAmelCase : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCAmelCase : int = [*signature.parameters.keys()] _lowerCAmelCase : Any = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def a ( self ): '''simple docstring''' def check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ): _lowerCAmelCase : Dict = model_class(snake_case__ ) model.to(snake_case__ ) model.eval() with torch.no_grad(): _lowerCAmelCase : Optional[Any] = model(**self._prepare_for_class(snake_case__ , snake_case__ ) ) _lowerCAmelCase : List[str] = outputs.hidden_states _lowerCAmelCase : List[str] = 5 self.assertEqual(len(snake_case__ ) , snake_case__ ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. _lowerCAmelCase : List[Any] = 2 for i in range(len(snake_case__ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : Optional[int] = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCAmelCase : Any = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*snake_case__ ) @slow def a ( self ): '''simple docstring''' for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase : Dict = MobileViTVaModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) def lowercase (): """simple docstring""" _lowerCAmelCase : Any = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def a ( self ): '''simple docstring''' return ( MobileViTImageProcessor.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ) if is_vision_available() else None ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = MobileViTVaForImageClassification.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ).to( snake_case__ ) _lowerCAmelCase : str = self.default_image_processor _lowerCAmelCase : Any = prepare_img() _lowerCAmelCase : Optional[int] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Tuple = model(**snake_case__ ) # verify the logits _lowerCAmelCase : Optional[Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , snake_case__ ) _lowerCAmelCase : Tuple = torch.tensor([-1.6336E00, -7.3204E-02, -5.1883E-01] ).to(snake_case__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Any = model.to(snake_case__ ) _lowerCAmelCase : int = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Optional[int] = prepare_img() _lowerCAmelCase : Dict = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : int = model(**snake_case__ ) _lowerCAmelCase : Dict = outputs.logits # verify the logits _lowerCAmelCase : str = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , snake_case__ ) _lowerCAmelCase : Any = torch.tensor( [ [[7.0863, 7.1525, 6.8201], [6.6931, 6.8770, 6.8933], [6.2978, 7.0366, 6.9636]], [[-3.7134, -3.6712, -3.6675], [-3.5825, -3.3549, -3.4777], [-3.3435, -3.3979, -3.2857]], [[-2.9329, -2.8003, -2.7369], [-3.0564, -2.4780, -2.0207], [-2.6889, -1.9298, -1.7640]], ] , device=snake_case__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : List[Any] = model.to(snake_case__ ) _lowerCAmelCase : str = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Tuple = prepare_img() _lowerCAmelCase : List[str] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Any = model(**snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.logits.detach().cpu() _lowerCAmelCase : Any = image_processor.post_process_semantic_segmentation(outputs=snake_case__ , target_sizes=[(50, 60)] ) _lowerCAmelCase : List[Any] = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , snake_case__ ) _lowerCAmelCase : List[str] = image_processor.post_process_semantic_segmentation(outputs=snake_case__ ) _lowerCAmelCase : Tuple = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , snake_case__ )
630
0
'''simple docstring''' from __future__ import annotations from random import random from typing import Generic, TypeVar lowerCAmelCase : Tuple = TypeVar("""KT""") lowerCAmelCase : Dict = TypeVar("""VT""") class UpperCamelCase__ ( Generic[KT, VT] ): """simple docstring""" def __init__( self , snake_case__ = "root" , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : Tuple = key _lowerCAmelCase : Union[str, Any] = value _lowerCAmelCase : list[Node[KT, VT]] = [] def __repr__( self ): '''simple docstring''' return F'Node({self.key}: {self.value})' @property def a ( self ): '''simple docstring''' return len(self.forward ) class UpperCamelCase__ ( Generic[KT, VT] ): """simple docstring""" def __init__( self , snake_case__ = 0.5 , snake_case__ = 16 ): '''simple docstring''' _lowerCAmelCase : Node[KT, VT] = Node[KT, VT]() _lowerCAmelCase : Optional[Any] = 0 _lowerCAmelCase : int = p _lowerCAmelCase : Tuple = max_level def __str__( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = list(self ) if len(snake_case__ ) == 0: return F'SkipList(level={self.level})' _lowerCAmelCase : List[Any] = max((len(str(snake_case__ ) ) for item in items) , default=4 ) _lowerCAmelCase : Tuple = max(snake_case__ , 4 ) + 4 _lowerCAmelCase : List[Any] = self.head _lowerCAmelCase : str = [] _lowerCAmelCase : Union[str, Any] = node.forward.copy() lines.append(F'[{node.key}]'.ljust(snake_case__ , '-' ) + '* ' * len(snake_case__ ) ) lines.append(' ' * label_size + '| ' * len(snake_case__ ) ) while len(node.forward ) != 0: _lowerCAmelCase : Union[str, Any] = node.forward[0] lines.append( F'[{node.key}]'.ljust(snake_case__ , '-' ) + ' '.join(str(n.key ) if n.key == node.key else '|' for n in forwards ) ) lines.append(' ' * label_size + '| ' * len(snake_case__ ) ) _lowerCAmelCase : Dict = node.forward lines.append('None'.ljust(snake_case__ ) + '* ' * len(snake_case__ ) ) return F'SkipList(level={self.level})\n' + "\n".join(snake_case__ ) def __iter__( self ): '''simple docstring''' _lowerCAmelCase : Any = self.head while len(node.forward ) != 0: yield node.forward[0].key _lowerCAmelCase : str = node.forward[0] def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = 1 while random() < self.p and level < self.max_level: level += 1 return level def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [] _lowerCAmelCase : int = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: _lowerCAmelCase : int = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(snake_case__ ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : str = self._locate_node(snake_case__ ) if node is not None: for i, update_node in enumerate(snake_case__ ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: _lowerCAmelCase : Optional[int] = node.forward[i] else: _lowerCAmelCase : Optional[Any] = update_node.forward[:i] def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self._locate_node(snake_case__ ) if node is not None: _lowerCAmelCase : Union[str, Any] = value else: _lowerCAmelCase : str = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , snake_case__ ): update_vector.append(self.head ) _lowerCAmelCase : List[Any] = level _lowerCAmelCase : Optional[int] = Node(snake_case__ , snake_case__ ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(snake_case__ ) else: _lowerCAmelCase : Any = new_node def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = self._locate_node(snake_case__ ) if node is not None: return node.value return None def lowercase (): """simple docstring""" _lowerCAmelCase : Any = SkipList() skip_list.insert('Key1' , 3 ) skip_list.insert('Key2' , 1_2 ) skip_list.insert('Key3' , 4_1 ) skip_list.insert('Key4' , -1_9 ) _lowerCAmelCase : Any = skip_list.head _lowerCAmelCase : List[str] = {} while node.level != 0: _lowerCAmelCase : str = node.forward[0] _lowerCAmelCase : str = node.value assert len(_A ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 1_2 assert all_values["Key3"] == 4_1 assert all_values["Key4"] == -1_9 def lowercase (): """simple docstring""" _lowerCAmelCase : Union[str, Any] = SkipList() skip_list.insert('Key1' , 1_0 ) skip_list.insert('Key1' , 1_2 ) skip_list.insert('Key5' , 7 ) skip_list.insert('Key7' , 1_0 ) skip_list.insert('Key10' , 5 ) skip_list.insert('Key7' , 7 ) skip_list.insert('Key5' , 5 ) skip_list.insert('Key10' , 1_0 ) _lowerCAmelCase : int = skip_list.head _lowerCAmelCase : Tuple = {} while node.level != 0: _lowerCAmelCase : int = node.forward[0] _lowerCAmelCase : Tuple = node.value if len(_A ) != 4: print() assert len(_A ) == 4 assert all_values["Key1"] == 1_2 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 1_0 def lowercase (): """simple docstring""" _lowerCAmelCase : Tuple = SkipList() assert skip_list.find('Some key' ) is None def lowercase (): """simple docstring""" _lowerCAmelCase : int = SkipList() skip_list.insert('Key2' , 2_0 ) assert skip_list.find('Key2' ) == 2_0 skip_list.insert('Some Key' , 1_0 ) skip_list.insert('Key2' , 8 ) skip_list.insert('V' , 1_3 ) assert skip_list.find('Y' ) is None assert skip_list.find('Key2' ) == 8 assert skip_list.find('Some Key' ) == 1_0 assert skip_list.find('V' ) == 1_3 def lowercase (): """simple docstring""" _lowerCAmelCase : Optional[int] = SkipList() skip_list.delete('Some key' ) assert len(skip_list.head.forward ) == 0 def lowercase (): """simple docstring""" _lowerCAmelCase : Any = SkipList() skip_list.insert('Key1' , 1_2 ) skip_list.insert('V' , 1_3 ) skip_list.insert('X' , 1_4 ) skip_list.insert('Key2' , 1_5 ) skip_list.delete('V' ) skip_list.delete('Key2' ) assert skip_list.find('V' ) is None assert skip_list.find('Key2' ) is None def lowercase (): """simple docstring""" _lowerCAmelCase : str = SkipList() skip_list.insert('Key1' , 1_2 ) skip_list.insert('V' , 1_3 ) skip_list.insert('X' , 1_4 ) skip_list.insert('Key2' , 1_5 ) skip_list.delete('V' ) assert skip_list.find('V' ) is None assert skip_list.find('X' ) == 1_4 assert skip_list.find('Key1' ) == 1_2 assert skip_list.find('Key2' ) == 1_5 skip_list.delete('X' ) assert skip_list.find('V' ) is None assert skip_list.find('X' ) is None assert skip_list.find('Key1' ) == 1_2 assert skip_list.find('Key2' ) == 1_5 skip_list.delete('Key1' ) assert skip_list.find('V' ) is None assert skip_list.find('X' ) is None assert skip_list.find('Key1' ) is None assert skip_list.find('Key2' ) == 1_5 skip_list.delete('Key2' ) assert skip_list.find('V' ) is None assert skip_list.find('X' ) is None assert skip_list.find('Key1' ) is None assert skip_list.find('Key2' ) is None def lowercase (): """simple docstring""" _lowerCAmelCase : Optional[int] = SkipList() skip_list.insert('Key1' , 1_2 ) skip_list.insert('V' , 1_3 ) skip_list.insert('X' , 1_4_2 ) skip_list.insert('Key2' , 1_5 ) skip_list.delete('X' ) def traverse_keys(_A ): yield node.key for forward_node in node.forward: yield from traverse_keys(_A ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def lowercase (): """simple docstring""" def is_sorted(_A ): return all(next_item >= item for item, next_item in zip(_A , lst[1:] ) ) _lowerCAmelCase : List[str] = SkipList() for i in range(1_0 ): skip_list.insert(_A , _A ) assert is_sorted(list(_A ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(_A ) ) skip_list.insert(-1_2 , -1_2 ) skip_list.insert(7_7 , 7_7 ) assert is_sorted(list(_A ) ) def lowercase (): """simple docstring""" for _ in range(1_0_0 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def lowercase (): """simple docstring""" _lowerCAmelCase : Optional[int] = SkipList() skip_list.insert(2 , '2' ) skip_list.insert(4 , '4' ) skip_list.insert(6 , '4' ) skip_list.insert(4 , '5' ) skip_list.insert(8 , '4' ) skip_list.insert(9 , '4' ) skip_list.delete(4 ) print(_A ) if __name__ == "__main__": import doctest doctest.testmod() main()
712
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = FlaxXLMRobertaModel.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Dict = 'The dog is cute and lives in the garden house' _lowerCAmelCase : List[str] = jnp.array([tokenizer.encode(snake_case__ )] ) _lowerCAmelCase : Optional[int] = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim _lowerCAmelCase : Tuple = jnp.array( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) _lowerCAmelCase : Union[str, Any] = model(snake_case__ )['last_hidden_state'] self.assertEqual(output.shape , snake_case__ ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , snake_case__ , atol=1E-3 ) )
630
0
'''simple docstring''' import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights _lowerCAmelCase : Union[str, Any] = FlaxDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=snake_case__ , cache_dir=snake_case__ ) _lowerCAmelCase : Union[str, Any] = [t[-1] for t in os.walk(os.path.join(snake_case__ , os.listdir(snake_case__ )[0] , 'snapshots' ) )] _lowerCAmelCase : Tuple = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('.bin' ) for f in files ) @slow @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = FlaxStableDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-pipe' , safety_checker=snake_case__ ) _lowerCAmelCase : List[str] = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _lowerCAmelCase : List[str] = jax.random.PRNGKey(0 ) _lowerCAmelCase : Optional[Any] = 4 _lowerCAmelCase : str = jax.device_count() _lowerCAmelCase : Union[str, Any] = num_samples * [prompt] _lowerCAmelCase : List[Any] = pipeline.prepare_inputs(snake_case__ ) # shard inputs and rng _lowerCAmelCase : List[Any] = replicate(snake_case__ ) _lowerCAmelCase : Tuple = jax.random.split(snake_case__ , snake_case__ ) _lowerCAmelCase : int = shard(snake_case__ ) _lowerCAmelCase : List[str] = pipeline(snake_case__ , snake_case__ , snake_case__ , snake_case__ , jit=snake_case__ ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.151_4745 ) < 1E-3 assert np.abs(np.abs(snake_case__ , dtype=np.floataa ).sum() - 4_9947.875 ) < 5E-1 _lowerCAmelCase : Dict = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(snake_case__ ) == num_samples def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='flax' , safety_checker=snake_case__ ) _lowerCAmelCase : int = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _lowerCAmelCase : str = jax.random.PRNGKey(0 ) _lowerCAmelCase : Optional[Any] = 50 _lowerCAmelCase : List[str] = jax.device_count() _lowerCAmelCase : Union[str, Any] = num_samples * [prompt] _lowerCAmelCase : str = pipeline.prepare_inputs(snake_case__ ) # shard inputs and rng _lowerCAmelCase : Optional[Any] = replicate(snake_case__ ) _lowerCAmelCase : Dict = jax.random.split(snake_case__ , snake_case__ ) _lowerCAmelCase : str = shard(snake_case__ ) _lowerCAmelCase : Optional[Any] = pipeline(snake_case__ , snake_case__ , snake_case__ , snake_case__ , jit=snake_case__ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0565_2401) ) < 1E-3 assert np.abs((np.abs(snake_case__ , dtype=np.floataa ).sum() - 238_3808.2) ) < 5E-1 def a ( self ): '''simple docstring''' _lowerCAmelCase : str = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=snake_case__ ) _lowerCAmelCase : Any = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _lowerCAmelCase : List[Any] = jax.random.PRNGKey(0 ) _lowerCAmelCase : Optional[int] = 50 _lowerCAmelCase : Optional[Any] = jax.device_count() _lowerCAmelCase : Optional[Any] = num_samples * [prompt] _lowerCAmelCase : Optional[Any] = pipeline.prepare_inputs(snake_case__ ) # shard inputs and rng _lowerCAmelCase : List[str] = replicate(snake_case__ ) _lowerCAmelCase : str = jax.random.split(snake_case__ , snake_case__ ) _lowerCAmelCase : List[Any] = shard(snake_case__ ) _lowerCAmelCase : List[Any] = pipeline(snake_case__ , snake_case__ , snake_case__ , snake_case__ , jit=snake_case__ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0400_3906) ) < 1E-3 assert np.abs((np.abs(snake_case__ , dtype=np.floataa ).sum() - 237_3516.75) ) < 5E-1 def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa ) _lowerCAmelCase : List[str] = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _lowerCAmelCase : Any = jax.random.PRNGKey(0 ) _lowerCAmelCase : Union[str, Any] = 50 _lowerCAmelCase : Optional[Any] = jax.device_count() _lowerCAmelCase : List[Any] = num_samples * [prompt] _lowerCAmelCase : Dict = pipeline.prepare_inputs(snake_case__ ) # shard inputs and rng _lowerCAmelCase : Union[str, Any] = replicate(snake_case__ ) _lowerCAmelCase : str = jax.random.split(snake_case__ , snake_case__ ) _lowerCAmelCase : Dict = shard(snake_case__ ) _lowerCAmelCase : Dict = pipeline(snake_case__ , snake_case__ , snake_case__ , snake_case__ , jit=snake_case__ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0400_3906) ) < 1E-3 assert np.abs((np.abs(snake_case__ , dtype=np.floataa ).sum() - 237_3516.75) ) < 5E-1 def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = FlaxDDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='scaled_linear' , set_alpha_to_one=snake_case__ , steps_offset=1 , ) _lowerCAmelCase : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , scheduler=snake_case__ , safety_checker=snake_case__ , ) _lowerCAmelCase : Tuple = scheduler.create_state() _lowerCAmelCase : str = scheduler_state _lowerCAmelCase : List[Any] = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _lowerCAmelCase : List[str] = jax.random.PRNGKey(0 ) _lowerCAmelCase : List[str] = 50 _lowerCAmelCase : str = jax.device_count() _lowerCAmelCase : List[str] = num_samples * [prompt] _lowerCAmelCase : Union[str, Any] = pipeline.prepare_inputs(snake_case__ ) # shard inputs and rng _lowerCAmelCase : Dict = replicate(snake_case__ ) _lowerCAmelCase : Tuple = jax.random.split(snake_case__ , snake_case__ ) _lowerCAmelCase : Union[str, Any] = shard(snake_case__ ) _lowerCAmelCase : Tuple = pipeline(snake_case__ , snake_case__ , snake_case__ , snake_case__ , jit=snake_case__ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_4504_3945) ) < 1E-3 assert np.abs((np.abs(snake_case__ , dtype=np.floataa ).sum() - 234_7693.5) ) < 5E-1 def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = ( 'A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of' ' field, close up, split lighting, cinematic' ) _lowerCAmelCase : List[str] = jax.device_count() _lowerCAmelCase : List[str] = num_samples * [prompt] _lowerCAmelCase : List[str] = jax.random.split(jax.random.PRNGKey(0 ) , snake_case__ ) _lowerCAmelCase : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=snake_case__ , ) _lowerCAmelCase : str = replicate(snake_case__ ) _lowerCAmelCase : int = pipeline.prepare_inputs(snake_case__ ) _lowerCAmelCase : Any = shard(snake_case__ ) _lowerCAmelCase : List[str] = pipeline(snake_case__ , snake_case__ , snake_case__ , jit=snake_case__ ).images assert images.shape == (num_samples, 1, 512, 512, 3) _lowerCAmelCase : Optional[Any] = images[2, 0, 256, 10:17, 1] # With memory efficient attention _lowerCAmelCase : str = FlaxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='bf16' , dtype=jnp.bfloataa , safety_checker=snake_case__ , use_memory_efficient_attention=snake_case__ , ) _lowerCAmelCase : List[str] = replicate(snake_case__ ) _lowerCAmelCase : Optional[Any] = pipeline.prepare_inputs(snake_case__ ) _lowerCAmelCase : List[Any] = shard(snake_case__ ) _lowerCAmelCase : Optional[Any] = pipeline(snake_case__ , snake_case__ , snake_case__ , jit=snake_case__ ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) _lowerCAmelCase : Tuple = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
713
'''simple docstring''' def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Dict = len(_A ) while cur > 1: # Find the maximum number in arr _lowerCAmelCase : int = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi _lowerCAmelCase : Dict = arr[mi::-1] + arr[mi + 1 : len(_A )] # Reverse whole list _lowerCAmelCase : Optional[int] = arr[cur - 1 :: -1] + arr[cur : len(_A )] cur -= 1 return arr if __name__ == "__main__": lowerCAmelCase : Tuple = input("""Enter numbers separated by a comma:\n""").strip() lowerCAmelCase : Tuple = [int(item) for item in user_input.split(""",""")] print(pancake_sort(unsorted))
630
0
lowerCAmelCase : Optional[Any] = [ """VerificationMode""", """Version""", """disable_progress_bar""", """enable_progress_bar""", """is_progress_bar_enabled""", """experimental""", ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
714
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) lowerCAmelCase : str = { """EleutherAI/gpt-j-6B""": """https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json""", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "gptj" __magic_name__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , snake_case__=5_0400 , snake_case__=2048 , snake_case__=4096 , snake_case__=28 , snake_case__=16 , snake_case__=64 , snake_case__=None , snake_case__="gelu_new" , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=1E-5 , snake_case__=0.02 , snake_case__=True , snake_case__=5_0256 , snake_case__=5_0256 , snake_case__=False , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : int = vocab_size _lowerCAmelCase : Optional[int] = n_positions _lowerCAmelCase : Optional[int] = n_embd _lowerCAmelCase : Optional[int] = n_layer _lowerCAmelCase : str = n_head _lowerCAmelCase : Tuple = n_inner _lowerCAmelCase : Tuple = rotary_dim _lowerCAmelCase : Optional[int] = activation_function _lowerCAmelCase : Any = resid_pdrop _lowerCAmelCase : List[str] = embd_pdrop _lowerCAmelCase : int = attn_pdrop _lowerCAmelCase : Any = layer_norm_epsilon _lowerCAmelCase : Optional[int] = initializer_range _lowerCAmelCase : List[str] = use_cache _lowerCAmelCase : Dict = bos_token_id _lowerCAmelCase : Any = eos_token_id super().__init__( bos_token_id=snake_case__ , eos_token_id=snake_case__ , tie_word_embeddings=snake_case__ , **snake_case__ ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = "default" , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' super().__init__(snake_case__ , task=snake_case__ , patching_specs=snake_case__ , use_past=snake_case__ ) if not getattr(self._config , 'pad_token_id' , snake_case__ ): # TODO: how to do that better? _lowerCAmelCase : Any = 0 @property def a ( self ): '''simple docstring''' _lowerCAmelCase : str = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(snake_case__ , direction='inputs' ) _lowerCAmelCase : int = {0: 'batch', 1: 'past_sequence + sequence'} else: _lowerCAmelCase : int = {0: 'batch', 1: 'sequence'} return common_inputs @property def a ( self ): '''simple docstring''' return self._config.n_layer @property def a ( self ): '''simple docstring''' return self._config.n_head def a ( self , snake_case__ , snake_case__ = -1 , snake_case__ = -1 , snake_case__ = False , snake_case__ = None , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = super(snake_case__ , self ).generate_dummy_inputs( snake_case__ , batch_size=snake_case__ , seq_length=snake_case__ , is_pair=snake_case__ , framework=snake_case__ ) # We need to order the input in the way they appears in the forward() _lowerCAmelCase : Any = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = common_inputs['input_ids'].shape # Not using the same length for past_key_values _lowerCAmelCase : Any = seqlen + 2 _lowerCAmelCase : Optional[int] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _lowerCAmelCase : Tuple = [ (torch.zeros(snake_case__ ), torch.zeros(snake_case__ )) for _ in range(self.num_layers ) ] _lowerCAmelCase : Tuple = common_inputs['attention_mask'] if self.use_past: _lowerCAmelCase : Any = ordered_inputs['attention_mask'].dtype _lowerCAmelCase : Union[str, Any] = torch.cat( [ordered_inputs['attention_mask'], torch.ones(snake_case__ , snake_case__ , dtype=snake_case__ )] , dim=1 ) return ordered_inputs @property def a ( self ): '''simple docstring''' return 13
630
0
'''simple docstring''' from __future__ import annotations def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Optional[int] = len(_A ) # We need to create solution object to save path. _lowerCAmelCase : List[str] = [[0 for _ in range(_A )] for _ in range(_A )] _lowerCAmelCase : Optional[int] = run_maze(_A , 0 , 0 , _A ) if solved: print('\n'.join(str(_A ) for row in solutions ) ) else: print('No solution exists!' ) return solved def lowercase (_A , _A , _A , _A ): """simple docstring""" _lowerCAmelCase : List[Any] = len(_A ) # Final check point. if i == j == (size - 1): _lowerCAmelCase : Any = 1 return True _lowerCAmelCase : Dict = (not i < 0) and (not j < 0) # Check lower bounds _lowerCAmelCase : Any = (i < size) and (j < size) # Check upper bounds if lower_flag and upper_flag: # check for already visited and block points. _lowerCAmelCase : str = (not solutions[i][j]) and (not maze[i][j]) if block_flag: # check visited _lowerCAmelCase : Union[str, Any] = 1 # check for directions if ( run_maze(_A , i + 1 , _A , _A ) or run_maze(_A , _A , j + 1 , _A ) or run_maze(_A , i - 1 , _A , _A ) or run_maze(_A , _A , j - 1 , _A ) ): return True _lowerCAmelCase : Tuple = 0 return False return False if __name__ == "__main__": import doctest doctest.testmod()
715
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Any = { """configuration_x_clip""": [ """XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XCLIPConfig""", """XCLIPTextConfig""", """XCLIPVisionConfig""", ], """processing_x_clip""": ["""XCLIPProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """XCLIPModel""", """XCLIPPreTrainedModel""", """XCLIPTextModel""", """XCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys lowerCAmelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowerCAmelCase : int = False class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = VersatileDiffusionTextToImagePipeline.from_pretrained('shi-labs/versatile-diffusion' ) # remove text_unet pipe.remove_unused_weights() pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : List[Any] = 'A painting of a squirrel eating a burger ' _lowerCAmelCase : str = torch.manual_seed(0 ) _lowerCAmelCase : Tuple = pipe( prompt=snake_case__ , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(snake_case__ ) _lowerCAmelCase : Union[str, Any] = VersatileDiffusionTextToImagePipeline.from_pretrained(snake_case__ ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : int = generator.manual_seed(0 ) _lowerCAmelCase : Any = pipe( prompt=snake_case__ , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = VersatileDiffusionTextToImagePipeline.from_pretrained( 'shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : str = 'A painting of a squirrel eating a burger ' _lowerCAmelCase : Optional[int] = torch.manual_seed(0 ) _lowerCAmelCase : str = pipe( prompt=snake_case__ , generator=snake_case__ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _lowerCAmelCase : Optional[Any] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _lowerCAmelCase : Optional[int] = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
716
'''simple docstring''' import math from datetime import datetime, timedelta def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Optional[Any] = year % 1_9 _lowerCAmelCase : Any = year % 4 _lowerCAmelCase : Optional[int] = year % 7 _lowerCAmelCase : int = math.floor(year / 1_0_0 ) _lowerCAmelCase : Dict = math.floor((1_3 + 8 * leap_day_inhibits) / 2_5 ) _lowerCAmelCase : Optional[Any] = leap_day_inhibits / 4 _lowerCAmelCase : Dict = ( 1_5 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 3_0 _lowerCAmelCase : List[Any] = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 _lowerCAmelCase : Dict = (1_9 * metonic_cycle + secular_moon_shift) % 3_0 # PHM -> Paschal Full Moon _lowerCAmelCase : Union[str, Any] = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 2_9 and days_from_phm_to_sunday == 6: return datetime(_A , 4 , 1_9 ) elif days_to_add == 2_8 and days_from_phm_to_sunday == 6: return datetime(_A , 4 , 1_8 ) else: return datetime(_A , 3 , 2_2 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (19_94, 20_00, 20_10, 20_21, 20_23): lowerCAmelCase : List[str] = """will be""" if year > datetime.now().year else """was""" print(F'''Easter in {year} {tense} {gauss_easter(year)}''')
630
0
'''simple docstring''' def lowercase (_A ): """simple docstring""" stooge(_A , 0 , len(_A ) - 1 ) return arr def lowercase (_A , _A , _A ): """simple docstring""" if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _lowerCAmelCase : List[str] = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _lowerCAmelCase : Union[str, Any] = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(_A , _A , (h - t) ) # Recursively sort last 2/3 elements stooge(_A , i + t , (_A) ) # Recursively sort first 2/3 elements stooge(_A , _A , (h - t) ) if __name__ == "__main__": lowerCAmelCase : Optional[int] = input("""Enter numbers separated by a comma:\n""").strip() lowerCAmelCase : Dict = [int(item) for item in user_input.split(""",""")] print(stooge_sort(unsorted))
717
'''simple docstring''' import unittest from knapsack import greedy_knapsack as kp class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = [10, 20, 30, 40, 50, 60] _lowerCAmelCase : Union[str, Any] = [2, 4, 6, 8, 10, 12] _lowerCAmelCase : Dict = 100 self.assertEqual(kp.calc_profit(snake_case__ , snake_case__ , snake_case__ ) , 210 ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'max_weight must greater than zero.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'Weight can not be negative.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'Profit can not be negative.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex(snake_case__ , 'max_weight must greater than zero.' ) def a ( self ): '''simple docstring''' self.assertRaisesRegex( snake_case__ , 'The length of profit and weight must be same.' ) if __name__ == "__main__": unittest.main()
630
0
'''simple docstring''' def lowercase (_A ): """simple docstring""" assert ( isinstance(_A , _A ) and number_of_steps > 0 ), f'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 _lowerCAmelCase : Any = 1, 1 for _ in range(number_of_steps - 1 ): _lowerCAmelCase : Any = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
718
'''simple docstring''' def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : Optional[int] = (boundary[1] - boundary[0]) / steps _lowerCAmelCase : Any = boundary[0] _lowerCAmelCase : List[str] = boundary[1] _lowerCAmelCase : Tuple = make_points(_A , _A , _A ) _lowerCAmelCase : Tuple = 0.0 y += (h / 2.0) * f(_A ) for i in x_i: # print(i) y += h * f(_A ) y += (h / 2.0) * f(_A ) return y def lowercase (_A , _A , _A ): """simple docstring""" _lowerCAmelCase : Tuple = a + h while x < (b - h): yield x _lowerCAmelCase : Any = x + h def lowercase (_A ): # enter your function here """simple docstring""" _lowerCAmelCase : int = (x - 0) * (x - 0) return y def lowercase (): """simple docstring""" _lowerCAmelCase : Optional[Any] = 0.0 # Lower bound of integration _lowerCAmelCase : Dict = 1.0 # Upper bound of integration _lowerCAmelCase : Optional[Any] = 10.0 # define number of steps or resolution _lowerCAmelCase : Optional[int] = [a, b] # define boundary of integration _lowerCAmelCase : List[Any] = method_a(_A , _A ) print(f'y = {y}' ) if __name__ == "__main__": main()
630
0
'''simple docstring''' from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = False , snake_case__ = False , snake_case__ = None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' super().__init__( features=snake_case__ , cache_dir=snake_case__ , keep_in_memory=snake_case__ , streaming=snake_case__ , num_proc=snake_case__ , **snake_case__ , ) _lowerCAmelCase : int = Generator( cache_dir=snake_case__ , features=snake_case__ , generator=snake_case__ , gen_kwargs=snake_case__ , **snake_case__ , ) def a ( self ): '''simple docstring''' if self.streaming: _lowerCAmelCase : Optional[int] = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: _lowerCAmelCase : List[Any] = None _lowerCAmelCase : Optional[int] = None _lowerCAmelCase : int = None _lowerCAmelCase : Tuple = None self.builder.download_and_prepare( download_config=snake_case__ , download_mode=snake_case__ , verification_mode=snake_case__ , base_path=snake_case__ , num_proc=self.num_proc , ) _lowerCAmelCase : Optional[Any] = self.builder.as_dataset( split='train' , verification_mode=snake_case__ , in_memory=self.keep_in_memory ) return dataset
719
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) lowerCAmelCase : int = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys lowerCAmelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=30 , snake_case__=2 , snake_case__=3 , snake_case__=True , snake_case__=True , snake_case__=32 , snake_case__=2 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=10 , snake_case__=0.02 , snake_case__=3 , snake_case__=None , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = parent _lowerCAmelCase : str = batch_size _lowerCAmelCase : Optional[int] = image_size _lowerCAmelCase : Optional[int] = patch_size _lowerCAmelCase : Optional[int] = num_channels _lowerCAmelCase : int = is_training _lowerCAmelCase : Tuple = use_labels _lowerCAmelCase : Optional[Any] = hidden_size _lowerCAmelCase : List[str] = num_hidden_layers _lowerCAmelCase : Any = num_attention_heads _lowerCAmelCase : Optional[Any] = intermediate_size _lowerCAmelCase : Tuple = hidden_act _lowerCAmelCase : List[str] = hidden_dropout_prob _lowerCAmelCase : Tuple = attention_probs_dropout_prob _lowerCAmelCase : Union[str, Any] = type_sequence_label_size _lowerCAmelCase : Dict = initializer_range _lowerCAmelCase : Union[str, Any] = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) _lowerCAmelCase : Tuple = (image_size // patch_size) ** 2 _lowerCAmelCase : Any = num_patches + 1 def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase : int = None if self.use_labels: _lowerCAmelCase : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCAmelCase : int = self.get_config() return config, pixel_values, labels def a ( self ): '''simple docstring''' return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def a ( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[str] = TFViTModel(config=snake_case__ ) _lowerCAmelCase : List[Any] = model(snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. _lowerCAmelCase : Tuple = self.image_size // 2 _lowerCAmelCase : str = pixel_values[:, :, :image_size, :image_size] _lowerCAmelCase : Optional[int] = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) _lowerCAmelCase : Tuple = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def a ( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = self.type_sequence_label_size _lowerCAmelCase : Optional[Any] = TFViTForImageClassification(snake_case__ ) _lowerCAmelCase : Dict = model(snake_case__ , labels=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. _lowerCAmelCase : List[Any] = self.image_size // 2 _lowerCAmelCase : str = pixel_values[:, :, :image_size, :image_size] _lowerCAmelCase : Union[str, Any] = model(snake_case__ , interpolate_pos_encoding=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images _lowerCAmelCase : int = 1 _lowerCAmelCase : Any = TFViTForImageClassification(snake_case__ ) _lowerCAmelCase : Tuple = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCAmelCase : Union[str, Any] = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.prepare_config_and_inputs() _lowerCAmelCase : Tuple = config_and_inputs _lowerCAmelCase : int = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () __magic_name__ = ( {"feature-extraction": TFViTModel, "image-classification": TFViTForImageClassification} if is_tf_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = TFViTModelTester(self ) _lowerCAmelCase : Any = ConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ , hidden_size=37 ) def a ( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='ViT does not use inputs_embeds' ) def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : List[str] = model_class(snake_case__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) _lowerCAmelCase : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case__ , tf.keras.layers.Layer ) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : Dict = model_class(snake_case__ ) _lowerCAmelCase : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCAmelCase : Any = [*signature.parameters.keys()] _lowerCAmelCase : Tuple = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = TFViTModel.from_pretrained('google/vit-base-patch16-224' ) self.assertIsNotNone(snake_case__ ) def lowercase (): """simple docstring""" _lowerCAmelCase : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def a ( self ): '''simple docstring''' return ViTImageProcessor.from_pretrained('google/vit-base-patch16-224' ) if is_vision_available() else None @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = TFViTForImageClassification.from_pretrained('google/vit-base-patch16-224' ) _lowerCAmelCase : int = self.default_image_processor _lowerCAmelCase : Dict = prepare_img() _lowerCAmelCase : Optional[int] = image_processor(images=snake_case__ , return_tensors='tf' ) # forward pass _lowerCAmelCase : Any = model(**snake_case__ ) # verify the logits _lowerCAmelCase : Tuple = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , snake_case__ ) _lowerCAmelCase : Union[str, Any] = tf.constant([-0.2744, 0.8215, -0.0836] ) tf.debugging.assert_near(outputs.logits[0, :3] , snake_case__ , atol=1E-4 )
720
'''simple docstring''' from collections import Counter from timeit import timeit def lowercase (_A = "" , ): """simple docstring""" return sum(c % 2 for c in Counter(input_str.replace(' ' , '' ).lower() ).values() ) < 2 def lowercase (_A = "" ): """simple docstring""" if len(_A ) == 0: return True _lowerCAmelCase : Union[str, Any] = input_str.replace(' ' , '' ).lower() # character_freq_dict: Stores the frequency of every character in the input string _lowerCAmelCase : dict[str, int] = {} for character in lower_case_input_str: _lowerCAmelCase : Union[str, Any] = character_freq_dict.get(_A , 0 ) + 1 _lowerCAmelCase : List[Any] = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def lowercase (_A = "" ): """simple docstring""" print('\nFor string = ' , _A , ':' ) print( '> can_string_be_rearranged_as_palindrome_counter()' , '\tans =' , can_string_be_rearranged_as_palindrome_counter(_A ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome_counter(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) print( '> can_string_be_rearranged_as_palindrome()' , '\tans =' , can_string_be_rearranged_as_palindrome(_A ) , '\ttime =' , timeit( 'z.can_string_be_rearranged_as_palindrome(z.check_str)' , setup='import __main__ as z' , ) , 'seconds' , ) if __name__ == "__main__": lowerCAmelCase : Tuple = input( """Enter string to determine if it can be rearranged as a palindrome or not: """ ).strip() benchmark(check_str) lowerCAmelCase : Optional[Any] = can_string_be_rearranged_as_palindrome_counter(check_str) print(F'''{check_str} can {"" if status else "not "}be rearranged as a palindrome''')
630
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: lowerCAmelCase : int = None lowerCAmelCase : Tuple = logging.get_logger(__name__) lowerCAmelCase : Tuple = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} lowerCAmelCase : Tuple = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", }, """tokenizer_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json""", }, } lowerCAmelCase : Dict = { """albert-base-v1""": 5_12, """albert-large-v1""": 5_12, """albert-xlarge-v1""": 5_12, """albert-xxlarge-v1""": 5_12, """albert-base-v2""": 5_12, """albert-large-v2""": 5_12, """albert-xlarge-v2""": 5_12, """albert-xxlarge-v2""": 5_12, } lowerCAmelCase : Optional[int] = """▁""" class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = AlbertTokenizer def __init__( self , snake_case__=None , snake_case__=None , snake_case__=True , snake_case__=True , snake_case__=False , snake_case__="[CLS]" , snake_case__="[SEP]" , snake_case__="<unk>" , snake_case__="[SEP]" , snake_case__="<pad>" , snake_case__="[CLS]" , snake_case__="[MASK]" , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Optional[int] = ( AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ , normalized=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token ) super().__init__( snake_case__ , tokenizer_file=snake_case__ , do_lower_case=snake_case__ , remove_space=snake_case__ , keep_accents=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , pad_token=snake_case__ , cls_token=snake_case__ , mask_token=snake_case__ , **snake_case__ , ) _lowerCAmelCase : Dict = do_lower_case _lowerCAmelCase : Tuple = remove_space _lowerCAmelCase : Optional[int] = keep_accents _lowerCAmelCase : Optional[Any] = vocab_file _lowerCAmelCase : Tuple = False if not self.vocab_file else True def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = [self.sep_token_id] _lowerCAmelCase : Optional[int] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : List[str] = [self.sep_token_id] _lowerCAmelCase : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCAmelCase : Optional[Any] = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ): copyfile(self.vocab_file , snake_case__ ) return (out_vocab_file,)
721
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase : str = logging.get_logger(__name__) lowerCAmelCase : int = { """facebook/data2vec-text-base""": """https://huggingface.co/data2vec/resolve/main/config.json""", } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "data2vec-text" def __init__( self , snake_case__=3_0522 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=2 , snake_case__=0.02 , snake_case__=1E-12 , snake_case__=1 , snake_case__=0 , snake_case__=2 , snake_case__="absolute" , snake_case__=True , snake_case__=None , **snake_case__ , ): '''simple docstring''' super().__init__(pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ , **snake_case__ ) _lowerCAmelCase : List[Any] = vocab_size _lowerCAmelCase : Tuple = hidden_size _lowerCAmelCase : Dict = num_hidden_layers _lowerCAmelCase : int = num_attention_heads _lowerCAmelCase : str = hidden_act _lowerCAmelCase : Any = intermediate_size _lowerCAmelCase : Any = hidden_dropout_prob _lowerCAmelCase : Optional[int] = attention_probs_dropout_prob _lowerCAmelCase : str = max_position_embeddings _lowerCAmelCase : Any = type_vocab_size _lowerCAmelCase : int = initializer_range _lowerCAmelCase : List[str] = layer_norm_eps _lowerCAmelCase : List[Any] = position_embedding_type _lowerCAmelCase : str = use_cache _lowerCAmelCase : Union[str, Any] = classifier_dropout class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" @property def a ( self ): '''simple docstring''' if self.task == "multiple-choice": _lowerCAmelCase : Union[str, Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _lowerCAmelCase : List[Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
630
0
'''simple docstring''' import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "char" __magic_name__ = "bpe" __magic_name__ = "wp" lowerCAmelCase : List[str] = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = ["image_processor", "char_tokenizer"] __magic_name__ = "ViTImageProcessor" __magic_name__ = "MgpstrTokenizer" def __init__( self , snake_case__=None , snake_case__=None , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Tuple = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case__ , ) _lowerCAmelCase : Optional[Any] = kwargs.pop('feature_extractor' ) _lowerCAmelCase : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) _lowerCAmelCase : List[Any] = tokenizer _lowerCAmelCase : Optional[int] = AutoTokenizer.from_pretrained('gpt2' ) _lowerCAmelCase : Any = AutoTokenizer.from_pretrained('bert-base-uncased' ) super().__init__(snake_case__ , snake_case__ ) def __call__( self , snake_case__=None , snake_case__=None , snake_case__=None , **snake_case__ ): '''simple docstring''' if images is None and text is None: raise ValueError('You need to specify either an `images` or `text` input to process.' ) if images is not None: _lowerCAmelCase : str = self.image_processor(snake_case__ , return_tensors=snake_case__ , **snake_case__ ) if text is not None: _lowerCAmelCase : List[Any] = self.char_tokenizer(snake_case__ , return_tensors=snake_case__ , **snake_case__ ) if text is None: return inputs elif images is None: return encodings else: _lowerCAmelCase : int = encodings['input_ids'] return inputs def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[str] = sequences _lowerCAmelCase : Dict = char_preds.size(0 ) _lowerCAmelCase : Union[str, Any] = self._decode_helper(snake_case__ , 'char' ) _lowerCAmelCase : Optional[int] = self._decode_helper(snake_case__ , 'bpe' ) _lowerCAmelCase : Union[str, Any] = self._decode_helper(snake_case__ , 'wp' ) _lowerCAmelCase : Dict = [] _lowerCAmelCase : Union[str, Any] = [] for i in range(snake_case__ ): _lowerCAmelCase : Dict = [char_scores[i], bpe_scores[i], wp_scores[i]] _lowerCAmelCase : Any = [char_strs[i], bpe_strs[i], wp_strs[i]] _lowerCAmelCase : List[str] = scores.index(max(snake_case__ ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) _lowerCAmelCase : Dict = {} _lowerCAmelCase : List[Any] = final_strs _lowerCAmelCase : List[str] = final_scores _lowerCAmelCase : str = char_strs _lowerCAmelCase : Optional[int] = bpe_strs _lowerCAmelCase : Dict = wp_strs return out def a ( self , snake_case__ , snake_case__ ): '''simple docstring''' if format == DecodeType.CHARACTER: _lowerCAmelCase : str = self.char_decode _lowerCAmelCase : Union[str, Any] = 1 _lowerCAmelCase : str = '[s]' elif format == DecodeType.BPE: _lowerCAmelCase : Optional[Any] = self.bpe_decode _lowerCAmelCase : Dict = 2 _lowerCAmelCase : Dict = '#' elif format == DecodeType.WORDPIECE: _lowerCAmelCase : Optional[int] = self.wp_decode _lowerCAmelCase : Optional[Any] = 102 _lowerCAmelCase : Optional[Any] = '[SEP]' else: raise ValueError(F'Format {format} is not supported.' ) _lowerCAmelCase : Any = [], [] _lowerCAmelCase : Any = pred_logits.size(0 ) _lowerCAmelCase : Union[str, Any] = pred_logits.size(1 ) _lowerCAmelCase : List[str] = pred_logits.topk(1 , dim=-1 , largest=snake_case__ , sorted=snake_case__ ) _lowerCAmelCase : int = preds_index.view(-1 , snake_case__ )[:, 1:] _lowerCAmelCase : Any = decoder(snake_case__ ) _lowerCAmelCase : str = torch.nn.functional.softmax(snake_case__ , dim=2 ).max(dim=2 ) _lowerCAmelCase : str = preds_max_prob[:, 1:] for index in range(snake_case__ ): _lowerCAmelCase : List[str] = preds_str[index].find(snake_case__ ) _lowerCAmelCase : Optional[Any] = preds_str[index][:pred_eos] _lowerCAmelCase : Optional[Any] = preds_index[index].cpu().tolist() _lowerCAmelCase : int = pred_index.index(snake_case__ ) if eos_token in pred_index else -1 _lowerCAmelCase : Dict = preds_max_prob[index][: pred_eos_index + 1] _lowerCAmelCase : List[Any] = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(snake_case__ ) conf_scores.append(snake_case__ ) return dec_strs, conf_scores def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = [seq.replace(' ' , '' ) for seq in self.char_tokenizer.batch_decode(snake_case__ )] return decode_strs def a ( self , snake_case__ ): '''simple docstring''' return self.bpe_tokenizer.batch_decode(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = [seq.replace(' ' , '' ) for seq in self.wp_tokenizer.batch_decode(snake_case__ )] return decode_strs
700
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins lowerCAmelCase : List[str] = ["""tests.fixtures.files""", """tests.fixtures.hub""", """tests.fixtures.fsspec"""] def lowercase (_A , _A ): """simple docstring""" for item in items: if any(marker in item.keywords for marker in ['integration', 'unit'] ): continue item.add_marker(pytest.mark.unit ) def lowercase (_A ): """simple docstring""" config.addinivalue_line('markers' , 'torchaudio_latest: mark test to run with torchaudio>=0.12' ) @pytest.fixture(autouse=_A ) def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : str = tmp_path_factory.getbasetemp() / 'cache' _lowerCAmelCase : Dict = test_hf_cache_home / 'datasets' _lowerCAmelCase : List[Any] = test_hf_cache_home / 'metrics' _lowerCAmelCase : List[Any] = test_hf_cache_home / 'modules' monkeypatch.setattr('datasets.config.HF_DATASETS_CACHE' , str(_A ) ) monkeypatch.setattr('datasets.config.HF_METRICS_CACHE' , str(_A ) ) monkeypatch.setattr('datasets.config.HF_MODULES_CACHE' , str(_A ) ) _lowerCAmelCase : Dict = test_hf_datasets_cache / 'downloads' monkeypatch.setattr('datasets.config.DOWNLOADED_DATASETS_PATH' , str(_A ) ) _lowerCAmelCase : Union[str, Any] = test_hf_datasets_cache / 'downloads' / 'extracted' monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH' , str(_A ) ) @pytest.fixture(autouse=_A , scope='session' ) def lowercase (): """simple docstring""" datasets.disable_progress_bar() @pytest.fixture(autouse=_A ) def lowercase (_A ): """simple docstring""" monkeypatch.setattr('datasets.config.HF_UPDATE_DOWNLOAD_COUNTS' , _A ) @pytest.fixture def lowercase (_A ): """simple docstring""" monkeypatch.setattr('sqlalchemy.util.deprecations.SILENCE_UBER_WARNING' , _A )
630
0
import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCAmelCase : Union[str, Any] = version.parse(importlib_metadata.version("""nltk""")) if NLTK_VERSION >= version.Version("""3.6.4"""): from nltk import word_tokenize lowerCAmelCase : Union[str, Any] = """\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } """ lowerCAmelCase : Optional[Any] = """\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. """ lowerCAmelCase : Tuple = """ Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: 'meteor': meteor score. Examples: >>> meteor = datasets.load_metric('meteor') >>> predictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"] >>> references = [\"It is a guide to action that ensures that the military will forever heed Party commands\"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results[\"meteor\"], 4)) 0.6944 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase__ ( datasets.Metric ): """simple docstring""" def a ( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'] , reference_urls=[ 'https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score', 'https://en.wikipedia.org/wiki/METEOR', ] , ) def a ( self , snake_case__ ): '''simple docstring''' import nltk nltk.download('wordnet' ) if NLTK_VERSION >= version.Version('3.6.5' ): nltk.download('punkt' ) if NLTK_VERSION >= version.Version('3.6.6' ): nltk.download('omw-1.4' ) def a ( self , snake_case__ , snake_case__ , snake_case__=0.9 , snake_case__=3 , snake_case__=0.5 ): '''simple docstring''' if NLTK_VERSION >= version.Version('3.6.5' ): _lowerCAmelCase : Dict = [ meteor_score.single_meteor_score( word_tokenize(snake_case__ ) , word_tokenize(snake_case__ ) , alpha=snake_case__ , beta=snake_case__ , gamma=snake_case__ ) for ref, pred in zip(snake_case__ , snake_case__ ) ] else: _lowerCAmelCase : List[Any] = [ meteor_score.single_meteor_score(snake_case__ , snake_case__ , alpha=snake_case__ , beta=snake_case__ , gamma=snake_case__ ) for ref, pred in zip(snake_case__ , snake_case__ ) ] return {"meteor": np.mean(snake_case__ )}
701
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCAmelCase : str = logging.get_logger(__name__) # General docstring lowerCAmelCase : Optional[Any] = """RegNetConfig""" # Base docstring lowerCAmelCase : int = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = [1, 10_88, 7, 7] # Image classification docstring lowerCAmelCase : Any = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = """tabby, tabby cat""" lowerCAmelCase : Tuple = [ """facebook/regnet-y-040""", # See all regnet models at https://huggingface.co/models?filter=regnet ] class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 3 , snake_case__ = 1 , snake_case__ = 1 , snake_case__ = "relu" , **snake_case__ , ): '''simple docstring''' super().__init__(**snake_case__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _lowerCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _lowerCAmelCase : List[Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=snake_case__ , strides=snake_case__ , padding='VALID' , groups=snake_case__ , use_bias=snake_case__ , name='convolution' , ) _lowerCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) _lowerCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.convolution(self.padding(snake_case__ ) ) _lowerCAmelCase : Union[str, Any] = self.normalization(snake_case__ ) _lowerCAmelCase : int = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = config.num_channels _lowerCAmelCase : List[Any] = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name='embedder' , ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = shape_list(snake_case__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( 'Make sure that the channel dimension of the pixel values match with the one set in the configuration.' ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _lowerCAmelCase : List[Any] = tf.transpose(snake_case__ , perm=(0, 2, 3, 1) ) _lowerCAmelCase : Tuple = self.embedder(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=1 , strides=snake_case__ , use_bias=snake_case__ , name='convolution' ) _lowerCAmelCase : Tuple = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' return self.normalization(self.convolution(snake_case__ ) , training=snake_case__ ) class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Any = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) _lowerCAmelCase : str = [ tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='relu' , name='attention.0' ), tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='sigmoid' , name='attention.2' ), ] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.pooler(snake_case__ ) for layer_module in self.attention: _lowerCAmelCase : Tuple = layer_module(snake_case__ ) _lowerCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Optional[int] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _lowerCAmelCase : Any = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.2' ), ] _lowerCAmelCase : List[str] = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = hidden_state for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) _lowerCAmelCase : int = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : Tuple = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : List[str] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) _lowerCAmelCase : Tuple = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetSELayer(snake_case__ , reduced_channels=int(round(in_channels / 4 ) ) , name='layer.2' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.3' ), ] _lowerCAmelCase : Tuple = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = hidden_state for layer_module in self.layers: _lowerCAmelCase : List[Any] = layer_module(snake_case__ ) _lowerCAmelCase : Tuple = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : str = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 2 , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Dict = TFRegNetXLayer if config.layer_type == 'x' else TFRegNetYLayer _lowerCAmelCase : Optional[int] = [ # downsampling is done in the first layer with stride of 2 layer(snake_case__ , snake_case__ , snake_case__ , stride=snake_case__ , name='layers.0' ), *[layer(snake_case__ , snake_case__ , snake_case__ , name=F'layers.{i+1}' ) for i in range(depth - 1 )], ] def a ( self , snake_case__ ): '''simple docstring''' for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( snake_case__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name='stages.0' , ) ) _lowerCAmelCase : Union[str, Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(snake_case__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(snake_case__ , snake_case__ , snake_case__ , depth=snake_case__ , name=F'stages.{i+1}' ) ) def a ( self , snake_case__ , snake_case__ = False , snake_case__ = True ): '''simple docstring''' _lowerCAmelCase : List[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _lowerCAmelCase : str = hidden_states + (hidden_state,) _lowerCAmelCase : List[str] = stage_module(snake_case__ ) if output_hidden_states: _lowerCAmelCase : Dict = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=snake_case__ , hidden_states=snake_case__ ) @keras_serializable class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" __magic_name__ = RegNetConfig def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = config _lowerCAmelCase : Union[str, Any] = TFRegNetEmbeddings(snake_case__ , name='embedder' ) _lowerCAmelCase : Optional[int] = TFRegNetEncoder(snake_case__ , name='encoder' ) _lowerCAmelCase : Dict = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) @unpack_inputs def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' _lowerCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : int = self.embedder(snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[str] = self.encoder( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[Any] = encoder_outputs[0] _lowerCAmelCase : Tuple = self.pooler(snake_case__ ) # Change to NCHW output format have uniformity in the modules _lowerCAmelCase : Optional[int] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) _lowerCAmelCase : Optional[Any] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _lowerCAmelCase : Union[str, Any] = tuple([tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=snake_case__ , pooler_output=snake_case__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = RegNetConfig __magic_name__ = "regnet" __magic_name__ = "pixel_values" @property def a ( self ): '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 224, 224) , dtype=tf.floataa )} lowerCAmelCase : List[Any] = r""" Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. """ lowerCAmelCase : Dict = r""" Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : List[str] = TFRegNetMainLayer(snake_case__ , name='regnet' ) @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Any = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : str = self.regnet( pixel_values=snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[Any] = config.num_labels _lowerCAmelCase : Optional[Any] = TFRegNetMainLayer(snake_case__ , name='regnet' ) # classification head _lowerCAmelCase : Optional[int] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name='classifier.1' ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def a ( self , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : Dict = self.regnet( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.pooler_output if return_dict else outputs[1] _lowerCAmelCase : List[Any] = self.classifier[0](snake_case__ ) _lowerCAmelCase : Tuple = self.classifier[1](snake_case__ ) _lowerCAmelCase : int = None if labels is None else self.hf_compute_loss(labels=snake_case__ , logits=snake_case__ ) if not return_dict: _lowerCAmelCase : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=snake_case__ , logits=snake_case__ , hidden_states=outputs.hidden_states )
630
0
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def a ( self ): '''simple docstring''' torch.manual_seed(0 ) _lowerCAmelCase : Any = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('AttnDownBlock2D', 'DownBlock2D') , up_block_types=('UpBlock2D', 'AttnUpBlock2D') , ) return model @property def a ( self ): '''simple docstring''' torch.manual_seed(0 ) _lowerCAmelCase : Dict = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('CrossAttnDownBlock2D', 'DownBlock2D') , up_block_types=('UpBlock2D', 'CrossAttnUpBlock2D') , cross_attention_dim=10 , ) return model @property def a ( self ): '''simple docstring''' torch.manual_seed(0 ) _lowerCAmelCase : int = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('DownEncoderBlock2D', 'DownEncoderBlock2D') , up_block_types=('UpDecoderBlock2D', 'UpDecoderBlock2D') , ) _lowerCAmelCase : int = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=('AttnDownBlock2D', 'DownBlock2D') , up_block_types=('UpBlock2D', 'AttnUpBlock2D') , ) return vqvae, unet @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase : int = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) _lowerCAmelCase : List[Any] = DDPMScheduler() _lowerCAmelCase : Union[str, Any] = AudioDiffusionPipeline(vqvae=snake_case__ , unet=self.dummy_unet , mel=snake_case__ , scheduler=snake_case__ ) _lowerCAmelCase : Union[str, Any] = pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : List[str] = torch.Generator(device=snake_case__ ).manual_seed(42 ) _lowerCAmelCase : Optional[int] = pipe(generator=snake_case__ , steps=4 ) _lowerCAmelCase : List[Any] = output.audios[0] _lowerCAmelCase : Tuple = output.images[0] _lowerCAmelCase : Tuple = torch.Generator(device=snake_case__ ).manual_seed(42 ) _lowerCAmelCase : str = pipe(generator=snake_case__ , steps=4 , return_dict=snake_case__ ) _lowerCAmelCase : Optional[Any] = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) _lowerCAmelCase : Tuple = np.frombuffer(image.tobytes() , dtype='uint8' )[:10] _lowerCAmelCase : Optional[Any] = np.frombuffer(image_from_tuple.tobytes() , dtype='uint8' )[:10] _lowerCAmelCase : Dict = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 _lowerCAmelCase : List[str] = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) _lowerCAmelCase : Any = DDIMScheduler() _lowerCAmelCase : Tuple = self.dummy_vqvae_and_unet _lowerCAmelCase : List[str] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=snake_case__ , scheduler=snake_case__ ) _lowerCAmelCase : Dict = pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) np.random.seed(0 ) _lowerCAmelCase : Optional[int] = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) _lowerCAmelCase : List[Any] = torch.Generator(device=snake_case__ ).manual_seed(42 ) _lowerCAmelCase : Union[str, Any] = pipe(raw_audio=snake_case__ , generator=snake_case__ , start_step=5 , steps=10 ) _lowerCAmelCase : List[Any] = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) _lowerCAmelCase : Union[str, Any] = np.frombuffer(image.tobytes() , dtype='uint8' )[:10] _lowerCAmelCase : List[Any] = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 _lowerCAmelCase : int = self.dummy_unet_condition _lowerCAmelCase : List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=snake_case__ , mel=snake_case__ , scheduler=snake_case__ ) _lowerCAmelCase : Optional[Any] = pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) np.random.seed(0 ) _lowerCAmelCase : List[Any] = torch.rand((1, 1, 10) ) _lowerCAmelCase : Tuple = pipe(generator=snake_case__ , encoding=snake_case__ ) _lowerCAmelCase : Optional[Any] = output.images[0] _lowerCAmelCase : Dict = np.frombuffer(image.tobytes() , dtype='uint8' )[:10] _lowerCAmelCase : Optional[int] = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = torch_device _lowerCAmelCase : Optional[Any] = DiffusionPipeline.from_pretrained('teticio/audio-diffusion-ddim-256' ) _lowerCAmelCase : str = pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : str = torch.Generator(device=snake_case__ ).manual_seed(42 ) _lowerCAmelCase : Dict = pipe(generator=snake_case__ ) _lowerCAmelCase : Any = output.audios[0] _lowerCAmelCase : Dict = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] _lowerCAmelCase : int = np.frombuffer(image.tobytes() , dtype='uint8' )[:10] _lowerCAmelCase : Dict = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
702
'''simple docstring''' from typing import Any def lowercase (_A ): """simple docstring""" if not input_list: return [] _lowerCAmelCase : Optional[int] = [input_list.count(_A ) for value in input_list] _lowerCAmelCase : int = max(_A ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(_A ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
630
0
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) lowerCAmelCase : List[str] = """\ Text data. Second line of data.""" lowerCAmelCase : Any = """file""" @pytest.fixture(scope='session' ) def lowercase (_A ): """simple docstring""" _lowerCAmelCase : int = tmp_path_factory.mktemp('data' ) / (FILE_PATH + '.zstd') _lowerCAmelCase : Tuple = bytes(_A , 'utf-8' ) with zstd.open(_A , 'wb' ) as f: f.write(_A ) return path @pytest.fixture def lowercase (_A ): """simple docstring""" with open(os.path.join(tmpfs.local_root_dir , _A ) , 'w' ) as f: f.write(_A ) return FILE_PATH @pytest.mark.parametrize('compression_format' , ['gzip', 'xz', 'zstd'] ) def lowercase (_A , _A , _A , _A , _A , _A ): """simple docstring""" _lowerCAmelCase : List[Any] = {'gzip': gz_file, 'xz': xz_file, 'zstd': zstd_path} _lowerCAmelCase : Any = input_paths[compression_format] _lowerCAmelCase : Union[str, Any] = tmp_path / 'cache' _lowerCAmelCase : List[Any] = DownloadConfig(cache_dir=_A , extract_compressed_file=_A ) _lowerCAmelCase : str = cached_path(_A , download_config=_A ) with open(_A ) as f: _lowerCAmelCase : List[Any] = f.read() with open(_A ) as f: _lowerCAmelCase : List[str] = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('default_extracted' , [True, False] ) @pytest.mark.parametrize('default_cache_dir' , [True, False] ) def lowercase (_A , _A , _A , _A , _A ): """simple docstring""" _lowerCAmelCase : int = 'custom_cache' _lowerCAmelCase : str = 'custom_extracted_dir' _lowerCAmelCase : Optional[Any] = tmp_path / 'custom_extracted_path' if default_extracted: _lowerCAmelCase : List[str] = ('downloads' if default_cache_dir else custom_cache_dir, 'extracted') else: monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_DIR' , _A ) monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH' , str(_A ) ) _lowerCAmelCase : List[str] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) _lowerCAmelCase : List[Any] = xz_file _lowerCAmelCase : Dict = ( DownloadConfig(extract_compressed_file=_A ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=_A ) ) _lowerCAmelCase : Any = cached_path(_A , download_config=_A ) assert Path(_A ).parent.parts[-2:] == expected def lowercase (_A ): """simple docstring""" _lowerCAmelCase : List[str] = str(Path(_A ).resolve() ) assert cached_path(_A ) == text_file # relative path _lowerCAmelCase : List[str] = str(Path(_A ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(_A ) == text_file def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Dict = str(tmp_path.resolve() / '__missing_file__.txt' ) with pytest.raises(_A ): cached_path(_A ) # relative path _lowerCAmelCase : Tuple = './__missing_file__.txt' with pytest.raises(_A ): cached_path(_A ) def lowercase (_A ): """simple docstring""" _lowerCAmelCase : str = get_from_cache(f'tmp://{tmpfs_file}' ) with open(_A ) as f: _lowerCAmelCase : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch('datasets.config.HF_DATASETS_OFFLINE' , _A ) def lowercase (): """simple docstring""" with pytest.raises(_A ): cached_path('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE' , _A ) def lowercase (_A ): """simple docstring""" _lowerCAmelCase : int = tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(_A ): http_get('https://huggingface.co' , temp_file=_A ) with pytest.raises(_A ): http_head('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE' , _A ) def lowercase (_A ): """simple docstring""" _lowerCAmelCase : int = tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(_A ): ftp_get('ftp://huggingface.co' , temp_file=_A ) with pytest.raises(_A ): ftp_head('ftp://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE' , _A ) def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(_A ): fsspec_get('s3://huggingface.co' , temp_file=_A ) with pytest.raises(_A ): fsspec_head('s3://huggingface.co' )
703
'''simple docstring''' from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
630
0
'''simple docstring''' import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = VideoToVideoSDPipeline __magic_name__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({"video"} ) - {"image", "width", "height"} __magic_name__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"video"} ) - {"image"} __magic_name__ = PipelineTesterMixin.required_optional_params - {"latents"} __magic_name__ = False # No `output_type`. __magic_name__ = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def a ( self ): '''simple docstring''' torch.manual_seed(0 ) _lowerCAmelCase : int = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'DownBlock3D') , up_block_types=('UpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D') , cross_attention_dim=32 , attention_head_dim=4 , ) _lowerCAmelCase : Dict = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , ) torch.manual_seed(0 ) _lowerCAmelCase : Tuple = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _lowerCAmelCase : int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) _lowerCAmelCase : str = CLIPTextModel(snake_case__ ) _lowerCAmelCase : Union[str, Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _lowerCAmelCase : List[str] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def a ( self , snake_case__ , snake_case__=0 ): '''simple docstring''' _lowerCAmelCase : Tuple = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith('mps' ): _lowerCAmelCase : Union[str, Any] = torch.manual_seed(snake_case__ ) else: _lowerCAmelCase : Dict = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) _lowerCAmelCase : int = { 'prompt': 'A painting of a squirrel eating a burger', 'video': video, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'pt', } return inputs def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase : str = self.get_dummy_components() _lowerCAmelCase : Dict = VideoToVideoSDPipeline(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = sd_pipe.to(snake_case__ ) sd_pipe.set_progress_bar_config(disable=snake_case__ ) _lowerCAmelCase : Any = self.get_dummy_inputs(snake_case__ ) _lowerCAmelCase : Dict = 'np' _lowerCAmelCase : List[str] = sd_pipe(**snake_case__ ).frames _lowerCAmelCase : Dict = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) _lowerCAmelCase : Tuple = np.array([106, 117, 113, 174, 137, 112, 148, 151, 131] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def a ( self ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case__ , expected_max_diff=5E-3 ) @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='`num_images_per_prompt` argument is not supported for this pipeline.' ) def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = VideoToVideoSDPipeline.from_pretrained('cerspense/zeroscope_v2_XL' , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames _lowerCAmelCase : str = torch.Generator(device='cpu' ).manual_seed(0 ) _lowerCAmelCase : str = torch.randn((1, 10, 3, 1024, 576) , generator=snake_case__ ) _lowerCAmelCase : List[Any] = video.to('cuda' ) _lowerCAmelCase : Optional[int] = 'Spiderman is surfing' _lowerCAmelCase : List[str] = pipe(snake_case__ , video=snake_case__ , generator=snake_case__ , num_inference_steps=3 , output_type='pt' ).frames _lowerCAmelCase : Any = np.array([-1.045_8984, -1.127_9297, -0.966_3086, -0.9150_3906, -0.7509_7656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1E-2
704
'''simple docstring''' import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase : int = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = {"""vocab_file""": """spiece.model"""} lowerCAmelCase : Optional[int] = { """vocab_file""": { """AI-Sweden/gpt-sw3-126m""": """https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-350m""": """https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-1.6b""": """https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-6.7b""": """https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-20b""": """https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model""", } } lowerCAmelCase : Union[str, Any] = { """AI-Sweden/gpt-sw3-126m""": 20_48, """AI-Sweden/gpt-sw3-350m""": 20_48, """AI-Sweden/gpt-sw3-1.6b""": 20_48, """AI-Sweden/gpt-sw3-6.7b""": 20_48, """AI-Sweden/gpt-sw3-20b""": 20_48, } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = ["input_ids", "attention_mask"] def __init__( self , snake_case__ , snake_case__=False , snake_case__=False , snake_case__=False , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__ = None , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : Tuple = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase : List[Any] = kwargs.get('name_or_path' ) if name_or_path is None: logger.warning( 'name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,' ' you are testing the model, this can safely be ignored' ) _lowerCAmelCase : Any = 'None' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing _lowerCAmelCase : str = '<|endoftext|>' if eos_token is None else eos_token _lowerCAmelCase : Tuple = '<unk>' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: _lowerCAmelCase : List[str] = unk_token if pad_token is None else pad_token _lowerCAmelCase : Optional[int] = eos_token if bos_token is None else bos_token else: _lowerCAmelCase : Tuple = '<pad>' if pad_token is None else pad_token _lowerCAmelCase : Union[str, Any] = '<s>' if bos_token is None else bos_token super().__init__( do_lower_case=snake_case__ , remove_space=snake_case__ , keep_accents=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , pad_token=snake_case__ , sp_model_kwargs=self.sp_model_kwargs , **snake_case__ , ) _lowerCAmelCase : Union[str, Any] = do_lower_case _lowerCAmelCase : Optional[int] = remove_space _lowerCAmelCase : Any = keep_accents _lowerCAmelCase : Optional[int] = vocab_file _lowerCAmelCase : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case__ ) # Used for whitespace normalization in input texts # fmt : off _lowerCAmelCase : Optional[Any] = {' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', '', '„'} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing _lowerCAmelCase : Optional[Any] = re.compile( F'[{"".join(map(snake_case__ , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]' ) def __getstate__( self ): '''simple docstring''' _lowerCAmelCase : List[str] = self.__dict__.copy() _lowerCAmelCase : int = None return state def __setstate__( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowerCAmelCase : int = {} _lowerCAmelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def a ( self ): '''simple docstring''' return len(self.sp_model ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.non_printing_characters_re.sub('' , snake_case__ ) # Normalize whitespaces _lowerCAmelCase : Tuple = ''.join([char if char not in self.whitespaces else ' ' for char in text] ) # NFC Unicode normalization _lowerCAmelCase : Union[str, Any] = unicodedata.normalize('NFC' , snake_case__ ) return text def a ( self , snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : str = self.preprocess_text(snake_case__ ) return self.sp_model.encode(snake_case__ , out_type=snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.PieceToId(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case__ ) @staticmethod def a ( snake_case__ ): '''simple docstring''' return out_string def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : int = [] _lowerCAmelCase : Optional[Any] = '' _lowerCAmelCase : Tuple = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(snake_case__ ) + token _lowerCAmelCase : Union[str, Any] = True _lowerCAmelCase : List[Any] = [] else: current_sub_tokens.append(snake_case__ ) _lowerCAmelCase : List[Any] = False out_string += self.sp_model.decode(snake_case__ ) return out_string def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = {self.convert_ids_to_tokens(snake_case__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _lowerCAmelCase : int = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case__ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case__ , 'wb' ) as fi: _lowerCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (out_vocab_file,) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' if isinstance(snake_case__ , snake_case__ ): _lowerCAmelCase : Optional[Any] = self.preprocess_text(snake_case__ ) _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) else: _lowerCAmelCase : Tuple = [self.preprocess_text(snake_case__ ) for t in text] _lowerCAmelCase : List[str] = self.sp_model.encode(snake_case__ ) if return_tensors is True or return_tensors == "pt": _lowerCAmelCase : int = torch.tensor(snake_case__ ) return token_ids def a ( self , snake_case__ ): '''simple docstring''' return self.sp_model.decode(snake_case__ ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = [F'User: {text}' if is_user else F'Bot: {text}' for is_user, text in conversation.iter_texts()] _lowerCAmelCase : str = ( F'{self.eos_token}{self.bos_token}' + F'{self.bos_token}'.join(snake_case__ ) + F'{self.bos_token}Bot:' ) return self.encode(text=snake_case__ )
630
0
import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class UpperCamelCase__ : """simple docstring""" def a ( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' return None class UpperCamelCase__ : """simple docstring""" def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' return None class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" __magic_name__ = [ # (model_name, model_kwargs) ("bert-base-cased", {}), ("gpt2", {"use_cache": False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def a ( self ): '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(snake_case__ , 'tf' , 12 , **snake_case__ ) @require_torch @slow def a ( self ): '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(snake_case__ , 'pt' , 12 , **snake_case__ ) @require_torch @slow def a ( self ): '''simple docstring''' from transformers import BertModel _lowerCAmelCase : Optional[int] = ['[UNK]', '[SEP]', '[CLS]', '[PAD]', '[MASK]', 'some', 'other', 'words'] with NamedTemporaryFile(mode='w+t' ) as vocab_file: vocab_file.write('\n'.join(snake_case__ ) ) vocab_file.flush() _lowerCAmelCase : List[Any] = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: _lowerCAmelCase : Optional[int] = BertModel(BertConfig(vocab_size=len(snake_case__ ) ) ) model.save_pretrained(snake_case__ ) self._test_export(snake_case__ , 'pt' , 12 , snake_case__ ) @require_tf @slow def a ( self ): '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCAmelCase : int = self._test_export(snake_case__ , 'tf' , 12 , **snake_case__ ) _lowerCAmelCase : str = quantize(Path(snake_case__ ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(snake_case__ ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) @require_torch @slow def a ( self ): '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCAmelCase : Dict = self._test_export(snake_case__ , 'pt' , 12 , **snake_case__ ) _lowerCAmelCase : Optional[Any] = quantize(snake_case__ ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(snake_case__ ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__=None , **snake_case__ ): '''simple docstring''' try: # Compute path with TemporaryDirectory() as tempdir: _lowerCAmelCase : List[str] = Path(snake_case__ ).joinpath('model.onnx' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , **snake_case__ ) return path except Exception as e: self.fail(snake_case__ ) @require_torch @require_tokenizers @slow def a ( self ): '''simple docstring''' from transformers import BertModel _lowerCAmelCase : str = BertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) _lowerCAmelCase : Dict = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(snake_case__ , snake_case__ , 'pt' ) @require_tf @require_tokenizers @slow def a ( self ): '''simple docstring''' from transformers import TFBertModel _lowerCAmelCase : str = TFBertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) _lowerCAmelCase : Dict = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(snake_case__ , snake_case__ , 'tf' ) def a ( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[str] = FeatureExtractionPipeline(snake_case__ , snake_case__ ) _lowerCAmelCase : Tuple = ['input_ids', 'token_type_ids', 'attention_mask', 'output_0', 'output_1'] _lowerCAmelCase : str = infer_shapes(snake_case__ , snake_case__ ) # Assert all variables are present self.assertEqual(len(snake_case__ ) , len(snake_case__ ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , snake_case__ ) self.assertSequenceEqual(variable_names[3:] , snake_case__ ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: 'batch', 1: 'sequence'} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['output_0'] , {0: 'batch', 1: 'sequence'} ) self.assertDictEqual(shapes['output_1'] , {0: 'batch'} ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = ['input_ids', 'attention_mask', 'token_type_ids'] _lowerCAmelCase : List[Any] = {'input_ids': [1, 2, 3, 4], 'attention_mask': [0, 0, 0, 0], 'token_type_ids': [1, 1, 1, 1]} _lowerCAmelCase : str = ensure_valid_input(FuncContiguousArgs() , snake_case__ , snake_case__ ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(snake_case__ ) , 3 ) # Should have exactly the same input names self.assertEqual(set(snake_case__ ) , set(snake_case__ ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(snake_case__ , (tokens['input_ids'], tokens['token_type_ids'], tokens['attention_mask']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) _lowerCAmelCase : List[Any] = ensure_valid_input(FuncNonContiguousArgs() , snake_case__ , snake_case__ ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(snake_case__ ) , 1 ) self.assertEqual(len(snake_case__ ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['input_ids'] ) self.assertEqual(ordered_input_names[0] , 'input_ids' ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = generate_identified_filename(Path('/home/something/my_fake_model.onnx' ) , '-test' ) self.assertEqual('/home/something/my_fake_model-test.onnx' , generated.as_posix() )
705
'''simple docstring''' import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = (DDPMScheduler,) def a ( self , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**snake_case__ ) return config def a ( self ): '''simple docstring''' for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=snake_case__ ) def a ( self ): '''simple docstring''' for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=snake_case__ , beta_end=snake_case__ ) def a ( self ): '''simple docstring''' for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=snake_case__ ) def a ( self ): '''simple docstring''' for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=snake_case__ ) def a ( self ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=snake_case__ ) def a ( self ): '''simple docstring''' self.check_over_configs(thresholding=snake_case__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=snake_case__ , prediction_type=snake_case__ , sample_max_value=snake_case__ , ) def a ( self ): '''simple docstring''' for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=snake_case__ ) def a ( self ): '''simple docstring''' for t in [0, 500, 999]: self.check_over_forward(time_step=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1E-5 def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[Any] = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : str = self.dummy_model() _lowerCAmelCase : str = self.dummy_sample_deter _lowerCAmelCase : Union[str, Any] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : List[Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Any = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Dict = pred_prev_sample _lowerCAmelCase : Dict = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : List[str] = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 258.9606 ) < 1E-2 assert abs(result_mean.item() - 0.3372 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.scheduler_classes[0] _lowerCAmelCase : Any = self.get_scheduler_config(prediction_type='v_prediction' ) _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = len(snake_case__ ) _lowerCAmelCase : Any = self.dummy_model() _lowerCAmelCase : Tuple = self.dummy_sample_deter _lowerCAmelCase : Optional[int] = torch.manual_seed(0 ) for t in reversed(range(snake_case__ ) ): # 1. predict noise residual _lowerCAmelCase : Union[str, Any] = model(snake_case__ , snake_case__ ) # 2. predict previous mean of sample x_t-1 _lowerCAmelCase : Dict = scheduler.step(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCAmelCase : Tuple = pred_prev_sample _lowerCAmelCase : Any = torch.sum(torch.abs(snake_case__ ) ) _lowerCAmelCase : Dict = torch.mean(torch.abs(snake_case__ ) ) assert abs(result_sum.item() - 202.0296 ) < 1E-2 assert abs(result_mean.item() - 0.2631 ) < 1E-3 def a ( self ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.scheduler_classes[0] _lowerCAmelCase : Optional[int] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=snake_case__ ) _lowerCAmelCase : Union[str, Any] = scheduler.timesteps for i, timestep in enumerate(snake_case__ ): if i == len(snake_case__ ) - 1: _lowerCAmelCase : str = -1 else: _lowerCAmelCase : Optional[Any] = timesteps[i + 1] _lowerCAmelCase : int = scheduler.previous_timestep(snake_case__ ) _lowerCAmelCase : int = prev_t.item() self.assertEqual(snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : Tuple = self.get_scheduler_config() _lowerCAmelCase : List[Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 51, 0] with self.assertRaises(snake_case__ , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : List[str] = self.get_scheduler_config() _lowerCAmelCase : Union[str, Any] = scheduler_class(**snake_case__ ) _lowerCAmelCase : Optional[int] = [100, 87, 50, 1, 0] _lowerCAmelCase : int = len(snake_case__ ) with self.assertRaises(snake_case__ , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=snake_case__ , timesteps=snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.scheduler_classes[0] _lowerCAmelCase : int = self.get_scheduler_config() _lowerCAmelCase : Any = scheduler_class(**snake_case__ ) _lowerCAmelCase : Any = [scheduler.config.num_train_timesteps] with self.assertRaises( snake_case__ , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=snake_case__ )
630
0
'''simple docstring''' def lowercase (_A = 1 , _A = 1_0_0_0 ): """simple docstring""" _lowerCAmelCase : str = 1 _lowerCAmelCase : Optional[Any] = 0 for divide_by_number in range(_A , digit + 1 ): _lowerCAmelCase : list[int] = [] _lowerCAmelCase : Dict = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(_A ): _lowerCAmelCase : List[str] = len(_A ) _lowerCAmelCase : int = divide_by_number else: has_been_divided.append(_A ) _lowerCAmelCase : str = now_divide * 1_0 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
706
'''simple docstring''' import socket def lowercase (): """simple docstring""" _lowerCAmelCase : Tuple = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCAmelCase : Optional[int] = socket.gethostname() _lowerCAmelCase : Any = 1_2_3_1_2 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCAmelCase : Union[str, Any] = sock.recv(1_0_2_4 ) if not data: break out_file.write(_A ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
630
0
'''simple docstring''' import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowercase (_A , _A , _A ): """simple docstring""" if openai_config_file == "": _lowerCAmelCase : str = OpenAIGPTConfig() else: _lowerCAmelCase : List[str] = OpenAIGPTConfig.from_json_file(_A ) _lowerCAmelCase : Union[str, Any] = OpenAIGPTModel(_A ) # Load weights from numpy load_tf_weights_in_openai_gpt(_A , _A , _A ) # Save pytorch-model _lowerCAmelCase : str = pytorch_dump_folder_path + '/' + WEIGHTS_NAME _lowerCAmelCase : int = pytorch_dump_folder_path + '/' + CONFIG_NAME print(f'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(model.state_dict() , _A ) print(f'Save configuration file to {pytorch_config_dump_path}' ) with open(_A , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCAmelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( """--openai_checkpoint_folder_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--openai_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) lowerCAmelCase : Optional[Any] = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
707
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase : Tuple = False lowerCAmelCase : str = True lowerCAmelCase : List[Any] = False if __name__ == "__main__": lowerCAmelCase : Any = argparse.ArgumentParser() parser.add_argument( """--repo_path""", default=None, type=str, required=True, help="""The config json file corresponding to the architecture.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") lowerCAmelCase : Optional[int] = parser.parse_args() lowerCAmelCase : int = { """image_size""": """sample_size""", """num_res_blocks""": """layers_per_block""", """block_channels""": """block_out_channels""", """down_blocks""": """down_block_types""", """up_blocks""": """up_block_types""", """downscale_freq_shift""": """freq_shift""", """resnet_num_groups""": """norm_num_groups""", """resnet_act_fn""": """act_fn""", """resnet_eps""": """norm_eps""", """num_head_channels""": """attention_head_dim""", } lowerCAmelCase : int = { """time_steps""": """time_proj""", """mid""": """mid_block""", """downsample_blocks""": """down_blocks""", """upsample_blocks""": """up_blocks""", } lowerCAmelCase : Optional[Any] = """""" if has_file(args.repo_path, """config.json""") else """unet""" with open(os.path.join(args.repo_path, subfolder, """config.json"""), """r""", encoding="""utf-8""") as reader: lowerCAmelCase : int = reader.read() lowerCAmelCase : List[str] = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, """config.json"""): lowerCAmelCase : str = UNetaDModel(**config) else: lowerCAmelCase : Union[str, Any] = UNetaDConditionModel if """ldm-text2im-large-256""" in args.repo_path else UNetaDModel lowerCAmelCase : Dict = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase : Union[str, Any] = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase : str = config[key] del config[key] lowerCAmelCase : Optional[int] = [k.replace("""UNetRes""", """""") for k in config["""down_block_types"""]] lowerCAmelCase : Dict = [k.replace("""UNetRes""", """""") for k in config["""up_block_types"""]] if do_only_weights: lowerCAmelCase : Tuple = torch.load(os.path.join(args.repo_path, subfolder, """diffusion_pytorch_model.bin""")) lowerCAmelCase : str = {} for param_key, param_value in state_dict.items(): if param_key.endswith(""".op.bias""") or param_key.endswith(""".op.weight"""): continue lowerCAmelCase : str = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split(""".""")[0] == key: lowerCAmelCase : Dict = param_value lowerCAmelCase : Tuple = True if not has_changed: lowerCAmelCase : Tuple = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
630
0
'''simple docstring''' import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline lowerCAmelCase : List[str] = argparse.ArgumentParser("""Stable Diffusion script with intel optimization""", add_help=False) parser.add_argument("""--dpm""", action="""store_true""", help="""Enable DPMSolver or not""") parser.add_argument("""--steps""", default=None, type=int, help="""Num inference steps""") lowerCAmelCase : Union[str, Any] = parser.parse_args() lowerCAmelCase : List[str] = """cpu""" lowerCAmelCase : Tuple = """a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings""" lowerCAmelCase : List[Any] = """path-to-your-trained-model""" lowerCAmelCase : List[str] = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: lowerCAmelCase : List[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) lowerCAmelCase : List[str] = pipe.to(device) # to channels last lowerCAmelCase : Optional[int] = pipe.unet.to(memory_format=torch.channels_last) lowerCAmelCase : Dict = pipe.vae.to(memory_format=torch.channels_last) lowerCAmelCase : Optional[Any] = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: lowerCAmelCase : Optional[int] = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex lowerCAmelCase : Dict = torch.randn(2, 4, 64, 64) lowerCAmelCase : List[Any] = torch.rand(1) * 9_99 lowerCAmelCase : str = torch.randn(2, 77, 7_68) lowerCAmelCase : Union[str, Any] = (sample, timestep, encoder_hidden_status) try: lowerCAmelCase : List[str] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: lowerCAmelCase : List[str] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) lowerCAmelCase : int = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) lowerCAmelCase : Dict = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: lowerCAmelCase : Optional[Any] = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute lowerCAmelCase : Optional[int] = 6_66 lowerCAmelCase : str = torch.Generator(device).manual_seed(seed) lowerCAmelCase : Any = {"""generator""": generator} if args.steps is not None: lowerCAmelCase : List[str] = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): lowerCAmelCase : Tuple = pipe(prompt, **generate_kwargs).images[0] # save image image.save("""generated.png""")
708
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = None ): '''simple docstring''' super().__init__() _lowerCAmelCase : Union[str, Any] = pad_token_id _lowerCAmelCase : List[Any] = max_length _lowerCAmelCase : Tuple = vocab _lowerCAmelCase : str = merges _lowerCAmelCase : List[str] = BytePairTokenizer(snake_case__ , snake_case__ , sequence_length=snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = [' '.join(snake_case__ ) for m in tokenizer.bpe_ranks.keys()] _lowerCAmelCase : Any = tokenizer.get_vocab() return cls(snake_case__ , snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = GPTaTokenizer.from_pretrained(snake_case__ , *snake_case__ , **snake_case__ ) return cls.from_tokenizer(snake_case__ , *snake_case__ , **snake_case__ ) @classmethod def a ( cls , snake_case__ ): '''simple docstring''' return cls(**snake_case__ ) def a ( self ): '''simple docstring''' return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' _lowerCAmelCase : str = self.tf_tokenizer(snake_case__ ) _lowerCAmelCase : str = tf.ones_like(snake_case__ ) if self.pad_token_id is not None: # pad the tokens up to max length _lowerCAmelCase : Optional[int] = max_length if max_length is not None else self.max_length if max_length is not None: _lowerCAmelCase , _lowerCAmelCase : str = pad_model_inputs( snake_case__ , max_seq_length=snake_case__ , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
630
0
'''simple docstring''' import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase : Any = logging.get_logger(__name__) lowerCAmelCase : Any = {"""vocab_file""": """vocab.json"""} lowerCAmelCase : Optional[int] = { """vocab_file""": { """mgp-str""": """https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json""", } } lowerCAmelCase : Dict = {"""mgp-str""": 27} class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , snake_case__ , snake_case__="[GO]" , snake_case__="[GO]" , snake_case__="[s]" , snake_case__="[GO]" , **snake_case__ ): '''simple docstring''' super().__init__( unk_token=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , pad_token=snake_case__ , **snake_case__ , ) with open(snake_case__ , encoding='utf-8' ) as vocab_handle: _lowerCAmelCase : Union[str, Any] = json.load(snake_case__ ) _lowerCAmelCase : int = {v: k for k, v in self.vocab.items()} @property def a ( self ): '''simple docstring''' return len(self.vocab ) def a ( self ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Tuple = [] for s in text: char_tokens.extend(snake_case__ ) return char_tokens def a ( self , snake_case__ ): '''simple docstring''' return self.vocab.get(snake_case__ , self.vocab.get(self.unk_token ) ) def a ( self , snake_case__ ): '''simple docstring''' return self.decoder.get(snake_case__ ) def a ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error('Vocabulary path ({}) should be a directory'.format(snake_case__ ) ) return _lowerCAmelCase : List[Any] = os.path.join( snake_case__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) with open(snake_case__ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=snake_case__ , ensure_ascii=snake_case__ ) + '\n' ) return (vocab_file,)
709
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Optional[int] = { """configuration_mctct""": ["""MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MCTCTConfig"""], """feature_extraction_mctct""": ["""MCTCTFeatureExtractor"""], """processing_mctct""": ["""MCTCTProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : List[str] = [ """MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MCTCTForCTC""", """MCTCTModel""", """MCTCTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys lowerCAmelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=True , snake_case__=True , snake_case__=99 , snake_case__=32 , snake_case__=5 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=16 , snake_case__=2 , snake_case__=0.02 , snake_case__=3 , snake_case__=4 , snake_case__=None , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = parent _lowerCAmelCase : Any = batch_size _lowerCAmelCase : Any = seq_length _lowerCAmelCase : Union[str, Any] = is_training _lowerCAmelCase : Tuple = use_token_type_ids _lowerCAmelCase : int = use_labels _lowerCAmelCase : List[Any] = vocab_size _lowerCAmelCase : Any = hidden_size _lowerCAmelCase : int = num_hidden_layers _lowerCAmelCase : List[str] = num_attention_heads _lowerCAmelCase : Optional[Any] = intermediate_size _lowerCAmelCase : Optional[int] = hidden_act _lowerCAmelCase : str = hidden_dropout_prob _lowerCAmelCase : int = attention_probs_dropout_prob _lowerCAmelCase : Union[str, Any] = max_position_embeddings _lowerCAmelCase : int = type_vocab_size _lowerCAmelCase : Union[str, Any] = type_sequence_label_size _lowerCAmelCase : Optional[int] = initializer_range _lowerCAmelCase : Tuple = num_labels _lowerCAmelCase : List[str] = num_choices _lowerCAmelCase : List[str] = scope _lowerCAmelCase : Optional[Any] = self.vocab_size - 1 def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase : List[str] = None if self.use_token_type_ids: _lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCAmelCase : int = None _lowerCAmelCase : List[Any] = None _lowerCAmelCase : Optional[int] = None if self.use_labels: _lowerCAmelCase : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCAmelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCAmelCase : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) _lowerCAmelCase : Tuple = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) _lowerCAmelCase : Any = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , *snake_case__ ): '''simple docstring''' _lowerCAmelCase : str = OpenAIGPTModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : str = model(snake_case__ , token_type_ids=snake_case__ , head_mask=snake_case__ ) _lowerCAmelCase : int = model(snake_case__ , token_type_ids=snake_case__ ) _lowerCAmelCase : Optional[int] = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , *snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = OpenAIGPTLMHeadModel(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Any = model(snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , *snake_case__ ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = OpenAIGPTDoubleHeadsModel(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : List[str] = model(snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , *snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[str] = self.num_labels _lowerCAmelCase : str = OpenAIGPTForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCAmelCase : Union[str, Any] = model(snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = self.prepare_config_and_inputs() ( _lowerCAmelCase ) : int = config_and_inputs _lowerCAmelCase : Any = { 'input_ids': input_ids, 'token_type_ids': token_type_ids, 'head_mask': head_mask, } return config, inputs_dict @require_torch class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) __magic_name__ = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly __magic_name__ = ( { "feature-extraction": OpenAIGPTModel, "text-classification": OpenAIGPTForSequenceClassification, "text-generation": OpenAIGPTLMHeadModel, "zero-shot": OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def a ( self , snake_case__ , snake_case__ , snake_case__=False ): '''simple docstring''' _lowerCAmelCase : List[Any] = super()._prepare_for_class(snake_case__ , snake_case__ , return_labels=snake_case__ ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": _lowerCAmelCase : Optional[Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=snake_case__ , ) _lowerCAmelCase : Tuple = inputs_dict['labels'] _lowerCAmelCase : Dict = inputs_dict['labels'] _lowerCAmelCase : Optional[Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=snake_case__ , ) _lowerCAmelCase : List[str] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=snake_case__ ) return inputs_dict def a ( self ): '''simple docstring''' _lowerCAmelCase : str = OpenAIGPTModelTester(self ) _lowerCAmelCase : Union[str, Any] = ConfigTester(self , config_class=snake_case__ , n_embd=37 ) def a ( self ): '''simple docstring''' self.config_tester.run_common_tests() def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*snake_case__ ) @slow def a ( self ): '''simple docstring''' for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase : Optional[Any] = OpenAIGPTModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) @require_torch class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = OpenAIGPTLMHeadModel.from_pretrained('openai-gpt' ) model.to(snake_case__ ) _lowerCAmelCase : int = torch.tensor([[481, 4735, 544]] , dtype=torch.long , device=snake_case__ ) # the president is _lowerCAmelCase : str = [ 481, 4735, 544, 246, 963, 870, 762, 239, 244, 4_0477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the _lowerCAmelCase : Dict = model.generate(snake_case__ , do_sample=snake_case__ ) self.assertListEqual(output_ids[0].tolist() , snake_case__ )
710
'''simple docstring''' lowerCAmelCase : Optional[int] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_00_00)] def lowercase (_A ): """simple docstring""" _lowerCAmelCase : str = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 1_0_0_0_0_0] number //= 1_0_0_0_0_0 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution lowerCAmelCase : list[bool | None] = [None] * 10_00_00_00 lowerCAmelCase : List[str] = True lowerCAmelCase : Union[str, Any] = False def lowercase (_A ): """simple docstring""" if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _lowerCAmelCase : Any = chain(next_number(_A ) ) _lowerCAmelCase : List[str] = number_chain while number < 1_0_0_0_0_0_0_0: _lowerCAmelCase : Tuple = number_chain number *= 1_0 return number_chain def lowercase (_A = 1_0_0_0_0_0_0_0 ): """simple docstring""" for i in range(1 , _A ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(_A ) if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution() = }''')
630
0
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCAmelCase : str = logging.get_logger(__name__) # General docstring lowerCAmelCase : Optional[Any] = """RegNetConfig""" # Base docstring lowerCAmelCase : int = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = [1, 10_88, 7, 7] # Image classification docstring lowerCAmelCase : Any = """facebook/regnet-y-040""" lowerCAmelCase : Optional[Any] = """tabby, tabby cat""" lowerCAmelCase : Tuple = [ """facebook/regnet-y-040""", # See all regnet models at https://huggingface.co/models?filter=regnet ] class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 3 , snake_case__ = 1 , snake_case__ = 1 , snake_case__ = "relu" , **snake_case__ , ): '''simple docstring''' super().__init__(**snake_case__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _lowerCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _lowerCAmelCase : List[Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=snake_case__ , strides=snake_case__ , padding='VALID' , groups=snake_case__ , use_bias=snake_case__ , name='convolution' , ) _lowerCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) _lowerCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.convolution(self.padding(snake_case__ ) ) _lowerCAmelCase : Union[str, Any] = self.normalization(snake_case__ ) _lowerCAmelCase : int = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = config.num_channels _lowerCAmelCase : List[Any] = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name='embedder' , ) def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = shape_list(snake_case__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( 'Make sure that the channel dimension of the pixel values match with the one set in the configuration.' ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _lowerCAmelCase : List[Any] = tf.transpose(snake_case__ , perm=(0, 2, 3, 1) ) _lowerCAmelCase : Tuple = self.embedder(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = tf.keras.layers.ConvaD( filters=snake_case__ , kernel_size=1 , strides=snake_case__ , use_bias=snake_case__ , name='convolution' ) _lowerCAmelCase : Tuple = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='normalization' ) def a ( self , snake_case__ , snake_case__ = False ): '''simple docstring''' return self.normalization(self.convolution(snake_case__ ) , training=snake_case__ ) class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Any = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) _lowerCAmelCase : str = [ tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='relu' , name='attention.0' ), tf.keras.layers.ConvaD(filters=snake_case__ , kernel_size=1 , activation='sigmoid' , name='attention.2' ), ] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = self.pooler(snake_case__ ) for layer_module in self.attention: _lowerCAmelCase : Tuple = layer_module(snake_case__ ) _lowerCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Optional[int] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _lowerCAmelCase : Any = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.2' ), ] _lowerCAmelCase : List[str] = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = hidden_state for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) _lowerCAmelCase : int = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : Tuple = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : List[str] = in_channels != out_channels or stride != 1 _lowerCAmelCase : Union[str, Any] = max(1 , out_channels // config.groups_width ) _lowerCAmelCase : Optional[Any] = ( TFRegNetShortCut(snake_case__ , stride=snake_case__ , name='shortcut' ) if should_apply_shortcut else tf.keras.layers.Activation('linear' , name='shortcut' ) ) _lowerCAmelCase : Tuple = [ TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=config.hidden_act , name='layer.0' ), TFRegNetConvLayer( snake_case__ , stride=snake_case__ , groups=snake_case__ , activation=config.hidden_act , name='layer.1' ), TFRegNetSELayer(snake_case__ , reduced_channels=int(round(in_channels / 4 ) ) , name='layer.2' ), TFRegNetConvLayer(snake_case__ , kernel_size=1 , activation=snake_case__ , name='layer.3' ), ] _lowerCAmelCase : Tuple = ACTaFN[config.hidden_act] def a ( self , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Dict = hidden_state for layer_module in self.layers: _lowerCAmelCase : List[Any] = layer_module(snake_case__ ) _lowerCAmelCase : Tuple = self.shortcut(snake_case__ ) hidden_state += residual _lowerCAmelCase : str = self.activation(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ = 2 , snake_case__ = 2 , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Dict = TFRegNetXLayer if config.layer_type == 'x' else TFRegNetYLayer _lowerCAmelCase : Optional[int] = [ # downsampling is done in the first layer with stride of 2 layer(snake_case__ , snake_case__ , snake_case__ , stride=snake_case__ , name='layers.0' ), *[layer(snake_case__ , snake_case__ , snake_case__ , name=F'layers.{i+1}' ) for i in range(depth - 1 )], ] def a ( self , snake_case__ ): '''simple docstring''' for layer_module in self.layers: _lowerCAmelCase : int = layer_module(snake_case__ ) return hidden_state class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : str = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( snake_case__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name='stages.0' , ) ) _lowerCAmelCase : Union[str, Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(snake_case__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(snake_case__ , snake_case__ , snake_case__ , depth=snake_case__ , name=F'stages.{i+1}' ) ) def a ( self , snake_case__ , snake_case__ = False , snake_case__ = True ): '''simple docstring''' _lowerCAmelCase : List[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _lowerCAmelCase : str = hidden_states + (hidden_state,) _lowerCAmelCase : List[str] = stage_module(snake_case__ ) if output_hidden_states: _lowerCAmelCase : Dict = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=snake_case__ , hidden_states=snake_case__ ) @keras_serializable class UpperCamelCase__ ( tf.keras.layers.Layer ): """simple docstring""" __magic_name__ = RegNetConfig def __init__( self , snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(**snake_case__ ) _lowerCAmelCase : Union[str, Any] = config _lowerCAmelCase : Union[str, Any] = TFRegNetEmbeddings(snake_case__ , name='embedder' ) _lowerCAmelCase : Optional[int] = TFRegNetEncoder(snake_case__ , name='encoder' ) _lowerCAmelCase : Dict = tf.keras.layers.GlobalAveragePoolingaD(keepdims=snake_case__ , name='pooler' ) @unpack_inputs def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' _lowerCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : int = self.embedder(snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[str] = self.encoder( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : List[Any] = encoder_outputs[0] _lowerCAmelCase : Tuple = self.pooler(snake_case__ ) # Change to NCHW output format have uniformity in the modules _lowerCAmelCase : Optional[int] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) _lowerCAmelCase : Optional[Any] = tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _lowerCAmelCase : Union[str, Any] = tuple([tf.transpose(snake_case__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=snake_case__ , pooler_output=snake_case__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = RegNetConfig __magic_name__ = "regnet" __magic_name__ = "pixel_values" @property def a ( self ): '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 224, 224) , dtype=tf.floataa )} lowerCAmelCase : List[Any] = r""" Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. """ lowerCAmelCase : Dict = r""" Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : List[str] = TFRegNetMainLayer(snake_case__ , name='regnet' ) @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def a ( self , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : Any = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : str = self.regnet( pixel_values=snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , SCREAMING_SNAKE_CASE_ , ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , *snake_case__ , **snake_case__ ): '''simple docstring''' super().__init__(snake_case__ , *snake_case__ , **snake_case__ ) _lowerCAmelCase : Optional[Any] = config.num_labels _lowerCAmelCase : Optional[Any] = TFRegNetMainLayer(snake_case__ , name='regnet' ) # classification head _lowerCAmelCase : Optional[int] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name='classifier.1' ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(snake_case__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=snake_case__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def a ( self , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__ = None , snake_case__=False , ): '''simple docstring''' _lowerCAmelCase : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _lowerCAmelCase : Dict = self.regnet( snake_case__ , output_hidden_states=snake_case__ , return_dict=snake_case__ , training=snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.pooler_output if return_dict else outputs[1] _lowerCAmelCase : List[Any] = self.classifier[0](snake_case__ ) _lowerCAmelCase : Tuple = self.classifier[1](snake_case__ ) _lowerCAmelCase : int = None if labels is None else self.hf_compute_loss(labels=snake_case__ , logits=snake_case__ ) if not return_dict: _lowerCAmelCase : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=snake_case__ , logits=snake_case__ , hidden_states=outputs.hidden_states )
711
'''simple docstring''' import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(snake_case__ , 'width_multiplier' ) ) class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__=13 , snake_case__=64 , snake_case__=2 , snake_case__=3 , snake_case__="swish" , snake_case__=3 , snake_case__=32 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=True , snake_case__=True , snake_case__=10 , snake_case__=None , snake_case__=0.25 , snake_case__=0.0 , snake_case__=0.0 , ): '''simple docstring''' _lowerCAmelCase : Tuple = parent _lowerCAmelCase : Optional[int] = batch_size _lowerCAmelCase : List[Any] = image_size _lowerCAmelCase : List[Any] = patch_size _lowerCAmelCase : List[str] = num_channels _lowerCAmelCase : Optional[Any] = make_divisible(512 * width_multiplier , divisor=8 ) _lowerCAmelCase : Optional[Any] = hidden_act _lowerCAmelCase : List[Any] = conv_kernel_size _lowerCAmelCase : Optional[Any] = output_stride _lowerCAmelCase : List[Any] = classifier_dropout_prob _lowerCAmelCase : str = use_labels _lowerCAmelCase : List[str] = is_training _lowerCAmelCase : Optional[int] = num_labels _lowerCAmelCase : List[str] = initializer_range _lowerCAmelCase : str = scope _lowerCAmelCase : Any = width_multiplier _lowerCAmelCase : Union[str, Any] = ffn_dropout _lowerCAmelCase : Optional[int] = attn_dropout def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase : Optional[Any] = None _lowerCAmelCase : Dict = None if self.use_labels: _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) _lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _lowerCAmelCase : Optional[int] = self.get_config() return config, pixel_values, labels, pixel_labels def a ( self ): '''simple docstring''' return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : str = model(snake_case__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.num_labels _lowerCAmelCase : List[Any] = MobileViTVaForImageClassification(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : int = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' _lowerCAmelCase : List[Any] = self.num_labels _lowerCAmelCase : Optional[int] = MobileViTVaForSemanticSegmentation(snake_case__ ) model.to(snake_case__ ) model.eval() _lowerCAmelCase : Dict = model(snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) _lowerCAmelCase : Any = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a ( self ): '''simple docstring''' _lowerCAmelCase : int = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Optional[int] = config_and_inputs _lowerCAmelCase : Tuple = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __magic_name__ = ( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) __magic_name__ = ( { "feature-extraction": MobileViTVaModel, "image-classification": MobileViTVaForImageClassification, "image-segmentation": MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False def a ( self ): '''simple docstring''' _lowerCAmelCase : int = MobileViTVaModelTester(self ) _lowerCAmelCase : Dict = MobileViTVaConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ ) def a ( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='MobileViTV2 does not use inputs_embeds' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not support input and output embeddings' ) def a ( self ): '''simple docstring''' pass @unittest.skip(reason='MobileViTV2 does not output attentions' ) def a ( self ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason='Got `CUDA error: misaligned address` for tests after this one being run.' ) def a ( self ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def a ( self ): '''simple docstring''' pass def a ( self ): '''simple docstring''' _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : str = model_class(snake_case__ ) _lowerCAmelCase : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCAmelCase : int = [*signature.parameters.keys()] _lowerCAmelCase : Any = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def a ( self ): '''simple docstring''' def check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ): _lowerCAmelCase : Dict = model_class(snake_case__ ) model.to(snake_case__ ) model.eval() with torch.no_grad(): _lowerCAmelCase : Optional[Any] = model(**self._prepare_for_class(snake_case__ , snake_case__ ) ) _lowerCAmelCase : List[str] = outputs.hidden_states _lowerCAmelCase : List[str] = 5 self.assertEqual(len(snake_case__ ) , snake_case__ ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. _lowerCAmelCase : List[Any] = 2 for i in range(len(snake_case__ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase : Optional[int] = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCAmelCase : Any = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*snake_case__ ) @slow def a ( self ): '''simple docstring''' for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase : Dict = MobileViTVaModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) def lowercase (): """simple docstring""" _lowerCAmelCase : Any = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def a ( self ): '''simple docstring''' return ( MobileViTImageProcessor.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ) if is_vision_available() else None ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = MobileViTVaForImageClassification.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ).to( snake_case__ ) _lowerCAmelCase : str = self.default_image_processor _lowerCAmelCase : Any = prepare_img() _lowerCAmelCase : Optional[int] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Tuple = model(**snake_case__ ) # verify the logits _lowerCAmelCase : Optional[Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , snake_case__ ) _lowerCAmelCase : Tuple = torch.tensor([-1.6336E00, -7.3204E-02, -5.1883E-01] ).to(snake_case__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Any = model.to(snake_case__ ) _lowerCAmelCase : int = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Optional[int] = prepare_img() _lowerCAmelCase : Dict = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : int = model(**snake_case__ ) _lowerCAmelCase : Dict = outputs.logits # verify the logits _lowerCAmelCase : str = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , snake_case__ ) _lowerCAmelCase : Any = torch.tensor( [ [[7.0863, 7.1525, 6.8201], [6.6931, 6.8770, 6.8933], [6.2978, 7.0366, 6.9636]], [[-3.7134, -3.6712, -3.6675], [-3.5825, -3.3549, -3.4777], [-3.3435, -3.3979, -3.2857]], [[-2.9329, -2.8003, -2.7369], [-3.0564, -2.4780, -2.0207], [-2.6889, -1.9298, -1.7640]], ] , device=snake_case__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , snake_case__ , atol=1E-4 ) ) @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : Any = MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : List[Any] = model.to(snake_case__ ) _lowerCAmelCase : str = MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) _lowerCAmelCase : Tuple = prepare_img() _lowerCAmelCase : List[str] = image_processor(images=snake_case__ , return_tensors='pt' ).to(snake_case__ ) # forward pass with torch.no_grad(): _lowerCAmelCase : Any = model(**snake_case__ ) _lowerCAmelCase : Optional[Any] = outputs.logits.detach().cpu() _lowerCAmelCase : Any = image_processor.post_process_semantic_segmentation(outputs=snake_case__ , target_sizes=[(50, 60)] ) _lowerCAmelCase : List[Any] = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , snake_case__ ) _lowerCAmelCase : List[str] = image_processor.post_process_semantic_segmentation(outputs=snake_case__ ) _lowerCAmelCase : Tuple = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , snake_case__ )
630
0
'''simple docstring''' import warnings from .state import AcceleratorState, GradientState warnings.filterwarnings("""ignore""", category=UserWarning, module="""torch.optim.lr_scheduler""") class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ = True , snake_case__ = False ): '''simple docstring''' _lowerCAmelCase : str = scheduler _lowerCAmelCase : Optional[int] = optimizers if isinstance(snake_case__ , (list, tuple) ) else [optimizers] _lowerCAmelCase : Optional[int] = split_batches _lowerCAmelCase : Any = step_with_optimizer _lowerCAmelCase : Tuple = GradientState() def a ( self , *snake_case__ , **snake_case__ ): '''simple docstring''' if not self.step_with_optimizer: # No link between scheduler and optimizer -> just step self.scheduler.step(*snake_case__ , **snake_case__ ) return # Otherwise, first make sure the optimizer was stepped. if not self.gradient_state.sync_gradients: if self.gradient_state.adjust_scheduler: self.scheduler._step_count += 1 return for opt in self.optimizers: if opt.step_was_skipped: return if self.split_batches: # Split batches -> the training dataloader batch size is not changed so one step per training step self.scheduler.step(*snake_case__ , **snake_case__ ) else: # Otherwise the training dataloader batch size was multiplied by `num_processes`, so we need to do # num_processes steps per training step _lowerCAmelCase : Dict = AcceleratorState().num_processes for _ in range(snake_case__ ): # Special case when using OneCycle and `drop_last` was not used if hasattr(self.scheduler , 'total_steps' ): if self.scheduler._step_count <= self.scheduler.total_steps: self.scheduler.step(*snake_case__ , **snake_case__ ) else: self.scheduler.step(*snake_case__ , **snake_case__ ) def a ( self ): '''simple docstring''' return self.scheduler.get_last_lr() def a ( self ): '''simple docstring''' return self.scheduler.state_dict() def a ( self , snake_case__ ): '''simple docstring''' self.scheduler.load_state_dict(snake_case__ ) def a ( self ): '''simple docstring''' return self.scheduler.get_lr() def a ( self , *snake_case__ , **snake_case__ ): '''simple docstring''' return self.scheduler.print_lr(*snake_case__ , **snake_case__ )
712
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @slow def a ( self ): '''simple docstring''' _lowerCAmelCase : List[str] = FlaxXLMRobertaModel.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained('xlm-roberta-base' ) _lowerCAmelCase : Dict = 'The dog is cute and lives in the garden house' _lowerCAmelCase : List[str] = jnp.array([tokenizer.encode(snake_case__ )] ) _lowerCAmelCase : Optional[int] = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim _lowerCAmelCase : Tuple = jnp.array( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) _lowerCAmelCase : Union[str, Any] = model(snake_case__ )['last_hidden_state'] self.assertEqual(output.shape , snake_case__ ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , snake_case__ , atol=1E-3 ) )
630
0
'''simple docstring''' from __future__ import annotations def lowercase (_A ): """simple docstring""" _lowerCAmelCase : List[Any] = [True] * limit _lowerCAmelCase : Optional[int] = False _lowerCAmelCase : Dict = False _lowerCAmelCase : Optional[Any] = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): _lowerCAmelCase : Tuple = i * 2 while index < limit: _lowerCAmelCase : Union[str, Any] = False _lowerCAmelCase : Any = index + i _lowerCAmelCase : Dict = [2] for i in range(3 , _A , 2 ): if is_prime[i]: primes.append(_A ) return primes def lowercase (_A = 1_0_0_0_0_0_0 ): """simple docstring""" _lowerCAmelCase : List[Any] = prime_sieve(_A ) _lowerCAmelCase : Optional[Any] = 0 _lowerCAmelCase : Any = 0 for i in range(len(_A ) ): for j in range(i + length , len(_A ) ): _lowerCAmelCase : List[str] = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: _lowerCAmelCase : Union[str, Any] = j - i _lowerCAmelCase : str = sol return largest if __name__ == "__main__": print(F'''{solution() = }''')
713
'''simple docstring''' def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Dict = len(_A ) while cur > 1: # Find the maximum number in arr _lowerCAmelCase : int = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi _lowerCAmelCase : Dict = arr[mi::-1] + arr[mi + 1 : len(_A )] # Reverse whole list _lowerCAmelCase : Optional[int] = arr[cur - 1 :: -1] + arr[cur : len(_A )] cur -= 1 return arr if __name__ == "__main__": lowerCAmelCase : Tuple = input("""Enter numbers separated by a comma:\n""").strip() lowerCAmelCase : Tuple = [int(item) for item in user_input.split(""",""")] print(pancake_sort(unsorted))
630
0
from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) lowerCAmelCase : str = { """EleutherAI/gpt-j-6B""": """https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json""", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "gptj" __magic_name__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , snake_case__=5_0400 , snake_case__=2048 , snake_case__=4096 , snake_case__=28 , snake_case__=16 , snake_case__=64 , snake_case__=None , snake_case__="gelu_new" , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=1E-5 , snake_case__=0.02 , snake_case__=True , snake_case__=5_0256 , snake_case__=5_0256 , snake_case__=False , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : int = vocab_size _lowerCAmelCase : Optional[int] = n_positions _lowerCAmelCase : Optional[int] = n_embd _lowerCAmelCase : Optional[int] = n_layer _lowerCAmelCase : str = n_head _lowerCAmelCase : Tuple = n_inner _lowerCAmelCase : Tuple = rotary_dim _lowerCAmelCase : Optional[int] = activation_function _lowerCAmelCase : Any = resid_pdrop _lowerCAmelCase : List[str] = embd_pdrop _lowerCAmelCase : int = attn_pdrop _lowerCAmelCase : Any = layer_norm_epsilon _lowerCAmelCase : Optional[int] = initializer_range _lowerCAmelCase : List[str] = use_cache _lowerCAmelCase : Dict = bos_token_id _lowerCAmelCase : Any = eos_token_id super().__init__( bos_token_id=snake_case__ , eos_token_id=snake_case__ , tie_word_embeddings=snake_case__ , **snake_case__ ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = "default" , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' super().__init__(snake_case__ , task=snake_case__ , patching_specs=snake_case__ , use_past=snake_case__ ) if not getattr(self._config , 'pad_token_id' , snake_case__ ): # TODO: how to do that better? _lowerCAmelCase : Any = 0 @property def a ( self ): '''simple docstring''' _lowerCAmelCase : str = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(snake_case__ , direction='inputs' ) _lowerCAmelCase : int = {0: 'batch', 1: 'past_sequence + sequence'} else: _lowerCAmelCase : int = {0: 'batch', 1: 'sequence'} return common_inputs @property def a ( self ): '''simple docstring''' return self._config.n_layer @property def a ( self ): '''simple docstring''' return self._config.n_head def a ( self , snake_case__ , snake_case__ = -1 , snake_case__ = -1 , snake_case__ = False , snake_case__ = None , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = super(snake_case__ , self ).generate_dummy_inputs( snake_case__ , batch_size=snake_case__ , seq_length=snake_case__ , is_pair=snake_case__ , framework=snake_case__ ) # We need to order the input in the way they appears in the forward() _lowerCAmelCase : Any = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowerCAmelCase : Optional[Any] = common_inputs['input_ids'].shape # Not using the same length for past_key_values _lowerCAmelCase : Any = seqlen + 2 _lowerCAmelCase : Optional[int] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _lowerCAmelCase : Tuple = [ (torch.zeros(snake_case__ ), torch.zeros(snake_case__ )) for _ in range(self.num_layers ) ] _lowerCAmelCase : Tuple = common_inputs['attention_mask'] if self.use_past: _lowerCAmelCase : Any = ordered_inputs['attention_mask'].dtype _lowerCAmelCase : Union[str, Any] = torch.cat( [ordered_inputs['attention_mask'], torch.ones(snake_case__ , snake_case__ , dtype=snake_case__ )] , dim=1 ) return ordered_inputs @property def a ( self ): '''simple docstring''' return 13
714
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) lowerCAmelCase : str = { """EleutherAI/gpt-j-6B""": """https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json""", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "gptj" __magic_name__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , snake_case__=5_0400 , snake_case__=2048 , snake_case__=4096 , snake_case__=28 , snake_case__=16 , snake_case__=64 , snake_case__=None , snake_case__="gelu_new" , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=1E-5 , snake_case__=0.02 , snake_case__=True , snake_case__=5_0256 , snake_case__=5_0256 , snake_case__=False , **snake_case__ , ): '''simple docstring''' _lowerCAmelCase : int = vocab_size _lowerCAmelCase : Optional[int] = n_positions _lowerCAmelCase : Optional[int] = n_embd _lowerCAmelCase : Optional[int] = n_layer _lowerCAmelCase : str = n_head _lowerCAmelCase : Tuple = n_inner _lowerCAmelCase : Tuple = rotary_dim _lowerCAmelCase : Optional[int] = activation_function _lowerCAmelCase : Any = resid_pdrop _lowerCAmelCase : List[str] = embd_pdrop _lowerCAmelCase : int = attn_pdrop _lowerCAmelCase : Any = layer_norm_epsilon _lowerCAmelCase : Optional[int] = initializer_range _lowerCAmelCase : List[str] = use_cache _lowerCAmelCase : Dict = bos_token_id _lowerCAmelCase : Any = eos_token_id super().__init__( bos_token_id=snake_case__ , eos_token_id=snake_case__ , tie_word_embeddings=snake_case__ , **snake_case__ ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = "default" , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' super().__init__(snake_case__ , task=snake_case__ , patching_specs=snake_case__ , use_past=snake_case__ ) if not getattr(self._config , 'pad_token_id' , snake_case__ ): # TODO: how to do that better? _lowerCAmelCase : Any = 0 @property def a ( self ): '''simple docstring''' _lowerCAmelCase : str = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(snake_case__ , direction='inputs' ) _lowerCAmelCase : int = {0: 'batch', 1: 'past_sequence + sequence'} else: _lowerCAmelCase : int = {0: 'batch', 1: 'sequence'} return common_inputs @property def a ( self ): '''simple docstring''' return self._config.n_layer @property def a ( self ): '''simple docstring''' return self._config.n_head def a ( self , snake_case__ , snake_case__ = -1 , snake_case__ = -1 , snake_case__ = False , snake_case__ = None , ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = super(snake_case__ , self ).generate_dummy_inputs( snake_case__ , batch_size=snake_case__ , seq_length=snake_case__ , is_pair=snake_case__ , framework=snake_case__ ) # We need to order the input in the way they appears in the forward() _lowerCAmelCase : Any = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = common_inputs['input_ids'].shape # Not using the same length for past_key_values _lowerCAmelCase : Any = seqlen + 2 _lowerCAmelCase : Optional[int] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _lowerCAmelCase : Tuple = [ (torch.zeros(snake_case__ ), torch.zeros(snake_case__ )) for _ in range(self.num_layers ) ] _lowerCAmelCase : Tuple = common_inputs['attention_mask'] if self.use_past: _lowerCAmelCase : Any = ordered_inputs['attention_mask'].dtype _lowerCAmelCase : Union[str, Any] = torch.cat( [ordered_inputs['attention_mask'], torch.ones(snake_case__ , snake_case__ , dtype=snake_case__ )] , dim=1 ) return ordered_inputs @property def a ( self ): '''simple docstring''' return 13
630
0
'''simple docstring''' from typing import Dict, Optional import numpy as np import datasets lowerCAmelCase : Optional[Any] = """ IoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union between the predicted segmentation and the ground truth. For binary (two classes) or multi-class segmentation, the mean IoU of the image is calculated by taking the IoU of each class and averaging them. """ lowerCAmelCase : Tuple = """ Args: predictions (`List[ndarray]`): List of predicted segmentation maps, each of shape (height, width). Each segmentation map can be of a different size. references (`List[ndarray]`): List of ground truth segmentation maps, each of shape (height, width). Each segmentation map can be of a different size. num_labels (`int`): Number of classes (categories). ignore_index (`int`): Index that will be ignored during evaluation. nan_to_num (`int`, *optional*): If specified, NaN values will be replaced by the number defined by the user. label_map (`dict`, *optional*): If specified, dictionary mapping old label indices to new label indices. reduce_labels (`bool`, *optional*, defaults to `False`): Whether or not to reduce all label values of segmentation maps by 1. Usually used for datasets where 0 is used for background, and background itself is not included in all classes of a dataset (e.g. ADE20k). The background label will be replaced by 255. Returns: `Dict[str, float | ndarray]` comprising various elements: - *mean_iou* (`float`): Mean Intersection-over-Union (IoU averaged over all categories). - *mean_accuracy* (`float`): Mean accuracy (averaged over all categories). - *overall_accuracy* (`float`): Overall accuracy on all images. - *per_category_accuracy* (`ndarray` of shape `(num_labels,)`): Per category accuracy. - *per_category_iou* (`ndarray` of shape `(num_labels,)`): Per category IoU. Examples: >>> import numpy as np >>> mean_iou = datasets.load_metric(\"mean_iou\") >>> # suppose one has 3 different segmentation maps predicted >>> predicted_1 = np.array([[1, 2], [3, 4], [5, 255]]) >>> actual_1 = np.array([[0, 3], [5, 4], [6, 255]]) >>> predicted_2 = np.array([[2, 7], [9, 2], [3, 6]]) >>> actual_2 = np.array([[1, 7], [9, 2], [3, 6]]) >>> predicted_3 = np.array([[2, 2, 3], [8, 2, 4], [3, 255, 2]]) >>> actual_3 = np.array([[1, 2, 2], [8, 2, 1], [3, 255, 1]]) >>> predicted = [predicted_1, predicted_2, predicted_3] >>> ground_truth = [actual_1, actual_2, actual_3] >>> results = mean_iou.compute(predictions=predicted, references=ground_truth, num_labels=10, ignore_index=255, reduce_labels=False) >>> print(results) # doctest: +NORMALIZE_WHITESPACE {'mean_iou': 0.47750000000000004, 'mean_accuracy': 0.5916666666666666, 'overall_accuracy': 0.5263157894736842, 'per_category_iou': array([0. , 0. , 0.375, 0.4 , 0.5 , 0. , 0.5 , 1. , 1. , 1. ]), 'per_category_accuracy': array([0. , 0. , 0.75 , 0.66666667, 1. , 0. , 0.5 , 1. , 1. , 1. ])} """ lowerCAmelCase : List[Any] = """\ @software{MMSegmentation_Contributors_OpenMMLab_Semantic_Segmentation_2020, author = {{MMSegmentation Contributors}}, license = {Apache-2.0}, month = {7}, title = {{OpenMMLab Semantic Segmentation Toolbox and Benchmark}}, url = {https://github.com/open-mmlab/mmsegmentation}, year = {2020} }""" def lowercase (_A , _A , _A , _A , _A = None , _A = False , ): """simple docstring""" if label_map is not None: for old_id, new_id in label_map.items(): _lowerCAmelCase : str = new_id # turn into Numpy arrays _lowerCAmelCase : Dict = np.array(_A ) _lowerCAmelCase : Union[str, Any] = np.array(_A ) if reduce_labels: _lowerCAmelCase : int = 2_5_5 _lowerCAmelCase : Union[str, Any] = label - 1 _lowerCAmelCase : Tuple = 2_5_5 _lowerCAmelCase : Any = label != ignore_index _lowerCAmelCase : Tuple = np.not_equal(_A , _A ) _lowerCAmelCase : Any = pred_label[mask] _lowerCAmelCase : str = np.array(_A )[mask] _lowerCAmelCase : List[Any] = pred_label[pred_label == label] _lowerCAmelCase : int = np.histogram(_A , bins=_A , range=(0, num_labels - 1) )[0] _lowerCAmelCase : int = np.histogram(_A , bins=_A , range=(0, num_labels - 1) )[0] _lowerCAmelCase : Union[str, Any] = np.histogram(_A , bins=_A , range=(0, num_labels - 1) )[0] _lowerCAmelCase : Optional[int] = area_pred_label + area_label - area_intersect return area_intersect, area_union, area_pred_label, area_label def lowercase (_A , _A , _A , _A , _A = None , _A = False , ): """simple docstring""" _lowerCAmelCase : Dict = np.zeros((num_labels,) , dtype=np.floataa ) _lowerCAmelCase : List[Any] = np.zeros((num_labels,) , dtype=np.floataa ) _lowerCAmelCase : Union[str, Any] = np.zeros((num_labels,) , dtype=np.floataa ) _lowerCAmelCase : Optional[Any] = np.zeros((num_labels,) , dtype=np.floataa ) for result, gt_seg_map in zip(_A , _A ): _lowerCAmelCase : Any = intersect_and_union( _A , _A , _A , _A , _A , _A ) total_area_intersect += area_intersect total_area_union += area_union total_area_pred_label += area_pred_label total_area_label += area_label return total_area_intersect, total_area_union, total_area_pred_label, total_area_label def lowercase (_A , _A , _A , _A , _A = None , _A = None , _A = False , ): """simple docstring""" _lowerCAmelCase : str = total_intersect_and_union( _A , _A , _A , _A , _A , _A ) # compute metrics _lowerCAmelCase : str = {} _lowerCAmelCase : Dict = total_area_intersect.sum() / total_area_label.sum() _lowerCAmelCase : Tuple = total_area_intersect / total_area_union _lowerCAmelCase : int = total_area_intersect / total_area_label _lowerCAmelCase : Optional[int] = np.nanmean(_A ) _lowerCAmelCase : Any = np.nanmean(_A ) _lowerCAmelCase : Optional[int] = all_acc _lowerCAmelCase : int = iou _lowerCAmelCase : Any = acc if nan_to_num is not None: _lowerCAmelCase : str = {metric: np.nan_to_num(_A , nan=_A ) for metric, metric_value in metrics.items()} return metrics @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase__ ( datasets.Metric ): """simple docstring""" def a ( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( # 1st Seq - height dim, 2nd - width dim { 'predictions': datasets.Sequence(datasets.Sequence(datasets.Value('uint16' ) ) ), 'references': datasets.Sequence(datasets.Sequence(datasets.Value('uint16' ) ) ), } ) , reference_urls=[ 'https://github.com/open-mmlab/mmsegmentation/blob/71c201b1813267d78764f306a297ca717827c4bf/mmseg/core/evaluation/metrics.py' ] , ) def a ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ = None , snake_case__ = None , snake_case__ = False , ): '''simple docstring''' _lowerCAmelCase : str = mean_iou( results=snake_case__ , gt_seg_maps=snake_case__ , num_labels=snake_case__ , ignore_index=snake_case__ , nan_to_num=snake_case__ , label_map=snake_case__ , reduce_labels=snake_case__ , ) return iou_result
715
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Any = { """configuration_x_clip""": [ """XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XCLIPConfig""", """XCLIPTextConfig""", """XCLIPVisionConfig""", ], """processing_x_clip""": ["""XCLIPProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """XCLIPModel""", """XCLIPPreTrainedModel""", """XCLIPTextModel""", """XCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys lowerCAmelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
630
0