code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Dict = logging.get_logger(__name__) set_seed(770) _UpperCAmelCase : Tuple = { """c_attn""": """att_proj""", """c_proj""": """out_proj""", """c_fc""": """in_proj""", """transformer.""": """""", """h.""": """layers.""", """ln_1""": """layernorm_1""", """ln_2""": """layernorm_2""", """ln_f""": """layernorm_final""", """wpe""": """position_embeds_layer""", """wte""": """input_embeds_layer""", } _UpperCAmelCase : Union[str, Any] = { """text_small""": { """repo_id""": """suno/bark""", """file_name""": """text.pt""", }, """coarse_small""": { """repo_id""": """suno/bark""", """file_name""": """coarse.pt""", }, """fine_small""": { """repo_id""": """suno/bark""", """file_name""": """fine.pt""", }, """text""": { """repo_id""": """suno/bark""", """file_name""": """text_2.pt""", }, """coarse""": { """repo_id""": """suno/bark""", """file_name""": """coarse_2.pt""", }, """fine""": { """repo_id""": """suno/bark""", """file_name""": """fine_2.pt""", }, } _UpperCAmelCase : List[Any] = os.path.dirname(os.path.abspath(__file__)) _UpperCAmelCase : List[Any] = os.path.join(os.path.expanduser("""~"""), """.cache""") _UpperCAmelCase : List[str] = os.path.join(os.getenv("""XDG_CACHE_HOME""", default_cache_dir), """suno""", """bark_v0""") def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__=False ): '''simple docstring''' snake_case_ = model_type if use_small: key += "_small" return os.path.join(UpperCamelCase__ , REMOTE_MODEL_PATHS[key]['file_name'] ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) hf_hub_download(repo_id=UpperCamelCase__ , filename=UpperCamelCase__ , local_dir=UpperCamelCase__ ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=False , UpperCamelCase__="text" ): '''simple docstring''' if model_type == "text": snake_case_ = BarkSemanticModel snake_case_ = BarkSemanticConfig snake_case_ = BarkSemanticGenerationConfig elif model_type == "coarse": snake_case_ = BarkCoarseModel snake_case_ = BarkCoarseConfig snake_case_ = BarkCoarseGenerationConfig elif model_type == "fine": snake_case_ = BarkFineModel snake_case_ = BarkFineConfig snake_case_ = BarkFineGenerationConfig else: raise NotImplementedError() snake_case_ = F'''{model_type}_small''' if use_small else model_type snake_case_ = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(UpperCamelCase__ ): logger.info(F'''{model_type} model not found, downloading into `{CACHE_DIR}`.''' ) _download(model_info['repo_id'] , model_info['file_name'] ) snake_case_ = torch.load(UpperCamelCase__ , map_location=UpperCamelCase__ ) # this is a hack snake_case_ = checkpoint['model_args'] if "input_vocab_size" not in model_args: snake_case_ = model_args['vocab_size'] snake_case_ = model_args['vocab_size'] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments snake_case_ = model_args.pop('n_head' ) snake_case_ = model_args.pop('n_embd' ) snake_case_ = model_args.pop('n_layer' ) snake_case_ = ConfigClass(**checkpoint['model_args'] ) snake_case_ = ModelClass(config=UpperCamelCase__ ) snake_case_ = GenerationConfigClass() snake_case_ = model_generation_config snake_case_ = checkpoint['model'] # fixup checkpoint snake_case_ = '_orig_mod.' for k, v in list(state_dict.items() ): if k.startswith(UpperCamelCase__ ): # replace part of the key with corresponding layer name in HF implementation snake_case_ = k[len(UpperCamelCase__ ) :] for old_layer_name in new_layer_name_dict: snake_case_ = new_k.replace(UpperCamelCase__ , new_layer_name_dict[old_layer_name] ) snake_case_ = state_dict.pop(UpperCamelCase__ ) snake_case_ = set(state_dict.keys() ) - set(model.state_dict().keys() ) snake_case_ = {k for k in extra_keys if not k.endswith('.attn.bias' )} snake_case_ = set(model.state_dict().keys() ) - set(state_dict.keys() ) snake_case_ = {k for k in missing_keys if not k.endswith('.attn.bias' )} if len(UpperCamelCase__ ) != 0: raise ValueError(F'''extra keys found: {extra_keys}''' ) if len(UpperCamelCase__ ) != 0: raise ValueError(F'''missing keys: {missing_keys}''' ) model.load_state_dict(UpperCamelCase__ , strict=UpperCamelCase__ ) snake_case_ = model.num_parameters(exclude_embeddings=UpperCamelCase__ ) snake_case_ = checkpoint['best_val_loss'].item() logger.info(F'''model loaded: {round(n_params/1E6 , 1 )}M params, {round(UpperCamelCase__ , 3 )} loss''' ) model.eval() model.to(UpperCamelCase__ ) del checkpoint, state_dict return model def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__=False , UpperCamelCase__="text" ): '''simple docstring''' if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() snake_case_ = 'cpu' # do conversion on cpu snake_case_ = _get_ckpt_path(UpperCamelCase__ , use_small=UpperCamelCase__ ) snake_case_ = _load_model(UpperCamelCase__ , UpperCamelCase__ , model_type=UpperCamelCase__ , use_small=UpperCamelCase__ ) # load bark initial model snake_case_ = _bark_load_model(UpperCamelCase__ , 'cpu' , model_type=UpperCamelCase__ , use_small=UpperCamelCase__ ) if model_type == "text": snake_case_ = bark_model['model'] if model.num_parameters(exclude_embeddings=UpperCamelCase__ ) != bark_model.get_num_params(): raise ValueError('initial and new models don\'t have the same number of parameters' ) # check if same output as the bark model snake_case_ = 5 snake_case_ = 10 if model_type in ["text", "coarse"]: snake_case_ = torch.randint(256 , (batch_size, sequence_length) , dtype=torch.int ) snake_case_ = bark_model(UpperCamelCase__ )[0] snake_case_ = model(UpperCamelCase__ ) # take last logits snake_case_ = output_new_model_total.logits[:, [-1], :] else: snake_case_ = 3 snake_case_ = 8 snake_case_ = torch.randint(256 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) snake_case_ = model(UpperCamelCase__ , UpperCamelCase__ ) snake_case_ = bark_model(UpperCamelCase__ , UpperCamelCase__ ) snake_case_ = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError('initial and new outputs don\'t have the same shape' ) if (output_new_model - output_old_model).abs().max().item() > 1E-3: raise ValueError('initial and new outputs are not equal' ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) model.save_pretrained(UpperCamelCase__ ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ): '''simple docstring''' snake_case_ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) snake_case_ = BarkSemanticConfig.from_pretrained(os.path.join(UpperCamelCase__ , 'config.json' ) ) snake_case_ = BarkCoarseConfig.from_pretrained(os.path.join(UpperCamelCase__ , 'config.json' ) ) snake_case_ = BarkFineConfig.from_pretrained(os.path.join(UpperCamelCase__ , 'config.json' ) ) snake_case_ = EncodecConfig.from_pretrained('facebook/encodec_24khz' ) snake_case_ = BarkSemanticModel.from_pretrained(UpperCamelCase__ ) snake_case_ = BarkCoarseModel.from_pretrained(UpperCamelCase__ ) snake_case_ = BarkFineModel.from_pretrained(UpperCamelCase__ ) snake_case_ = EncodecModel.from_pretrained('facebook/encodec_24khz' ) snake_case_ = BarkConfig.from_sub_model_configs( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) snake_case_ = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) snake_case_ = BarkModel(UpperCamelCase__ ) snake_case_ = semantic snake_case_ = coarseAcoustic snake_case_ = fineAcoustic snake_case_ = codec snake_case_ = bark_generation_config Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) bark.save_pretrained(UpperCamelCase__ , repo_id=UpperCamelCase__ , push_to_hub=UpperCamelCase__ ) if __name__ == "__main__": _UpperCAmelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument("""model_type""", type=str, help="""text, coarse or fine.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--is_small""", action="""store_true""", help="""convert the small version instead of the large.""") _UpperCAmelCase : Optional[Any] = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
285
import gc import unittest from diffusers import FlaxStableDiffusionInpaintPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase ( unittest.TestCase ): def a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def a ( self ): snake_case_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-inpaint/init_image.png' ) snake_case_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png' ) snake_case_ = 'xvjiarui/stable-diffusion-2-inpainting' snake_case_ , snake_case_ = FlaxStableDiffusionInpaintPipeline.from_pretrained(snake_case , safety_checker=snake_case ) snake_case_ = 'Face of a yellow cat, high resolution, sitting on a park bench' snake_case_ = jax.random.PRNGKey(0 ) snake_case_ = 50 snake_case_ = jax.device_count() snake_case_ = num_samples * [prompt] snake_case_ = num_samples * [init_image] snake_case_ = num_samples * [mask_image] snake_case_ , snake_case_ , snake_case_ = pipeline.prepare_inputs(snake_case , snake_case , snake_case ) # shard inputs and rng snake_case_ = replicate(snake_case ) snake_case_ = jax.random.split(snake_case , jax.device_count() ) snake_case_ = shard(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = pipeline( snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , jit=snake_case ) snake_case_ = output.images.reshape(snake_case , 512 , 512 , 3 ) snake_case_ = images[0, 253:256, 253:256, -1] snake_case_ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) snake_case_ = jnp.array( [0.3_61_13_07, 0.37_64_97_36, 0.3_75_74_08, 0.38_21_39_53, 0.39_29_51_67, 0.3_84_16_31, 0.41_55_49_78, 0.4_13_74_75, 0.4_21_70_84] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
285
1
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) class a_ ( lowerCamelCase ): lowercase = ["""input_features""", """is_longer"""] def __init__( self , _SCREAMING_SNAKE_CASE=64 , _SCREAMING_SNAKE_CASE=48000 , _SCREAMING_SNAKE_CASE=480 , _SCREAMING_SNAKE_CASE=10 , _SCREAMING_SNAKE_CASE=1024 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE = 0 , _SCREAMING_SNAKE_CASE = 14000 , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = "fusion" , _SCREAMING_SNAKE_CASE = "repeatpad" , **_SCREAMING_SNAKE_CASE , ) -> Optional[Any]: """simple docstring""" super().__init__( feature_size=_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , padding_value=_SCREAMING_SNAKE_CASE , return_attention_mask=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) UpperCamelCase = top_db UpperCamelCase = truncation UpperCamelCase = padding UpperCamelCase = fft_window_size UpperCamelCase = (fft_window_size >> 1) + 1 UpperCamelCase = hop_length UpperCamelCase = max_length_s UpperCamelCase = max_length_s * sampling_rate UpperCamelCase = sampling_rate UpperCamelCase = frequency_min UpperCamelCase = frequency_max UpperCamelCase = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=_SCREAMING_SNAKE_CASE , min_frequency=_SCREAMING_SNAKE_CASE , max_frequency=_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , norm=_SCREAMING_SNAKE_CASE , mel_scale="""htk""" , ) UpperCamelCase = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=_SCREAMING_SNAKE_CASE , min_frequency=_SCREAMING_SNAKE_CASE , max_frequency=_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , norm="""slaney""" , mel_scale="""slaney""" , ) def A__ ( self ) -> Dict[str, Any]: """simple docstring""" UpperCamelCase = copy.deepcopy(self.__dict__ ) UpperCamelCase = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> np.ndarray: """simple docstring""" UpperCamelCase = spectrogram( _SCREAMING_SNAKE_CASE , window_function(self.fft_window_size , """hann""" ) , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=_SCREAMING_SNAKE_CASE , log_mel="""dB""" , ) return log_mel_spectrogram.T def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" UpperCamelCase = np.array_split(list(range(0 , total_frames - chunk_frames + 1 ) ) , 3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk UpperCamelCase = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk UpperCamelCase = [0] # randomly choose index for each part UpperCamelCase = np.random.choice(ranges[0] ) UpperCamelCase = np.random.choice(ranges[1] ) UpperCamelCase = np.random.choice(ranges[2] ) UpperCamelCase = mel[idx_front : idx_front + chunk_frames, :] UpperCamelCase = mel[idx_middle : idx_middle + chunk_frames, :] UpperCamelCase = mel[idx_back : idx_back + chunk_frames, :] UpperCamelCase = torch.tensor(mel[None, None, :] ) UpperCamelCase = torch.nn.functional.interpolate( _SCREAMING_SNAKE_CASE , size=[chunk_frames, 64] , mode="""bilinear""" , align_corners=_SCREAMING_SNAKE_CASE ) UpperCamelCase = mel_shrink[0][0].numpy() UpperCamelCase = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0 ) return mel_fusion def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> np.array: """simple docstring""" if waveform.shape[0] > max_length: if truncation == "rand_trunc": UpperCamelCase = True # random crop to max_length (for compatibility) -> this should be handled by self.pad UpperCamelCase = len(_SCREAMING_SNAKE_CASE ) - max_length UpperCamelCase = np.random.randint(0 , overflow + 1 ) UpperCamelCase = waveform[idx : idx + max_length] UpperCamelCase = self._np_extract_fbank_features(_SCREAMING_SNAKE_CASE , self.mel_filters_slaney )[None, :] elif truncation == "fusion": UpperCamelCase = self._np_extract_fbank_features(_SCREAMING_SNAKE_CASE , self.mel_filters ) UpperCamelCase = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed UpperCamelCase = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. UpperCamelCase = np.stack([mel, mel, mel, mel] , axis=0 ) UpperCamelCase = False else: UpperCamelCase = self._random_mel_fusion(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCamelCase = True else: raise NotImplementedError(F"data_truncating {truncation} not implemented" ) else: UpperCamelCase = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": UpperCamelCase = int(max_length / len(_SCREAMING_SNAKE_CASE ) ) UpperCamelCase = np.stack(np.tile(_SCREAMING_SNAKE_CASE , n_repeat + 1 ) )[:max_length] if padding == "repeatpad": UpperCamelCase = int(max_length / len(_SCREAMING_SNAKE_CASE ) ) UpperCamelCase = np.stack(np.tile(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) UpperCamelCase = np.pad(_SCREAMING_SNAKE_CASE , (0, max_length - waveform.shape[0]) , mode="""constant""" , constant_values=0 ) if truncation == "fusion": UpperCamelCase = self._np_extract_fbank_features(_SCREAMING_SNAKE_CASE , self.mel_filters ) UpperCamelCase = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0 ) else: UpperCamelCase = self._np_extract_fbank_features(_SCREAMING_SNAKE_CASE , self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , ) -> BatchFeature: """simple docstring""" UpperCamelCase = truncation if truncation is not None else self.truncation UpperCamelCase = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" F" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" F" was sampled with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) UpperCamelCase = isinstance(_SCREAMING_SNAKE_CASE , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}" ) UpperCamelCase = is_batched_numpy or ( isinstance(_SCREAMING_SNAKE_CASE , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCamelCase = [np.asarray(_SCREAMING_SNAKE_CASE , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(_SCREAMING_SNAKE_CASE , np.ndarray ): UpperCamelCase = np.asarray(_SCREAMING_SNAKE_CASE , dtype=np.floataa ) elif isinstance(_SCREAMING_SNAKE_CASE , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCamelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCamelCase = [np.asarray(_SCREAMING_SNAKE_CASE )] # convert to mel spectrogram, truncate and pad if needed. UpperCamelCase = [ self._get_input_mel(_SCREAMING_SNAKE_CASE , max_length if max_length else self.nb_max_samples , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for waveform in raw_speech ] UpperCamelCase = [] UpperCamelCase = [] for mel, longer in padded_inputs: input_mel.append(_SCREAMING_SNAKE_CASE ) is_longer.append(_SCREAMING_SNAKE_CASE ) if truncation == "fusion" and sum(_SCREAMING_SNAKE_CASE ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer UpperCamelCase = np.random.randint(0 , len(_SCREAMING_SNAKE_CASE ) ) UpperCamelCase = True if isinstance(input_mel[0] , _SCREAMING_SNAKE_CASE ): UpperCamelCase = [np.asarray(_SCREAMING_SNAKE_CASE , dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool UpperCamelCase = [[longer] for longer in is_longer] UpperCamelCase = {"""input_features""": input_mel, """is_longer""": is_longer} UpperCamelCase = BatchFeature(_SCREAMING_SNAKE_CASE ) if return_tensors is not None: UpperCamelCase = input_features.convert_to_tensors(_SCREAMING_SNAKE_CASE ) return input_features
356
'''simple docstring''' import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, ) from transformers.models.focalnet.modeling_focalnet import FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a_ : def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=[32, 64, 128] , _SCREAMING_SNAKE_CASE=[1, 2, 1] , _SCREAMING_SNAKE_CASE=[2, 2, 4] , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=2.0 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=0.0_2 , _SCREAMING_SNAKE_CASE=1e-5 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=10 , _SCREAMING_SNAKE_CASE=8 , _SCREAMING_SNAKE_CASE=["stage1", "stage2"] , _SCREAMING_SNAKE_CASE=[1, 2] , ) -> Optional[Any]: """simple docstring""" UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = image_size UpperCamelCase = patch_size UpperCamelCase = num_channels UpperCamelCase = embed_dim UpperCamelCase = hidden_sizes UpperCamelCase = depths UpperCamelCase = num_heads UpperCamelCase = window_size UpperCamelCase = mlp_ratio UpperCamelCase = qkv_bias UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = drop_path_rate UpperCamelCase = hidden_act UpperCamelCase = use_absolute_embeddings UpperCamelCase = patch_norm UpperCamelCase = layer_norm_eps UpperCamelCase = initializer_range UpperCamelCase = is_training UpperCamelCase = scope UpperCamelCase = use_labels UpperCamelCase = type_sequence_label_size UpperCamelCase = encoder_stride UpperCamelCase = out_features UpperCamelCase = out_indices def A__ ( self ) -> Any: """simple docstring""" UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase = None if self.use_labels: UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase = self.get_config() return config, pixel_values, labels def A__ ( self ) -> str: """simple docstring""" return FocalNetConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , hidden_sizes=self.hidden_sizes , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" UpperCamelCase = FocalNetModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) UpperCamelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) UpperCamelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" UpperCamelCase = FocalNetBackbone(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size, 8, 8] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[:-1] ) # verify backbone works with out_features=None UpperCamelCase = None UpperCamelCase = FocalNetBackbone(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size * 2, 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" UpperCamelCase = FocalNetForMaskedImageModeling(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCamelCase = 1 UpperCamelCase = FocalNetForMaskedImageModeling(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[str]: """simple docstring""" UpperCamelCase = self.type_sequence_label_size UpperCamelCase = FocalNetForImageClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCamelCase = 1 UpperCamelCase = FocalNetForImageClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def A__ ( self ) -> List[str]: """simple docstring""" UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase ,UpperCamelCase ,UpperCamelCase = config_and_inputs UpperCamelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class a_ ( lowerCamelCase , lowerCamelCase , unittest.TestCase ): lowercase = ( ( FocalNetModel, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetBackbone, ) if is_torch_available() else () ) lowercase = ( {"""feature-extraction""": FocalNetModel, """image-classification""": FocalNetForImageClassification} if is_torch_available() else {} ) lowercase = False lowercase = False lowercase = False lowercase = False lowercase = False def A__ ( self ) -> int: """simple docstring""" UpperCamelCase = FocalNetModelTester(self ) UpperCamelCase = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , embed_dim=37 , has_text_modality=_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> List[Any]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A__ ( self ) -> Tuple: """simple docstring""" return def A__ ( self ) -> str: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> List[str]: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @unittest.skip(reason="""FocalNet does not use inputs_embeds""" ) def A__ ( self ) -> int: """simple docstring""" pass @unittest.skip(reason="""FocalNet does not use feedforward chunking""" ) def A__ ( self ) -> int: """simple docstring""" pass def A__ ( self ) -> str: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_SCREAMING_SNAKE_CASE , nn.Linear ) ) def A__ ( self ) -> List[Any]: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) UpperCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase = [*signature.parameters.keys()] UpperCamelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" UpperCamelCase = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): UpperCamelCase = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) UpperCamelCase = outputs.hidden_states UpperCamelCase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) # FocalNet has a different seq_length UpperCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCamelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) UpperCamelCase = outputs.reshaped_hidden_states self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) UpperCamelCase ,UpperCamelCase ,UpperCamelCase ,UpperCamelCase = reshaped_hidden_states[0].shape UpperCamelCase = ( reshaped_hidden_states[0].view(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def A__ ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes[:-1]: UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def A__ ( self ) -> List[Any]: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = 3 UpperCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) UpperCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCamelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) UpperCamelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes[:-1]: UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True self.check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) @slow def A__ ( self ) -> Union[str, Any]: """simple docstring""" for model_name in FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase = FocalNetModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> Any: """simple docstring""" UpperCamelCase ,UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = _config_zero_init(_SCREAMING_SNAKE_CASE ) for model_class in self.all_model_classes: UpperCamelCase = model_class(config=_SCREAMING_SNAKE_CASE ) for name, param in model.named_parameters(): if "embeddings" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"Parameter {name} of model {model_class} seems not properly initialized" , ) @require_vision @require_torch class a_ ( unittest.TestCase ): @cached_property def A__ ( self ) -> List[str]: """simple docstring""" return AutoImageProcessor.from_pretrained("""microsoft/focalnet-tiny""" ) if is_vision_available() else None @slow def A__ ( self ) -> str: """simple docstring""" UpperCamelCase = FocalNetForImageClassification.from_pretrained("""microsoft/focalnet-tiny""" ).to(_SCREAMING_SNAKE_CASE ) UpperCamelCase = self.default_image_processor UpperCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) UpperCamelCase = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).to(_SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): UpperCamelCase = model(**_SCREAMING_SNAKE_CASE ) # verify the logits UpperCamelCase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) UpperCamelCase = torch.tensor([0.2_1_6_6, -0.4_3_6_8, 0.2_1_9_1] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) ) self.assertTrue(outputs.logits.argmax(dim=-1 ).item() , 281 ) @require_torch class a_ ( lowerCamelCase , unittest.TestCase ): lowercase = (FocalNetBackbone,) if is_torch_available() else () lowercase = FocalNetConfig lowercase = False def A__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = FocalNetModelTester(self )
183
0
"""simple docstring""" def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ): '''simple docstring''' __lowerCAmelCase = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("All input parameters must be positive" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("Relative densities cannot be greater than one" ) else: __lowerCAmelCase = 1 - (matter_density + radiation_density + dark_energy) __lowerCAmelCase = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) __lowerCAmelCase = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation A : Tuple = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
57
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase ={"configuration_vit_msn": ["VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP", "ViTMSNConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase =[ "VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST", "ViTMSNModel", "ViTMSNForImageClassification", "ViTMSNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys __UpperCAmelCase =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
67
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __a = {"configuration_xglm": ["XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XGLMConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["XGLMTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["XGLMTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XGLMForCausalLM", "XGLMModel", "XGLMPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "FlaxXGLMForCausalLM", "FlaxXGLMModel", "FlaxXGLMPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXGLMForCausalLM", "TFXGLMModel", "TFXGLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure)
43
'''simple docstring''' import numpy as np def __snake_case( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = 1e-12 , _lowerCAmelCase = 100 , ) -> tuple[float, np.ndarray]: assert np.shape(_lowerCAmelCase )[0] == np.shape(_lowerCAmelCase )[1] # Ensure proper dimensionality. assert np.shape(_lowerCAmelCase )[0] == np.shape(_lowerCAmelCase )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(_lowerCAmelCase ) == np.iscomplexobj(_lowerCAmelCase ) snake_case__ : str = np.iscomplexobj(_lowerCAmelCase ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(_lowerCAmelCase , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. snake_case__ : Tuple = False snake_case__ : Any = 0 snake_case__ : List[str] = 0 snake_case__ : Dict = 1e12 while not convergence: # Multiple matrix by the vector. snake_case__ : Optional[int] = np.dot(_lowerCAmelCase , _lowerCAmelCase ) # Normalize the resulting output vector. snake_case__ : Optional[Any] = w / np.linalg.norm(_lowerCAmelCase ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) snake_case__ : List[str] = vector.conj().T if is_complex else vector.T snake_case__ : Optional[int] = np.dot(_lowerCAmelCase , np.dot(_lowerCAmelCase , _lowerCAmelCase ) ) # Check convergence. snake_case__ : Union[str, Any] = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: snake_case__ : Dict = True snake_case__ : Union[str, Any] = lambda_ if is_complex: snake_case__ : int = np.real(lambda_ ) return lambda_, vector def __snake_case( ) -> None: snake_case__ : int = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) snake_case__ : Tuple = np.array([41, 4, 20] ) snake_case__ : Dict = real_input_matrix.astype(np.complexaaa ) snake_case__ : Optional[int] = np.triu(1J * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T snake_case__ : Dict = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": snake_case__ : Dict = real_input_matrix snake_case__ : Optional[Any] = real_vector elif problem_type == "complex": snake_case__ : Optional[Any] = complex_input_matrix snake_case__ : Optional[Any] = complex_vector # Our implementation. snake_case__ , snake_case__ : Tuple = power_iteration(_lowerCAmelCase , _lowerCAmelCase ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). snake_case__ , snake_case__ : Dict = np.linalg.eigh(_lowerCAmelCase ) # Last eigenvalue is the maximum one. snake_case__ : Optional[int] = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. snake_case__ : Any = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(_lowerCAmelCase ) - np.abs(_lowerCAmelCase ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
43
1
"""simple docstring""" import collections import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = '▁' UpperCamelCase_ = {'vocab_file': 'prophetnet.tokenizer'} UpperCamelCase_ = { 'vocab_file': { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/prophetnet.tokenizer' ), } } UpperCamelCase_ = { 'microsoft/xprophetnet-large-wiki100-cased': {'do_lower_case': False}, } UpperCamelCase_ = { 'microsoft/xprophetnet-large-wiki100-cased': 512, } def UpperCamelCase ( UpperCAmelCase ) ->Dict: """simple docstring""" a_ = collections.OrderedDict() with open(UpperCAmelCase , "r" , encoding="utf-8" ) as reader: a_ = reader.readlines() for index, token in enumerate(UpperCAmelCase ): a_ = token.rstrip("\n" ) a_ = index return vocab class snake_case ( SCREAMING_SNAKE_CASE_ ): a_ : Dict = VOCAB_FILES_NAMES a_ : Tuple = PRETRAINED_VOCAB_FILES_MAP a_ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ : Union[str, Any] = ["""input_ids""", """attention_mask"""] def __init__( self , __UpperCAmelCase , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="[UNK]" , __UpperCAmelCase="[PAD]" , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[MASK]" , __UpperCAmelCase = None , **__UpperCAmelCase , ) ->None: a_ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) try: import sentencepiece as spm except ImportError: logger.warning( "You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece" " pip install sentencepiece") raise a_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(__UpperCAmelCase)) a_ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # put special tokens and [unused] tokens into the vocab a_ = {"[PAD]": 0, "[CLS]": 1, "[SEP]": 2, "[UNK]": 3, "[MASK]": 4} for i in range(10): a_ = F'''[unused{i}]''' a_ = 5 + i # The first "real" token "," has position 15 in the embedding vocab and position 3 in the spm vocab a_ = 12 a_ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} for k in self.fairseq_tokens_to_ids.keys(): self.unique_no_split_tokens.append(__UpperCAmelCase) def __getstate__( self) ->List[str]: a_ = self.__dict__.copy() a_ = None return state def __setstate__( self , __UpperCAmelCase) ->Optional[int]: a_ = d try: import sentencepiece as spm except ImportError: logger.warning( "You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece" " pip install sentencepiece") raise # for backward compatibility if not hasattr(self , "sp_model_kwargs"): a_ = {} a_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False) ->List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase) if token_ids_a is None: return ([0] * len(__UpperCAmelCase)) + [1] return ([0] * len(__UpperCAmelCase)) + [1] + ([0] * len(__UpperCAmelCase)) + [1] def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None) ->List[int]: a_ = [self.sep_token_id] if token_ids_a is None: return len(token_ids_a + sep) * [0] return len(token_ids_a + sep + sep + token_ids_a + sep) * [0] @property def UpperCAmelCase__ ( self) ->str: return len(self.sp_model) + self.fairseq_offset def UpperCAmelCase__ ( self) ->int: a_ = {self.convert_ids_to_tokens(__UpperCAmelCase): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def UpperCAmelCase__ ( self , __UpperCAmelCase) ->str: return self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase) def UpperCAmelCase__ ( self , __UpperCAmelCase) ->Optional[Any]: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] a_ = self.sp_model.PieceToId(__UpperCAmelCase) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def UpperCAmelCase__ ( self , __UpperCAmelCase) ->List[Any]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset) def UpperCAmelCase__ ( self , __UpperCAmelCase) ->List[Any]: a_ = "".join(__UpperCAmelCase).replace(__UpperCAmelCase , " ").strip() return out_string def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None) ->Tuple[str]: if not os.path.isdir(__UpperCAmelCase): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''') return a_ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(__UpperCAmelCase) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , __UpperCAmelCase) elif not os.path.isfile(self.vocab_file): with open(__UpperCAmelCase , "wb") as fi: a_ = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase) return (out_vocab_file,) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None) ->List[int]: if token_ids_a is None: return token_ids_a + [self.sep_token_id] a_ = [self.sep_token_id] return token_ids_a + sep + token_ids_a + sep
243
"""simple docstring""" def UpperCamelCase ( UpperCAmelCase ) ->str: """simple docstring""" return " ".join( "".join(word[::-1] ) if len(UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('Hey wollef sroirraw'))
243
1
import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path A : Union[str, Any] = [ {'dataset': 'wikipedia', 'config_name': '20220301.de'}, {'dataset': 'wikipedia', 'config_name': '20220301.en'}, {'dataset': 'wikipedia', 'config_name': '20220301.fr'}, {'dataset': 'wikipedia', 'config_name': '20220301.frr'}, {'dataset': 'wikipedia', 'config_name': '20220301.it'}, {'dataset': 'wikipedia', 'config_name': '20220301.simple'}, {'dataset': 'snli', 'config_name': 'plain_text'}, {'dataset': 'eli5', 'config_name': 'LFQA_reddit'}, {'dataset': 'wiki40b', 'config_name': 'en'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.compressed'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.no_index'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.multiset.no_index'}, {'dataset': 'natural_questions', 'config_name': 'default'}, ] def __lowerCAmelCase ( a__=True ) -> Tuple: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=a ) ) class __A( a ): snake_case_ = None snake_case_ = None def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> List[Any]: '''simple docstring''' with TemporaryDirectory() as tmp_dir: __a = dataset_module_factory(_snake_case , cache_dir=_snake_case ) __a = import_main_class(dataset_module.module_path , dataset=_snake_case ) __a = builder_cls( cache_dir=_snake_case , config_name=_snake_case , hash=dataset_module.hash , ) __a = '''/'''.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=_snake_case ).replace(os.sep , '''/''' ), config.DATASET_INFO_FILENAME, ] ) __a = cached_path(_snake_case , cache_dir=_snake_case ) self.assertTrue(os.path.exists(_snake_case ) ) @pytest.mark.integration def __lowerCAmelCase ( a__ ) -> List[str]: __a = tmp_path_factory.mktemp('''test_hf_gcp''' ) / '''test_wikipedia_simple''' __a = dataset_module_factory('''wikipedia''' , cache_dir=a__ ) __a = import_main_class(dataset_module.module_path ) __a = builder_cls( cache_dir=a__ , config_name='''20220301.frr''' , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam __a = None builder_instance.download_and_prepare() __a = builder_instance.as_dataset() assert ds @pytest.mark.integration def __lowerCAmelCase ( a__ ) -> Optional[Any]: __a = dataset_module_factory('''wikipedia''' , cache_dir=a__ ) __a = import_main_class(dataset_module.module_path , dataset=a__ ) __a = builder_cls( cache_dir=a__ , config_name='''20220301.frr''' , hash=dataset_module.hash , ) __a = builder_instance.as_streaming_dataset() assert ds assert isinstance(a__ , a__ ) assert "train" in ds assert isinstance(ds['''train'''] , a__ ) assert next(iter(ds['''train'''] ) )
33
import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class __A( a ): snake_case_ = 0 snake_case_ = False snake_case_ = 3.0 class __A( unittest.TestCase ): def SCREAMING_SNAKE_CASE_ ( self ) -> Dict: '''simple docstring''' self.assertDictEqual(MockClass().to_kwargs() , {} ) self.assertDictEqual(MockClass(a=2 ).to_kwargs() , {'''a''': 2} ) self.assertDictEqual(MockClass(a=2 , b=_snake_case ).to_kwargs() , {'''a''': 2, '''b''': True} ) self.assertDictEqual(MockClass(a=2 , c=2.25 ).to_kwargs() , {'''a''': 2, '''c''': 2.25} ) @require_cuda def SCREAMING_SNAKE_CASE_ ( self ) -> str: '''simple docstring''' __a = GradScalerKwargs(init_scale=1_024 , growth_factor=2 ) AcceleratorState._reset_state() __a = Accelerator(mixed_precision='''fp16''' , kwargs_handlers=[scaler_handler] ) print(accelerator.use_fpaa ) __a = accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale , 1024.0 ) self.assertEqual(scaler._growth_factor , 2.0 ) # Check the other values are at the default self.assertEqual(scaler._backoff_factor , 0.5 ) self.assertEqual(scaler._growth_interval , 2_000 ) self.assertEqual(scaler._enabled , _snake_case ) @require_multi_gpu def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' __a = ['''torchrun''', F"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] execute_subprocess_async(_snake_case , env=os.environ.copy() ) if __name__ == "__main__": A : List[str] = DistributedDataParallelKwargs(bucket_cap_mb=1_5, find_unused_parameters=True) A : Optional[Any] = Accelerator(kwargs_handlers=[ddp_scaler]) A : int = torch.nn.Linear(1_0_0, 2_0_0) A : Optional[int] = accelerator.prepare(model) # Check the values changed in kwargs A : List[Any] = '' A : Tuple = model.bucket_bytes_cap // (1_0_2_4 * 1_0_2_4) if observed_bucket_cap_map != 1_5: error_msg += F"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += F"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += F"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += F"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += F"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
33
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} lowerCAmelCase__ = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } lowerCAmelCase__ = { """google/realm-cc-news-pretrained-embedder""": 5_1_2, """google/realm-cc-news-pretrained-encoder""": 5_1_2, """google/realm-cc-news-pretrained-scorer""": 5_1_2, """google/realm-cc-news-pretrained-openqa""": 5_1_2, """google/realm-orqa-nq-openqa""": 5_1_2, """google/realm-orqa-nq-reader""": 5_1_2, """google/realm-orqa-wq-openqa""": 5_1_2, """google/realm-orqa-wq-reader""": 5_1_2, } lowerCAmelCase__ = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class a__ ( snake_case ): """simple docstring""" __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = RealmTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=True , lowercase="[UNK]" , lowercase="[SEP]" , lowercase="[PAD]" , lowercase="[CLS]" , lowercase="[MASK]" , lowercase=True , lowercase=None , **lowercase , ) -> List[str]: '''simple docstring''' super().__init__( lowercase , tokenizer_file=lowercase , do_lower_case=lowercase , unk_token=lowercase , sep_token=lowercase , pad_token=lowercase , cls_token=lowercase , mask_token=lowercase , tokenize_chinese_chars=lowercase , strip_accents=lowercase , **lowercase , ) A__ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowercase ) != do_lower_case or normalizer_state.get("strip_accents" , lowercase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowercase ) != tokenize_chinese_chars ): A__ = getattr(lowercase , normalizer_state.pop("type" ) ) A__ = do_lower_case A__ = strip_accents A__ = tokenize_chinese_chars A__ = normalizer_class(**lowercase ) A__ = do_lower_case def UpperCamelCase ( self , lowercase , **lowercase ) -> Optional[int]: '''simple docstring''' A__ = PaddingStrategy.MAX_LENGTH A__ = text A__ = kwargs.pop("text_pair" , lowercase ) A__ = kwargs.pop("return_tensors" , lowercase ) A__ = { "input_ids": [], "attention_mask": [], "token_type_ids": [], } for idx, candidate_text in enumerate(lowercase ): if batch_text_pair is not None: A__ = batch_text_pair[idx] else: A__ = None A__ = super().__call__(lowercase , lowercase , return_tensors=lowercase , **lowercase ) A__ = encoded_candidates.get("input_ids" ) A__ = encoded_candidates.get("attention_mask" ) A__ = encoded_candidates.get("token_type_ids" ) if encoded_input_ids is not None: output_data["input_ids"].append(lowercase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(lowercase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(lowercase ) A__ = {key: item for key, item in output_data.items() if len(lowercase ) != 0} return BatchEncoding(lowercase , tensor_type=lowercase ) def UpperCamelCase ( self , lowercase , lowercase=None ) -> str: '''simple docstring''' A__ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase ( self , lowercase , lowercase = None ) -> List[int]: '''simple docstring''' A__ = [self.sep_token_id] A__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase ( self , lowercase , lowercase = None ) -> Tuple[str]: '''simple docstring''' A__ = self._tokenizer.model.save(lowercase , name=lowercase ) return tuple(lowercase )
68
from __future__ import annotations def lowerCAmelCase__ ( a__: dict , a__: str ) -> set[str]: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase = set(a__ ), [start] while stack: _UpperCAmelCase = stack.pop() explored.add(a__ ) # Differences from BFS: # 1) pop last element instead of first one # 2) add adjacent elements to stack without exploring them for adj in reversed(graph[v] ): if adj not in explored: stack.append(a__ ) return explored lowerCAmelCase__ :Tuple = { '''A''': ['''B''', '''C''', '''D'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F'''], '''D''': ['''B''', '''D'''], '''E''': ['''B''', '''F'''], '''F''': ['''C''', '''E''', '''G'''], '''G''': ['''F'''], } if __name__ == "__main__": import doctest doctest.testmod() print(depth_first_search(G, '''A'''))
329
0
from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse("3.8"): import importlib_metadata else: import importlib.metadata as importlib_metadata __UpperCAmelCase :Union[str, Any] = "" if version.parse(importlib_metadata.version("jiwer")) < version.parse("2.3.0"): class a ( tr.AbstractTransform ): """simple docstring""" def __init__( self : Tuple , snake_case : str = " " ) -> Optional[int]: __UpperCAmelCase : Optional[Any] = sentence_delimiter def lowerCamelCase__ ( self : int , snake_case : str ) -> Optional[int]: return list(snake_case ) def lowerCamelCase__ ( self : Any , snake_case : List[str] ) -> Optional[Any]: __UpperCAmelCase : List[Any] = [] for sent_idx, sentence in enumerate(snake_case ): chars.extend(self.process_string(snake_case ) ) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(snake_case ) - 1: chars.append(self.sentence_delimiter ) return chars __UpperCAmelCase :int = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: __UpperCAmelCase :int = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) __UpperCAmelCase :List[Any] = "\\n@inproceedings{inproceedings,\n author = {Morris, Andrew and Maier, Viktoria and Green, Phil},\n year = {2004},\n month = {01},\n pages = {},\n title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.}\n}\n" __UpperCAmelCase :Union[str, Any] = "\\nCharacter error rate (CER) is a common metric of the performance of an automatic speech recognition system.\n\nCER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information.\n\nCharacter error rate can be computed as:\n\nCER = (S + D + I) / N = (S + D + I) / (S + D + C)\n\nwhere\n\nS is the number of substitutions,\nD is the number of deletions,\nI is the number of insertions,\nC is the number of correct characters,\nN is the number of characters in the reference (N=S+D+C).\n\nCER's output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the\nperformance of the ASR system with a CER of 0 being a perfect score.\n" __UpperCAmelCase :Union[str, Any] = "\nComputes CER score of transcribed segments against references.\nArgs:\n references: list of references for each speech input.\n predictions: list of transcribtions to score.\n concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result.\nReturns:\n (float): the character error rate\n\nExamples:\n\n >>> predictions = [\"this is the prediction\", \"there is an other sample\"]\n >>> references = [\"this is the reference\", \"there is another one\"]\n >>> cer = datasets.load_metric(\"cer\")\n >>> cer_score = cer.compute(predictions=predictions, references=references)\n >>> print(cer_score)\n 0.34146341463414637\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): """simple docstring""" def lowerCamelCase__ ( self : Tuple ) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', '''https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates''', ] , ) def lowerCamelCase__ ( self : Optional[Any] , snake_case : List[str] , snake_case : Dict , snake_case : str=False ) -> str: if concatenate_texts: return jiwer.compute_measures( snake_case , snake_case , truth_transform=snake_case , hypothesis_transform=snake_case , )["wer"] __UpperCAmelCase : Optional[int] = 0 __UpperCAmelCase : Tuple = 0 for prediction, reference in zip(snake_case , snake_case ): __UpperCAmelCase : str = jiwer.compute_measures( snake_case , snake_case , truth_transform=snake_case , hypothesis_transform=snake_case , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
353
'''simple docstring''' from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class a : """simple docstring""" SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] SCREAMING_SNAKE_CASE : int SCREAMING_SNAKE_CASE : int SCREAMING_SNAKE_CASE : float SCREAMING_SNAKE_CASE : float SCREAMING_SNAKE_CASE : Tuple[int] def lowerCamelCase__ ( self : Any ) -> int: assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def lowerCamelCase__ ( self : Union[str, Any] ) -> str: return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def lowerCamelCase__ ( self : Any ) -> Optional[Any]: return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def lowerCamelCase__ ( self : Any ) -> torch.Tensor: __UpperCAmelCase : Dict = torch.arange(self.height * self.width ) __UpperCAmelCase : Dict = torch.stack( [ pixel_indices % self.width, torch.div(snake_case , self.width , rounding_mode='''trunc''' ), ] , axis=1 , ) return coords @property def lowerCamelCase__ ( self : Any ) -> int: __UpperCAmelCase , *__UpperCAmelCase : str = self.shape __UpperCAmelCase : Dict = int(np.prod(snake_case ) ) __UpperCAmelCase : Tuple = self.get_image_coords() __UpperCAmelCase : List[Any] = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __UpperCAmelCase : Any = self.get_camera_rays(snake_case ) __UpperCAmelCase : List[str] = rays.view(snake_case , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def lowerCamelCase__ ( self : Union[str, Any] , snake_case : torch.Tensor ) -> torch.Tensor: __UpperCAmelCase , *__UpperCAmelCase , __UpperCAmelCase : List[str] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __UpperCAmelCase : List[str] = coords.view(snake_case , -1 , 2 ) __UpperCAmelCase : Optional[Any] = self.resolution() __UpperCAmelCase : Tuple = self.fov() __UpperCAmelCase : Optional[int] = (flat.float() / (res - 1)) * 2 - 1 __UpperCAmelCase : Union[str, Any] = fracs * torch.tan(fov / 2 ) __UpperCAmelCase : str = fracs.view(snake_case , -1 , 2 ) __UpperCAmelCase : Any = ( self.z.view(snake_case , 1 , 3 ) + self.x.view(snake_case , 1 , 3 ) * fracs[:, :, :1] + self.y.view(snake_case , 1 , 3 ) * fracs[:, :, 1:] ) __UpperCAmelCase : Union[str, Any] = directions / directions.norm(dim=-1 , keepdim=snake_case ) __UpperCAmelCase : Union[str, Any] = torch.stack( [ torch.broadcast_to(self.origin.view(snake_case , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(snake_case , *snake_case , 2 , 3 ) def lowerCamelCase__ ( self : Any , snake_case : int , snake_case : int ) -> "DifferentiableProjectiveCamera": assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=snake_case , height=snake_case , x_fov=self.x_fov , y_fov=self.y_fov , ) def _a ( _lowercase : int ): '''simple docstring''' __UpperCAmelCase : str = [] __UpperCAmelCase : Optional[int] = [] __UpperCAmelCase : Union[str, Any] = [] __UpperCAmelCase : List[Any] = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): __UpperCAmelCase : Dict = np.array([np.sin(_lowercase ), np.cos(_lowercase ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __UpperCAmelCase : Any = -z * 4 __UpperCAmelCase : Dict = np.array([np.cos(_lowercase ), -np.sin(_lowercase ), 0.0] ) __UpperCAmelCase : List[str] = np.cross(_lowercase , _lowercase ) origins.append(_lowercase ) xs.append(_lowercase ) ys.append(_lowercase ) zs.append(_lowercase ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , x=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , y=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , z=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , width=_lowercase , height=_lowercase , x_fov=0.7 , y_fov=0.7 , shape=(1, len(_lowercase )) , )
240
0
from typing import Dict, List, Optional from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCamelCase = logging.get_logger(__name__) __lowerCamelCase = { """nielsr/canine-s""": 20_48, } # Unicode defines 1,114,112 total “codepoints” __lowerCamelCase = 1_11_41_12 # Below: Constants defining canonical codepoints for special, pseudo-characters. # Copied from https://github.com/google-research/language/blob/master/language/canine/special_codepoints.py __lowerCamelCase = 0 __lowerCamelCase = 0xe_000 __lowerCamelCase = 0xe_001 __lowerCamelCase = 0xe_002 __lowerCamelCase = 0xe_003 __lowerCamelCase = 0xe_004 # Maps special codepoints to human-readable names. __lowerCamelCase = { # Special symbols are represented using codepoints values that are valid, # but designated as "Private Use", meaning that they will never be assigned # characters by the Unicode Consortium, and are thus safe for use here. # # NOTE: Do *NOT* add any sort of [UNK_CHAR] here. They are explicitly # excluded and should fail with a hard error. CLS: "[CLS]", SEP: "[SEP]", BOS: "[BOS]", MASK: "[MASK]", PAD: "[PAD]", RESERVED: "[RESERVED]", } # Maps special codepoint human-readable names to their codepoint values. __lowerCamelCase = {name: codepoint for codepoint, name in SPECIAL_CODEPOINTS.items()} class UpperCAmelCase ( A_ ): A__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__(self : List[Any] , snake_case__ : Optional[Any]=chr(snake_case__ ) , snake_case__ : Optional[int]=chr(snake_case__ ) , snake_case__ : Optional[int]=chr(snake_case__ ) , snake_case__ : Dict=chr(snake_case__ ) , snake_case__ : List[Any]=chr(snake_case__ ) , snake_case__ : Tuple=chr(snake_case__ ) , snake_case__ : Optional[int]=False , snake_case__ : Optional[int]=20_48 , **snake_case__ : str , ) -> Tuple: '''simple docstring''' snake_case : Dict = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else bos_token snake_case : List[Any] = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else eos_token snake_case : Any = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else sep_token snake_case : Union[str, Any] = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else cls_token snake_case : str = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it snake_case : Union[str, Any] = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token super().__init__( bos_token=snake_case__ , eos_token=snake_case__ , sep_token=snake_case__ , cls_token=snake_case__ , pad_token=snake_case__ , mask_token=snake_case__ , add_prefix_space=snake_case__ , model_max_length=snake_case__ , **snake_case__ , ) # Creates a mapping for looking up the IDs of special symbols. snake_case : Dict[str, int] = {} for codepoint, name in SPECIAL_CODEPOINTS.items(): snake_case : List[str] = codepoint # Creates a mapping for looking up the string forms of special symbol IDs. snake_case : Dict[int, str] = { codepoint: name for name, codepoint in self._special_codepoints.items() } snake_case : Dict = UNICODE_VOCAB_SIZE snake_case : Dict = len(self._special_codepoints ) @property def _SCREAMING_SNAKE_CASE (self : Optional[Any] ) -> int: '''simple docstring''' return self._unicode_vocab_size def _SCREAMING_SNAKE_CASE (self : str , snake_case__ : str ) -> List[str]: '''simple docstring''' return list(snake_case__ ) def _SCREAMING_SNAKE_CASE (self : List[Any] , snake_case__ : str ) -> int: '''simple docstring''' try: return ord(snake_case__ ) except TypeError: raise ValueError(f"""invalid token: '{token}'""" ) def _SCREAMING_SNAKE_CASE (self : List[str] , snake_case__ : int ) -> str: '''simple docstring''' try: if index in SPECIAL_CODEPOINTS: return SPECIAL_CODEPOINTS[index] return chr(snake_case__ ) except TypeError: raise ValueError(f"""invalid id: {index}""" ) def _SCREAMING_SNAKE_CASE (self : Union[str, Any] , snake_case__ : Tuple ) -> str: '''simple docstring''' return "".join(snake_case__ ) def _SCREAMING_SNAKE_CASE (self : str , snake_case__ : List[int] , snake_case__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' snake_case : Any = [self.sep_token_id] snake_case : List[Any] = [self.cls_token_id] snake_case : str = cls + token_ids_a + sep if token_ids_a is not None: result += token_ids_a + sep return result def _SCREAMING_SNAKE_CASE (self : List[str] , snake_case__ : List[int] , snake_case__ : Optional[List[int]] = None , snake_case__ : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case__ , token_ids_a=snake_case__ , already_has_special_tokens=snake_case__ ) snake_case : Tuple = [1] + ([0] * len(snake_case__ )) + [1] if token_ids_a is not None: result += ([0] * len(snake_case__ )) + [1] return result def _SCREAMING_SNAKE_CASE (self : Optional[int] , snake_case__ : List[int] , snake_case__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' snake_case : str = [self.sep_token_id] snake_case : Optional[int] = [self.cls_token_id] snake_case : int = len(cls + token_ids_a + sep ) * [0] if token_ids_a is not None: result += len(token_ids_a + sep ) * [1] return result def _SCREAMING_SNAKE_CASE (self : Optional[Any] , snake_case__ : str , snake_case__ : Optional[str] = None ) -> List[Any]: '''simple docstring''' return ()
59
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def UpperCamelCase ( __lowerCamelCase : Optional[int] ): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings" , set() ) @pytest.fixture def UpperCamelCase ( __lowerCamelCase : str ): class UpperCAmelCase : def __init__(self : Optional[int] , snake_case__ : str ) -> Any: '''simple docstring''' snake_case : List[str] = metric_id class UpperCAmelCase : A__ : List[str] = [MetricMock(A_ ) for metric_id in ["accuracy", "mse", "precision", "codeparrot/apps_metric"]] def _SCREAMING_SNAKE_CASE (self : int ) -> List[str]: '''simple docstring''' return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub" , HfhMock() ) @pytest.mark.parametrize( "func, args" , [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def UpperCamelCase ( __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : Any , __lowerCamelCase : int , __lowerCamelCase : Any ): if "tmp_path" in args: snake_case : str = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(__lowerCamelCase , match="https://huggingface.co/docs/evaluate" ): func(*__lowerCamelCase )
59
1
"""simple docstring""" import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCamelCase ( _lowercase , unittest.TestCase ): UpperCAmelCase_ = LongformerTokenizer UpperCAmelCase_ = True UpperCAmelCase_ = LongformerTokenizerFast UpperCAmelCase_ = True def snake_case_ (self ) -> Any: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCamelCase = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] UpperCamelCase = dict(zip(__a , range(len(__a ) ) ) ) UpperCamelCase = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] UpperCamelCase = {"unk_token": "<unk>"} UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(__a ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(__a ) ) def snake_case_ (self , **__a ) -> List[Any]: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **__a ) def snake_case_ (self , **__a ) -> Any: kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **__a ) def snake_case_ (self , __a ) -> Any: UpperCamelCase = "lower newer" UpperCamelCase = "lower newer" return input_text, output_text def snake_case_ (self ) -> str: UpperCamelCase = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCamelCase = "lower newer" UpperCamelCase = ["l", "o", "w", "er", "\u0120", "n", "e", "w", "er"] UpperCamelCase = tokenizer.tokenize(__a ) # , add_prefix_space=True) self.assertListEqual(__a , __a ) UpperCamelCase = tokens + [tokenizer.unk_token] UpperCamelCase = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(__a ) , __a ) def snake_case_ (self ) -> Optional[Any]: UpperCamelCase = self.get_tokenizer() self.assertListEqual(tokenizer.encode("Hello world!" , add_special_tokens=__a ) , [0, 3_14_14, 2_32, 3_28, 2] ) self.assertListEqual( tokenizer.encode("Hello world! cécé herlolip 418" , add_special_tokens=__a ) , [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2] , ) @slow def snake_case_ (self ) -> Optional[int]: UpperCamelCase = self.tokenizer_class.from_pretrained("allenai/longformer-base-4096" ) UpperCamelCase = tokenizer.encode("sequence builders" , add_special_tokens=__a ) UpperCamelCase = tokenizer.encode("multi-sequence build" , add_special_tokens=__a ) UpperCamelCase = tokenizer.encode( "sequence builders" , add_special_tokens=__a , add_prefix_space=__a ) UpperCamelCase = tokenizer.encode( "sequence builders" , "multi-sequence build" , add_special_tokens=__a , add_prefix_space=__a ) UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__a ) UpperCamelCase = tokenizer.build_inputs_with_special_tokens(__a , __a ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def snake_case_ (self ) -> Tuple: UpperCamelCase = self.get_tokenizer() UpperCamelCase = "Encode this sequence." UpperCamelCase = tokenizer.byte_encoder[" ".encode("utf-8" )[0]] # Testing encoder arguments UpperCamelCase = tokenizer.encode(__a , add_special_tokens=__a , add_prefix_space=__a ) UpperCamelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(__a , __a ) UpperCamelCase = tokenizer.encode(__a , add_special_tokens=__a , add_prefix_space=__a ) UpperCamelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(__a , __a ) tokenizer.add_special_tokens({"bos_token": "<s>"} ) UpperCamelCase = tokenizer.encode(__a , add_special_tokens=__a ) UpperCamelCase = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(__a , __a ) # Testing spaces after special tokens UpperCamelCase = "<mask>" tokenizer.add_special_tokens( {"mask_token": AddedToken(__a , lstrip=__a , rstrip=__a )} ) # mask token has a left space UpperCamelCase = tokenizer.convert_tokens_to_ids(__a ) UpperCamelCase = "Encode <mask> sequence" UpperCamelCase = "Encode <mask>sequence" UpperCamelCase = tokenizer.encode(__a ) UpperCamelCase = encoded.index(__a ) UpperCamelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(__a , __a ) UpperCamelCase = tokenizer.encode(__a ) UpperCamelCase = encoded.index(__a ) UpperCamelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(__a , __a ) def snake_case_ (self ) -> Any: pass def snake_case_ (self ) -> Optional[int]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCamelCase = self.rust_tokenizer_class.from_pretrained(__a , **__a ) UpperCamelCase = self.tokenizer_class.from_pretrained(__a , **__a ) UpperCamelCase = "A, <mask> AllenNLP sentence." UpperCamelCase = tokenizer_r.encode_plus(__a , add_special_tokens=__a , return_token_type_ids=__a ) UpperCamelCase = tokenizer_p.encode_plus(__a , add_special_tokens=__a , return_token_type_ids=__a ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["token_type_ids"] ) , sum(tokens_p["token_type_ids"] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) , sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) , ) UpperCamelCase = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) UpperCamelCase = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["input_ids"] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( __a , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( __a , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) def snake_case_ (self ) -> Optional[int]: for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): UpperCamelCase = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=__a , add_prefix_space=__a , trim_offsets=__a ) UpperCamelCase = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) UpperCamelCase = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["add_prefix_space"] , __a ) self.assertEqual(post_processor_state["add_prefix_space"] , __a ) self.assertEqual(post_processor_state["trim_offsets"] , __a ) def snake_case_ (self ) -> Any: # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` and # `trim_offsets` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCamelCase = "hello" # `hello` is a token in the vocabulary of `pretrained_name` UpperCamelCase = F"{text_of_1_token} {text_of_1_token}" UpperCamelCase = self.rust_tokenizer_class.from_pretrained( __a , use_fast=__a , add_prefix_space=__a , trim_offsets=__a ) UpperCamelCase = tokenizer_r(__a , return_offsets_mapping=__a , add_special_tokens=__a ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__a )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__a ) + 1, len(__a ) + 1 + len(__a )) , ) UpperCamelCase = self.rust_tokenizer_class.from_pretrained( __a , use_fast=__a , add_prefix_space=__a , trim_offsets=__a ) UpperCamelCase = tokenizer_r(__a , return_offsets_mapping=__a , add_special_tokens=__a ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__a )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__a ) + 1, len(__a ) + 1 + len(__a )) , ) UpperCamelCase = self.rust_tokenizer_class.from_pretrained( __a , use_fast=__a , add_prefix_space=__a , trim_offsets=__a ) UpperCamelCase = tokenizer_r(__a , return_offsets_mapping=__a , add_special_tokens=__a ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__a )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__a ), len(__a ) + 1 + len(__a )) , ) UpperCamelCase = self.rust_tokenizer_class.from_pretrained( __a , use_fast=__a , add_prefix_space=__a , trim_offsets=__a ) UpperCamelCase = tokenizer_r(__a , return_offsets_mapping=__a , add_special_tokens=__a ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__a )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__a ), len(__a ) + 1 + len(__a )) , ) UpperCamelCase = F" {text}" # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) UpperCamelCase = self.rust_tokenizer_class.from_pretrained( __a , use_fast=__a , add_prefix_space=__a , trim_offsets=__a ) UpperCamelCase = tokenizer_r(__a , return_offsets_mapping=__a , add_special_tokens=__a ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(__a )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__a ) + 1, 1 + len(__a ) + 1 + len(__a )) , ) UpperCamelCase = self.rust_tokenizer_class.from_pretrained( __a , use_fast=__a , add_prefix_space=__a , trim_offsets=__a ) UpperCamelCase = tokenizer_r(__a , return_offsets_mapping=__a , add_special_tokens=__a ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(__a )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__a ), 1 + len(__a ) + 1 + len(__a )) , ) UpperCamelCase = self.rust_tokenizer_class.from_pretrained( __a , use_fast=__a , add_prefix_space=__a , trim_offsets=__a ) UpperCamelCase = tokenizer_r(__a , return_offsets_mapping=__a , add_special_tokens=__a ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(__a )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__a ), 1 + len(__a ) + 1 + len(__a )) , )
244
"""simple docstring""" import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict lowerCAmelCase__ = namedtuple( '''_TestCommandArgs''', [ '''dataset''', '''name''', '''cache_dir''', '''data_dir''', '''all_configs''', '''save_infos''', '''ignore_verifications''', '''force_redownload''', '''clear_cache''', ], defaults=[None, None, None, False, False, False, False, False], ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = _TestCommandArgs(dataset=_SCREAMING_SNAKE_CASE , all_configs=_SCREAMING_SNAKE_CASE , save_infos=_SCREAMING_SNAKE_CASE ) UpperCamelCase = TestCommand(*_SCREAMING_SNAKE_CASE ) test_command.run() UpperCamelCase = os.path.join(_SCREAMING_SNAKE_CASE , "README.md" ) assert os.path.exists(_SCREAMING_SNAKE_CASE ) UpperCamelCase = DatasetInfosDict.from_directory(_SCREAMING_SNAKE_CASE ) UpperCamelCase = DatasetInfosDict( { "default": DatasetInfo( features=Features( { "tokens": Sequence(Value("string" ) ), "ner_tags": Sequence( ClassLabel(names=["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] ) ), "langs": Sequence(Value("string" ) ), "spans": Sequence(Value("string" ) ), } ) , splits=[ { "name": "train", "num_bytes": 2_351_563, "num_examples": 10_000, }, { "name": "validation", "num_bytes": 238_418, "num_examples": 1_000, }, ] , download_size=3_940_680 , dataset_size=2_589_981 , ) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: UpperCamelCase , UpperCamelCase = getattr(dataset_infos["default"] , _SCREAMING_SNAKE_CASE ), getattr(expected_dataset_infos["default"] , _SCREAMING_SNAKE_CASE ) if key == "num_bytes": assert is_apercent_close(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif key == "splits": assert list(_SCREAMING_SNAKE_CASE ) == list(_SCREAMING_SNAKE_CASE ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes , expected[split].num_bytes ) else: result == expected
244
1
"""simple docstring""" import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __A : Optional[Any] = get_tests_dir("fixtures/spiece.model") @require_sentencepiece @require_tokenizers class _a ( __lowercase , unittest.TestCase): """simple docstring""" UpperCamelCase__ = AlbertTokenizer UpperCamelCase__ = AlbertTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = True def lowercase__ ( self : int )->int: super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase__ ( self : Union[str, Any] , __UpperCamelCase : List[str] )->Union[str, Any]: _UpperCAmelCase = '''this is a test''' _UpperCAmelCase = '''this is a test''' return input_text, output_text def lowercase__ ( self : List[str] )->List[Any]: _UpperCAmelCase = '''<pad>''' _UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase__ ) , UpperCAmelCase__ ) def lowercase__ ( self : Tuple )->int: _UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<pad>''' ) self.assertEqual(vocab_keys[1] , '''<unk>''' ) self.assertEqual(vocab_keys[-1] , '''▁eloquent''' ) self.assertEqual(len(UpperCAmelCase__ ) , 3_0_0_0_0 ) def lowercase__ ( self : str )->List[str]: self.assertEqual(self.get_tokenizer().vocab_size , 3_0_0_0_0 ) def lowercase__ ( self : Tuple )->Union[str, Any]: if not self.test_rust_tokenizer: return _UpperCAmelCase = self.get_tokenizer() _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = '''I was born in 92000, and this is falsé.''' _UpperCAmelCase = tokenizer.tokenize(UpperCAmelCase__ ) _UpperCAmelCase = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) _UpperCAmelCase = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) _UpperCAmelCase = self.get_rust_tokenizer() _UpperCAmelCase = tokenizer.encode(UpperCAmelCase__ ) _UpperCAmelCase = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def lowercase__ ( self : str )->str: _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) _UpperCAmelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(UpperCAmelCase__ , ['''▁this''', '''▁is''', '''▁a''', '''▁test'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [4_8, 2_5, 2_1, 1_2_8_9] ) _UpperCAmelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( UpperCAmelCase__ , ['''▁i''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''é''', '''.'''] ) _UpperCAmelCase = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , [3_1, 2_3, 3_8_6, 1_9, 5_6_1, 3_0_5_0, 1_5, 1_7, 4_8, 2_5, 8_2_5_6, 1_8, 1, 9] ) _UpperCAmelCase = tokenizer.convert_ids_to_tokens(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , ['''▁i''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''.'''] , ) def lowercase__ ( self : str )->Tuple: _UpperCAmelCase = AlbertTokenizer(UpperCAmelCase__ ) _UpperCAmelCase = tokenizer.encode('''sequence builders''' ) _UpperCAmelCase = tokenizer.encode('''multi-sequence build''' ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ ) _UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ , UpperCAmelCase__ ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def lowercase__ ( self : Dict )->int: # fmt: off _UpperCAmelCase = {'''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''input_ids''': [[2, 2_1_9_7_0, 1_3, 5, 6_0_9_2, 1_6_7, 2_8, 7_1_0_3, 2_1_5_3, 6_7_3, 8, 7_0_2_8, 1_2_0_5_1, 1_8, 1_7, 7_1_0_3, 2_1_5_3, 6_7_3, 8, 3_5_1_5, 1_8_6_8_4, 8, 4_4_6_1, 6, 1_9_2_7, 2_9_7, 8, 1_2_0_6_0, 2_6_0_7, 1_8, 1_3, 5, 4_4_6_1, 1_5, 1_0_5_3_8, 3_8, 8, 1_3_5, 1_5, 8_2_2, 5_8, 1_5, 9_9_3, 1_0_3_6_3, 1_5, 1_4_6_0, 8_0_0_5, 4_4_6_1, 1_5, 9_9_3, 2_5_5, 2_3_2_8, 9, 9, 9, 6, 2_6, 1_1_1_2, 8_1_6, 3_2_6_0, 1_3, 5, 1_0_3, 2_3_7_7, 6, 1_7, 1_1_1_2, 8_1_6, 2_7_8_2, 1_3, 5, 1_0_3, 1_0_6_4_1, 6, 2_9, 8_4, 2_5_1_2, 2_4_3_0, 7_8_2, 1_8_6_8_4, 2_7_6_1, 1_9, 8_0_8, 2_4_3_0, 2_5_5_6, 1_7, 8_5_5, 1_4_8_0, 9_4_7_7, 4_0_9_1, 1_2_8, 1_1_7_1_2, 1_5, 7_1_0_3, 2_1_5_3, 6_7_3, 1_7, 2_4_8_8_3, 9_9_9_0, 9, 3], [2, 1_1_5_0_2, 2_5, 1_0_0_6, 2_0, 7_8_2, 8, 1_1_8_0_9, 8_5_5, 1_7_3_2, 1_9_3_9_3, 1_8_6_6_7, 3_7, 3_6_7, 2_1_0_1_8, 6_9, 1_8_5_4, 3_4, 1_1_8_6_0, 1_9_1_2_4, 2_7, 1_5_6, 2_2_5, 1_7, 1_9_3, 4_1_4_1, 1_9, 6_5, 9_1_2_4, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 1_4, 2_2_3_1, 8_8_6, 2_3_8_5, 1_7_6_5_9, 8_4, 1_4, 1_6_7_9_2, 1_9_5_2, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''token_type_ids''': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase__ , model_name='''albert-base-v2''' , revision='''6b6560eaf5ff2e250b00c50f380c5389a9c2d82e''' , )
260
'''simple docstring''' import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def a_ ( lowerCamelCase : Dict ): lowerCAmelCase = {} lowerCAmelCase = tokenizer(example['content'] , truncation=lowerCamelCase )['input_ids'] lowerCAmelCase = len(example['content'] ) / len(output['input_ids'] ) return output __snake_case =HfArgumentParser(PretokenizationArguments) __snake_case =parser.parse_args() if args.num_workers is None: __snake_case =multiprocessing.cpu_count() __snake_case =AutoTokenizer.from_pretrained(args.tokenizer_dir) __snake_case =time.time() __snake_case =load_dataset(args.dataset_name, split="""train""") print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') __snake_case =time.time() __snake_case =ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') __snake_case =time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
4
0
'''simple docstring''' import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed __snake_case = { '''distilbert''': (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), '''roberta''': (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), '''bert''': (BertConfig, BertForMaskedLM, BertTokenizer), '''gpt2''': (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def a ( __a ) -> Union[str, Any]: '''simple docstring''' assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def a ( __a , __a ) -> Any: '''simple docstring''' if args.student_type == "roberta": UpperCamelCase__ :Dict = False elif args.student_type == "gpt2": UpperCamelCase__ :List[str] = False def a ( __a , __a ) -> List[Any]: '''simple docstring''' if args.student_type == "roberta": UpperCamelCase__ :Dict = False def a ( ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ :Union[str, Any] = argparse.ArgumentParser(description='''Training''' ) parser.add_argument('''--force''' , action='''store_true''' , help='''Overwrite dump_path if it already exists.''' ) parser.add_argument( '''--dump_path''' , type=__a , required=__a , help='''The output directory (log, checkpoints, parameters, etc.)''' ) parser.add_argument( '''--data_file''' , type=__a , required=__a , help='''The binarized file (tokenized + tokens_to_ids) and grouped by sequence.''' , ) parser.add_argument( '''--student_type''' , type=__a , choices=['''distilbert''', '''roberta''', '''gpt2'''] , required=__a , help='''The student type (DistilBERT, RoBERTa).''' , ) parser.add_argument('''--student_config''' , type=__a , required=__a , help='''Path to the student configuration.''' ) parser.add_argument( '''--student_pretrained_weights''' , default=__a , type=__a , help='''Load student initialization checkpoint.''' ) parser.add_argument( '''--teacher_type''' , choices=['''bert''', '''roberta''', '''gpt2'''] , required=__a , help='''Teacher type (BERT, RoBERTa).''' ) parser.add_argument('''--teacher_name''' , type=__a , required=__a , help='''The teacher model.''' ) parser.add_argument('''--temperature''' , default=2.0 , type=__a , help='''Temperature for the softmax temperature.''' ) parser.add_argument( '''--alpha_ce''' , default=0.5 , type=__a , help='''Linear weight for the distillation loss. Must be >=0.''' ) parser.add_argument( '''--alpha_mlm''' , default=0.0 , type=__a , help='''Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.''' , ) parser.add_argument('''--alpha_clm''' , default=0.5 , type=__a , help='''Linear weight for the CLM loss. Must be >=0.''' ) parser.add_argument('''--alpha_mse''' , default=0.0 , type=__a , help='''Linear weight of the MSE loss. Must be >=0.''' ) parser.add_argument( '''--alpha_cos''' , default=0.0 , type=__a , help='''Linear weight of the cosine embedding loss. Must be >=0.''' ) parser.add_argument( '''--mlm''' , action='''store_true''' , help='''The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.''' ) parser.add_argument( '''--mlm_mask_prop''' , default=0.1_5 , type=__a , help='''Proportion of tokens for which we need to make a prediction.''' , ) parser.add_argument('''--word_mask''' , default=0.8 , type=__a , help='''Proportion of tokens to mask out.''' ) parser.add_argument('''--word_keep''' , default=0.1 , type=__a , help='''Proportion of tokens to keep.''' ) parser.add_argument('''--word_rand''' , default=0.1 , type=__a , help='''Proportion of tokens to randomly replace.''' ) parser.add_argument( '''--mlm_smoothing''' , default=0.7 , type=__a , help='''Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).''' , ) parser.add_argument('''--token_counts''' , type=__a , help='''The token counts in the data_file for MLM.''' ) parser.add_argument( '''--restrict_ce_to_mask''' , action='''store_true''' , help='''If true, compute the distillation loss only the [MLM] prediction distribution.''' , ) parser.add_argument( '''--freeze_pos_embs''' , action='''store_true''' , help='''Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only.''' , ) parser.add_argument( '''--freeze_token_type_embds''' , action='''store_true''' , help='''Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only.''' , ) parser.add_argument('''--n_epoch''' , type=__a , default=3 , help='''Number of pass on the whole dataset.''' ) parser.add_argument('''--batch_size''' , type=__a , default=5 , help='''Batch size (for each process).''' ) parser.add_argument( '''--group_by_size''' , action='''store_false''' , help='''If true, group sequences that have similar length into the same batch. Default is true.''' , ) parser.add_argument( '''--gradient_accumulation_steps''' , type=__a , default=50 , help='''Gradient accumulation for larger training batches.''' , ) parser.add_argument('''--warmup_prop''' , default=0.0_5 , type=__a , help='''Linear warmup proportion.''' ) parser.add_argument('''--weight_decay''' , default=0.0 , type=__a , help='''Weight decay if we apply some.''' ) parser.add_argument('''--learning_rate''' , default=5e-4 , type=__a , help='''The initial learning rate for Adam.''' ) parser.add_argument('''--adam_epsilon''' , default=1e-6 , type=__a , help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' , default=5.0 , type=__a , help='''Max gradient norm.''' ) parser.add_argument('''--initializer_range''' , default=0.0_2 , type=__a , help='''Random initialization range.''' ) parser.add_argument( '''--fp16''' , action='''store_true''' , help='''Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit''' , ) parser.add_argument( '''--fp16_opt_level''' , type=__a , default='''O1''' , help=( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].''' '''See details at https://nvidia.github.io/apex/amp.html''' ) , ) parser.add_argument('''--n_gpu''' , type=__a , default=1 , help='''Number of GPUs in the node.''' ) parser.add_argument('''--local_rank''' , type=__a , default=-1 , help='''Distributed training - Local rank''' ) parser.add_argument('''--seed''' , type=__a , default=56 , help='''Random seed''' ) parser.add_argument('''--log_interval''' , type=__a , default=500 , help='''Tensorboard logging interval.''' ) parser.add_argument('''--checkpoint_interval''' , type=__a , default=4000 , help='''Checkpoint interval.''' ) UpperCamelCase__ :Dict = parser.parse_args() sanity_checks(__a ) # ARGS # init_gpu_params(__a ) set_seed(__a ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( f'''Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite''' ''' itUse `--force` if you want to overwrite it''' ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(f'''Experiment will be dumped and logged in {args.dump_path}''' ) # SAVE PARAMS # logger.info(f'''Param: {args}''' ) with open(os.path.join(args.dump_path , '''parameters.json''' ) , '''w''' ) as f: json.dump(vars(__a ) , __a , indent=4 ) git_log(args.dump_path ) UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ :Any = MODEL_CLASSES[args.student_type] UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ :Any = MODEL_CLASSES[args.teacher_type] # TOKENIZER # UpperCamelCase__ :Union[str, Any] = teacher_tokenizer_class.from_pretrained(args.teacher_name ) UpperCamelCase__ :str = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): UpperCamelCase__ :List[str] = tokenizer.all_special_tokens.index(__a ) UpperCamelCase__ :Dict = tokenizer.all_special_ids[idx] logger.info(f'''Special tokens {special_tok_ids}''' ) UpperCamelCase__ :Any = special_tok_ids UpperCamelCase__ :Any = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f'''Loading data from {args.data_file}''' ) with open(args.data_file , '''rb''' ) as fp: UpperCamelCase__ :int = pickle.load(__a ) if args.mlm: logger.info(f'''Loading token counts from {args.token_counts} (already pre-computed)''' ) with open(args.token_counts , '''rb''' ) as fp: UpperCamelCase__ :Any = pickle.load(__a ) UpperCamelCase__ :Optional[Any] = np.maximum(__a , 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): UpperCamelCase__ :Union[str, Any] = 0.0 # do not predict special tokens UpperCamelCase__ :str = torch.from_numpy(__a ) else: UpperCamelCase__ :Optional[Any] = None UpperCamelCase__ :str = LmSeqsDataset(params=__a , data=__a ) logger.info('''Data loader created.''' ) # STUDENT # logger.info(f'''Loading student config from {args.student_config}''' ) UpperCamelCase__ :Dict = student_config_class.from_pretrained(args.student_config ) UpperCamelCase__ :int = True if args.student_pretrained_weights is not None: logger.info(f'''Loading pretrained weights from {args.student_pretrained_weights}''' ) UpperCamelCase__ :str = student_model_class.from_pretrained(args.student_pretrained_weights , config=__a ) else: UpperCamelCase__ :Optional[int] = student_model_class(__a ) if args.n_gpu > 0: student.to(f'''cuda:{args.local_rank}''' ) logger.info('''Student loaded.''' ) # TEACHER # UpperCamelCase__ :List[Any] = teacher_model_class.from_pretrained(args.teacher_name , output_hidden_states=__a ) if args.n_gpu > 0: teacher.to(f'''cuda:{args.local_rank}''' ) logger.info(f'''Teacher loaded from {args.teacher_name}.''' ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(__a , __a ) if args.freeze_token_type_embds: freeze_token_type_embeddings(__a , __a ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() UpperCamelCase__ :str = Distiller( params=__a , dataset=__a , token_probs=__a , student=__a , teacher=__a ) distiller.train() logger.info('''Let\'s go get some drinks.''' ) if __name__ == "__main__": main()
219
'''simple docstring''' from __future__ import annotations __snake_case = [True] * 1000001 __snake_case = 2 while i * i <= 1000000: if seive[i]: for j in range(i * i, 1000001, i): __snake_case = False i += 1 def a ( __a ) -> bool: '''simple docstring''' return seive[n] def a ( __a ) -> bool: '''simple docstring''' return any(digit in '''02468''' for digit in str(__a ) ) def a ( __a = 1000000 ) -> list[int]: '''simple docstring''' UpperCamelCase__ :Any = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(__a ) and not contains_an_even_digit(__a ): UpperCamelCase__ :str = str(__a ) UpperCamelCase__ :List[str] = [int(str_num[j:] + str_num[:j] ) for j in range(len(__a ) )] if all(is_prime(__a ) for i in list_nums ): result.append(__a ) return result def a ( ) -> int: '''simple docstring''' return len(find_circular_primes() ) if __name__ == "__main__": print(F"""{len(find_circular_primes()) = }""")
219
1
def lowercase__ ( __snake_case : dict ): '''simple docstring''' UpperCAmelCase_ : set[int] = set() # To detect a back edge, keep track of vertices currently in the recursion stack UpperCAmelCase_ : set[int] = set() return any( node not in visited and depth_first_search(__snake_case , __snake_case , __snake_case , __snake_case ) for node in graph ) def lowercase__ ( __snake_case : dict , __snake_case : int , __snake_case : set , __snake_case : set ): '''simple docstring''' visited.add(__snake_case ) rec_stk.add(__snake_case ) for node in graph[vertex]: if node not in visited: if depth_first_search(__snake_case , __snake_case , __snake_case , __snake_case ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(__snake_case ) return False if __name__ == "__main__": from doctest import testmod testmod()
29
def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if edge <= 0 or not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise ValueError('''Length must be a positive.''' ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if edge <= 0 or not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise ValueError('''Length must be a positive.''' ) return ((15 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
101
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : List[Any] = [2, 2, 6, 2] if 'tiny' in model_name else [2, 2, 1_8, 2] SCREAMING_SNAKE_CASE_ : str = True if 'large' in model_name or 'huge' in model_name else False SCREAMING_SNAKE_CASE_ : Tuple = True if 'large' in model_name or 'huge' in model_name else False SCREAMING_SNAKE_CASE_ : Dict = True if 'large' in model_name or 'huge' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: SCREAMING_SNAKE_CASE_ : List[Any] = [3, 3, 3, 3] SCREAMING_SNAKE_CASE_ : Optional[Any] = [5, 5, 5, 5] elif "fl4" in model_name: SCREAMING_SNAKE_CASE_ : int = [4, 4, 4, 4] SCREAMING_SNAKE_CASE_ : List[str] = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: SCREAMING_SNAKE_CASE_ : int = [3, 3, 3, 3] if "lrf" in model_name: SCREAMING_SNAKE_CASE_ : int = [3, 3, 3, 3] else: SCREAMING_SNAKE_CASE_ : Optional[Any] = [2, 2, 2, 2] if "tiny" in model_name: SCREAMING_SNAKE_CASE_ : List[Any] = 9_6 elif "small" in model_name: SCREAMING_SNAKE_CASE_ : Tuple = 9_6 elif "base" in model_name: SCREAMING_SNAKE_CASE_ : Optional[int] = 1_2_8 elif "large" in model_name: SCREAMING_SNAKE_CASE_ : Optional[int] = 1_9_2 elif "xlarge" in model_name: SCREAMING_SNAKE_CASE_ : Union[str, Any] = 2_5_6 elif "huge" in model_name: SCREAMING_SNAKE_CASE_ : str = 3_5_2 # set label information SCREAMING_SNAKE_CASE_ : Tuple = 'huggingface/label-files' if "large" in model_name or "huge" in model_name: SCREAMING_SNAKE_CASE_ : List[str] = 'imagenet-22k-id2label.json' else: SCREAMING_SNAKE_CASE_ : Dict = 'imagenet-1k-id2label.json' SCREAMING_SNAKE_CASE_ : Optional[int] = json.load(open(hf_hub_download(A__, A__, repo_type='dataset' ), 'r' ) ) SCREAMING_SNAKE_CASE_ : Dict = {int(A__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE_ : int = {v: k for k, v in idalabel.items()} SCREAMING_SNAKE_CASE_ : str = FocalNetConfig( embed_dim=A__, depths=A__, focal_levels=A__, focal_windows=A__, use_conv_embed=A__, idalabel=A__, labelaid=A__, use_post_layernorm=A__, use_layerscale=A__, ) return config def a__ ( A__ ): if "patch_embed.proj" in name: SCREAMING_SNAKE_CASE_ : str = name.replace('patch_embed.proj', 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: SCREAMING_SNAKE_CASE_ : Union[str, Any] = name.replace('patch_embed.norm', 'embeddings.norm' ) if "layers" in name: SCREAMING_SNAKE_CASE_ : str = 'encoder.' + name if "encoder.layers" in name: SCREAMING_SNAKE_CASE_ : int = name.replace('encoder.layers', 'encoder.stages' ) if "downsample.proj" in name: SCREAMING_SNAKE_CASE_ : Union[str, Any] = name.replace('downsample.proj', 'downsample.projection' ) if "blocks" in name: SCREAMING_SNAKE_CASE_ : Optional[int] = name.replace('blocks', 'layers' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: SCREAMING_SNAKE_CASE_ : str = name.replace('modulation.f', 'modulation.projection_in' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: SCREAMING_SNAKE_CASE_ : Optional[int] = name.replace('modulation.h', 'modulation.projection_context' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: SCREAMING_SNAKE_CASE_ : Any = name.replace('modulation.proj', 'modulation.projection_out' ) if name == "norm.weight": SCREAMING_SNAKE_CASE_ : Optional[int] = 'layernorm.weight' if name == "norm.bias": SCREAMING_SNAKE_CASE_ : Optional[int] = 'layernorm.bias' if "head" in name: SCREAMING_SNAKE_CASE_ : Union[str, Any] = name.replace('head', 'classifier' ) else: SCREAMING_SNAKE_CASE_ : Optional[int] = 'focalnet.' + name return name def a__ ( A__, A__, A__=False ): # fmt: off SCREAMING_SNAKE_CASE_ : str = { 'focalnet-tiny': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth', 'focalnet-tiny-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth', 'focalnet-small': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth', 'focalnet-small-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth', 'focalnet-base': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth', 'focalnet-base-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth', 'focalnet-large-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth', 'focalnet-large-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth', 'focalnet-xlarge-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth', 'focalnet-xlarge-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth', } # fmt: on SCREAMING_SNAKE_CASE_ : str = model_name_to_url[model_name] print('Checkpoint URL: ', A__ ) SCREAMING_SNAKE_CASE_ : Optional[Any] = torch.hub.load_state_dict_from_url(A__, map_location='cpu' )['model'] # rename keys for key in state_dict.copy().keys(): SCREAMING_SNAKE_CASE_ : int = state_dict.pop(A__ ) SCREAMING_SNAKE_CASE_ : List[str] = val SCREAMING_SNAKE_CASE_ : Union[str, Any] = get_focalnet_config(A__ ) SCREAMING_SNAKE_CASE_ : Tuple = FocalNetForImageClassification(A__ ) model.eval() # load state dict model.load_state_dict(A__ ) # verify conversion SCREAMING_SNAKE_CASE_ : int = 'http://images.cocodataset.org/val2017/000000039769.jpg' SCREAMING_SNAKE_CASE_ : int = BitImageProcessor( do_resize=A__, size={'shortest_edge': 2_5_6}, resample=PILImageResampling.BILINEAR, do_center_crop=A__, crop_size=2_2_4, do_normalize=A__, image_mean=A__, image_std=A__, ) SCREAMING_SNAKE_CASE_ : List[str] = Image.open(requests.get(A__, stream=A__ ).raw ) SCREAMING_SNAKE_CASE_ : str = processor(images=A__, return_tensors='pt' ) SCREAMING_SNAKE_CASE_ : Optional[int] = transforms.Compose( [ transforms.Resize(2_5_6 ), transforms.CenterCrop(2_2_4 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_85, 0.4_56, 0.4_06], std=[0.2_29, 0.2_24, 0.2_25] ), ] ) SCREAMING_SNAKE_CASE_ : Tuple = image_transforms(A__ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values, A__, atol=1E-4 ) SCREAMING_SNAKE_CASE_ : List[Any] = model(**A__ ) SCREAMING_SNAKE_CASE_ : Tuple = outputs.logits.argmax(-1 ).item() print('Predicted class:', model.config.idalabel[predicted_class_idx] ) print('First values of logits:', outputs.logits[0, :3] ) if model_name == "focalnet-tiny": SCREAMING_SNAKE_CASE_ : Union[str, Any] = torch.tensor([0.21_66, -0.43_68, 0.21_91] ) elif model_name == "focalnet-tiny-lrf": SCREAMING_SNAKE_CASE_ : Tuple = torch.tensor([1.16_69, 0.01_25, -0.16_95] ) elif model_name == "focalnet-small": SCREAMING_SNAKE_CASE_ : Dict = torch.tensor([0.49_17, -0.04_30, 0.13_41] ) elif model_name == "focalnet-small-lrf": SCREAMING_SNAKE_CASE_ : Dict = torch.tensor([-0.25_88, -0.53_42, -0.23_31] ) elif model_name == "focalnet-base": SCREAMING_SNAKE_CASE_ : List[str] = torch.tensor([-0.16_55, -0.40_90, -0.17_30] ) elif model_name == "focalnet-base-lrf": SCREAMING_SNAKE_CASE_ : Dict = torch.tensor([0.53_06, -0.04_83, -0.39_28] ) assert torch.allclose(outputs.logits[0, :3], A__, atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(A__ ) processor.save_pretrained(A__ ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": lowerCAmelCase__ : int =argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='focalnet-tiny', type=str, help='Name of the FocalNet model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub.', ) lowerCAmelCase__ : Tuple =parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
162
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class __lowercase : """simple docstring""" _UpperCAmelCase = 42 _UpperCAmelCase = None _UpperCAmelCase = None def a__ ( ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = Node(1 ) SCREAMING_SNAKE_CASE_ : Any = Node(2 ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = Node(3 ) SCREAMING_SNAKE_CASE_ : int = Node(4 ) SCREAMING_SNAKE_CASE_ : List[str] = Node(5 ) return tree def a__ ( A__ ): return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def a__ ( A__ ): return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def a__ ( A__ ): return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def a__ ( A__ ): return (max(height(root.left ), height(root.right ) ) + 1) if root else 0 def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : list[Any] = [] if root is None: return output SCREAMING_SNAKE_CASE_ : int = deque([root] ) while process_queue: SCREAMING_SNAKE_CASE_ : List[str] = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def a__ ( A__, A__ ): SCREAMING_SNAKE_CASE_ : list[Any] = [] def populate_output(A__, A__ ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left, level - 1 ) populate_output(root.right, level - 1 ) populate_output(A__, A__ ) return output def a__ ( A__, A__ ): SCREAMING_SNAKE_CASE_ : list[Any] = [] def populate_output(A__, A__ ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right, level - 1 ) populate_output(root.left, level - 1 ) populate_output(A__, A__ ) return output def a__ ( A__ ): if root is None: return [] SCREAMING_SNAKE_CASE_ : list[Sequence[Node | None]] = [] SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 SCREAMING_SNAKE_CASE_ : Optional[Any] = height(A__ ) for h in range(1, height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(A__, A__ ) ) SCREAMING_SNAKE_CASE_ : List[Any] = 1 else: output.append(get_nodes_from_right_to_left(A__, A__ ) ) SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 return output def a__ ( ): # Main function for testing. SCREAMING_SNAKE_CASE_ : Optional[int] = make_tree() print(F'''In-order Traversal: {inorder(A__ )}''' ) print(F'''Pre-order Traversal: {preorder(A__ )}''' ) print(F'''Post-order Traversal: {postorder(A__ )}''', '\n' ) print(F'''Height of Tree: {height(A__ )}''', '\n' ) print('Complete Level Order Traversal: ' ) print(level_order(A__ ), '\n' ) print('Level-wise order Traversal: ' ) for level in range(1, height(A__ ) + 1 ): print(F'''Level {level}:''', get_nodes_from_left_to_right(A__, level=A__ ) ) print('\nZigZag order Traversal: ' ) print(zigzag(A__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
162
1
def _lowerCAmelCase ( __lowerCAmelCase ) -> int: """simple docstring""" snake_case__ : Dict = len(__lowerCAmelCase ) snake_case__ : List[str] = len(matrix[0] ) snake_case__ : Tuple = min(__lowerCAmelCase , __lowerCAmelCase ) for row in range(__lowerCAmelCase ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , __lowerCAmelCase ): snake_case__ : Any = matrix[col][row] / matrix[row][row] for i in range(__lowerCAmelCase , __lowerCAmelCase ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows snake_case__ : str = True for i in range(row + 1 , __lowerCAmelCase ): if matrix[i][row] != 0: snake_case__ , snake_case__ : Tuple = matrix[i], matrix[row] snake_case__ : Dict = False break if reduce: rank -= 1 for i in range(__lowerCAmelCase ): snake_case__ : Optional[int] = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
230
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. A__ = abspath(join(dirname(dirname(dirname(__file__))), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def _lowerCAmelCase ( __lowerCAmelCase ) -> str: """simple docstring""" from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(__lowerCAmelCase ) def _lowerCAmelCase ( __lowerCAmelCase ) -> List[Any]: """simple docstring""" from transformers.testing_utils import pytest_terminal_summary_main snake_case__ : Dict = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(__lowerCAmelCase , id=__lowerCAmelCase )
230
1
"""simple docstring""" import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence from datasets.features import ArrayaD, ClassLabel, Features, Image, Value from datasets.features.features import ArrayaDExtensionType, cast_to_python_objects from datasets.keyhash import DuplicatedKeysError, InvalidKeyError from .utils import require_pil class _lowercase ( __a ): """simple docstring""" def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' __UpperCamelCase =pa.array(TypedSequence([1, 2, 3] ) ) self.assertEqual(arr.type , pa.intaa() ) def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' with self.assertRaises(UpperCamelCase__ ): __UpperCamelCase =pa.array(TypedSequence([1, 2, 3] ) , type=pa.intaa() ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: '''simple docstring''' with self.assertRaises(UpperCamelCase__ ): __UpperCamelCase =pa.array(TypedSequence([1, 2, 3] , try_type=Value('''bool''' ) , type=Value('''int64''' ) ) ) def UpperCAmelCase_ ( self : Dict ) -> List[Any]: '''simple docstring''' __UpperCamelCase =pa.array(TypedSequence([1, 2, 3] , type=Value('''int32''' ) ) ) self.assertEqual(arr.type , pa.intaa() ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): __UpperCamelCase =pa.array(TypedSequence(['''foo''', '''bar'''] , type=Value('''int64''' ) ) ) def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' __UpperCamelCase =pa.array(TypedSequence([1, 2, 3] , try_type=Value('''int32''' ) ) ) self.assertEqual(arr.type , pa.intaa() ) def UpperCAmelCase_ ( self : Any ) -> Tuple: '''simple docstring''' __UpperCamelCase =pa.array(TypedSequence(['''foo''', '''bar'''] , try_type=Value('''int64''' ) ) ) self.assertEqual(arr.type , pa.string() ) def UpperCAmelCase_ ( self : Any ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase =pa.array(TypedSequence([[[1, 2, 3]]] , type=ArrayaD((1, 3) , '''int64''' ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , '''int64''' ) ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: '''simple docstring''' with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): __UpperCamelCase =pa.array(TypedSequence(['''foo''', '''bar'''] , type=ArrayaD((1, 3) , '''int64''' ) ) ) def UpperCAmelCase_ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' __UpperCamelCase =pa.array(TypedSequence([[[1, 2, 3]]] , try_type=ArrayaD((1, 3) , '''int64''' ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , '''int64''' ) ) def UpperCAmelCase_ ( self : int ) -> Dict: '''simple docstring''' __UpperCamelCase =pa.array(TypedSequence(['''foo''', '''bar'''] , try_type=ArrayaD((1, 3) , '''int64''' ) ) ) self.assertEqual(arr.type , pa.string() ) @require_pil def UpperCAmelCase_ ( self : List[str] ) -> Optional[int]: '''simple docstring''' import PIL.Image __UpperCamelCase =PIL.Image.fromarray(np.arange(10 , dtype=np.uinta ).reshape(2 , 5 ) ) with patch( '''datasets.arrow_writer.cast_to_python_objects''' , side_effect=UpperCamelCase__ ) as mock_cast_to_python_objects: __UpperCamelCase =pa.array(TypedSequence([{'''path''': None, '''bytes''': B'''image_bytes'''}, pil_image] , type=Image() ) ) __UpperCamelCase , __UpperCamelCase =mock_cast_to_python_objects.call_args_list[-1] self.assertIn('''optimize_list_casting''' , UpperCamelCase__ ) self.assertFalse(kwargs['''optimize_list_casting'''] ) def lowerCAmelCase (__UpperCamelCase : List[str] , __UpperCamelCase : int ): """simple docstring""" __UpperCamelCase =pa.BufferReader(__UpperCamelCase ) if isinstance(__UpperCamelCase , pa.Buffer ) else pa.memory_map(__UpperCamelCase ) __UpperCamelCase =pa.ipc.open_stream(__UpperCamelCase ) __UpperCamelCase =f.read_all() assert len(pa_table.to_batches() ) == expected_num_chunks assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} del pa_table @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def lowerCAmelCase (__UpperCamelCase : List[str] , __UpperCamelCase : List[Any] ): """simple docstring""" __UpperCamelCase =pa.BufferOutputStream() __UpperCamelCase =pa.schema(__UpperCamelCase ) if fields else None with ArrowWriter(stream=__UpperCamelCase , schema=__UpperCamelCase , writer_batch_size=__UpperCamelCase ) as writer: writer.write({'''col_1''': '''foo''', '''col_2''': 1} ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} ) __UpperCamelCase , __UpperCamelCase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: __UpperCamelCase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(__UpperCamelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def lowerCAmelCase (): """simple docstring""" __UpperCamelCase =pa.BufferOutputStream() __UpperCamelCase =Features({'''labels''': ClassLabel(names=['''neg''', '''pos'''] )} ) with ArrowWriter(stream=__UpperCamelCase , features=__UpperCamelCase ) as writer: writer.write({'''labels''': 0} ) writer.write({'''labels''': 1} ) __UpperCamelCase , __UpperCamelCase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == features.arrow_schema assert writer._schema.metadata == features.arrow_schema.metadata __UpperCamelCase =pa.BufferReader(output.getvalue() ) __UpperCamelCase =pa.ipc.open_stream(__UpperCamelCase ) __UpperCamelCase =f.read_all() __UpperCamelCase =pa_table.schema assert pa_table.num_rows == 2 assert schema == features.arrow_schema assert schema.metadata == features.arrow_schema.metadata assert features == Features.from_arrow_schema(__UpperCamelCase ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) def lowerCAmelCase (__UpperCamelCase : Dict ): """simple docstring""" __UpperCamelCase =pa.BufferOutputStream() with ArrowWriter( stream=__UpperCamelCase , writer_batch_size=__UpperCamelCase , hash_salt='''split_name''' , check_duplicates=__UpperCamelCase , ) as writer: with pytest.raises(__UpperCamelCase ): writer.write({'''col_1''': '''foo''', '''col_2''': 1} , key=[1, 2] ) __UpperCamelCase , __UpperCamelCase =writer.finalize() @pytest.mark.parametrize('''writer_batch_size''' , [None, 2, 1_0] ) def lowerCAmelCase (__UpperCamelCase : List[Any] ): """simple docstring""" __UpperCamelCase =pa.BufferOutputStream() with ArrowWriter( stream=__UpperCamelCase , writer_batch_size=__UpperCamelCase , hash_salt='''split_name''' , check_duplicates=__UpperCamelCase , ) as writer: with pytest.raises(__UpperCamelCase ): writer.write({'''col_1''': '''foo''', '''col_2''': 1} , key=1_0 ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} , key=1_0 ) __UpperCamelCase , __UpperCamelCase =writer.finalize() @pytest.mark.parametrize('''writer_batch_size''' , [None, 2, 1_0] ) def lowerCAmelCase (__UpperCamelCase : str ): """simple docstring""" __UpperCamelCase =pa.BufferOutputStream() with ArrowWriter( stream=__UpperCamelCase , writer_batch_size=__UpperCamelCase , hash_salt='''split_name''' , check_duplicates=__UpperCamelCase , ) as writer: writer.write({'''col_1''': '''foo''', '''col_2''': 1} , key=1 ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} , key=2 ) __UpperCamelCase , __UpperCamelCase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def lowerCAmelCase (__UpperCamelCase : Dict , __UpperCamelCase : Any ): """simple docstring""" __UpperCamelCase =pa.BufferOutputStream() __UpperCamelCase =pa.schema(__UpperCamelCase ) if fields else None with ArrowWriter(stream=__UpperCamelCase , schema=__UpperCamelCase , writer_batch_size=__UpperCamelCase ) as writer: writer.write_batch({'''col_1''': ['''foo''', '''bar'''], '''col_2''': [1, 2]} ) writer.write_batch({'''col_1''': [], '''col_2''': []} ) __UpperCamelCase , __UpperCamelCase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: __UpperCamelCase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(__UpperCamelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def lowerCAmelCase (__UpperCamelCase : List[Any] , __UpperCamelCase : str ): """simple docstring""" __UpperCamelCase =pa.BufferOutputStream() __UpperCamelCase =pa.schema(__UpperCamelCase ) if fields else None with ArrowWriter(stream=__UpperCamelCase , schema=__UpperCamelCase , writer_batch_size=__UpperCamelCase ) as writer: writer.write_table(pa.Table.from_pydict({'''col_1''': ['''foo''', '''bar'''], '''col_2''': [1, 2]} ) ) __UpperCamelCase , __UpperCamelCase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: __UpperCamelCase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(__UpperCamelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('''writer_batch_size''' , [None, 1, 1_0] ) @pytest.mark.parametrize( '''fields''' , [None, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}, {'''col_1''': pa.string(), '''col_2''': pa.intaa()}] ) def lowerCAmelCase (__UpperCamelCase : str , __UpperCamelCase : List[Any] ): """simple docstring""" __UpperCamelCase =pa.BufferOutputStream() __UpperCamelCase =pa.schema(__UpperCamelCase ) if fields else None with ArrowWriter(stream=__UpperCamelCase , schema=__UpperCamelCase , writer_batch_size=__UpperCamelCase ) as writer: writer.write_row(pa.Table.from_pydict({'''col_1''': ['''foo'''], '''col_2''': [1]} ) ) writer.write_row(pa.Table.from_pydict({'''col_1''': ['''bar'''], '''col_2''': [2]} ) ) __UpperCamelCase , __UpperCamelCase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: __UpperCamelCase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} assert writer._schema == pa.schema(__UpperCamelCase , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def lowerCAmelCase (): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __UpperCamelCase ={'''col_1''': pa.string(), '''col_2''': pa.intaa()} __UpperCamelCase =os.path.join(__UpperCamelCase , '''test.arrow''' ) with ArrowWriter(path=__UpperCamelCase , schema=pa.schema(__UpperCamelCase ) ) as writer: writer.write_batch({'''col_1''': ['''foo''', '''bar'''], '''col_2''': [1, 2]} ) __UpperCamelCase , __UpperCamelCase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == pa.schema(__UpperCamelCase , metadata=writer._schema.metadata ) _check_output(__UpperCamelCase , 1 ) def lowerCAmelCase (__UpperCamelCase : List[str] ): """simple docstring""" if pa.types.is_list(__UpperCamelCase ): return get_base_dtype(arr_type.value_type ) else: return arr_type def lowerCAmelCase (__UpperCamelCase : Union[str, Any] , __UpperCamelCase : Dict ): """simple docstring""" if isinstance(lst[0] , __UpperCamelCase ): change_first_primitive_element_in_list(lst[0] , __UpperCamelCase ) else: __UpperCamelCase =value @pytest.mark.parametrize('''optimized_int_type, expected_dtype''' , [(None, pa.intaa()), (Value('''int32''' ), pa.intaa())] ) @pytest.mark.parametrize('''sequence''' , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def lowerCAmelCase (__UpperCamelCase : int , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : int ): """simple docstring""" __UpperCamelCase =pa.array(TypedSequence(__UpperCamelCase , optimized_int_type=__UpperCamelCase ) ) assert get_base_dtype(arr.type ) == expected_dtype @pytest.mark.parametrize( '''col, expected_dtype''' , [ ('''attention_mask''', pa.inta()), ('''special_tokens_mask''', pa.inta()), ('''token_type_ids''', pa.inta()), ('''input_ids''', pa.intaa()), ('''other''', pa.intaa()), ] , ) @pytest.mark.parametrize('''sequence''' , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def lowerCAmelCase (__UpperCamelCase : Optional[Any] , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : Optional[Any] ): """simple docstring""" __UpperCamelCase =pa.array(OptimizedTypedSequence(__UpperCamelCase , col=__UpperCamelCase ) ) assert get_base_dtype(arr.type ) == expected_dtype # not in range if col != "other": # avoids errors due to in-place modifications __UpperCamelCase =copy.deepcopy(__UpperCamelCase ) __UpperCamelCase =np.iinfo(expected_dtype.to_pandas_dtype() ).max + 1 change_first_primitive_element_in_list(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase =pa.array(OptimizedTypedSequence(__UpperCamelCase , col=__UpperCamelCase ) ) assert get_base_dtype(arr.type ) == pa.intaa() @pytest.mark.parametrize('''raise_exception''' , [False, True] ) def lowerCAmelCase (__UpperCamelCase : str , __UpperCamelCase : List[Any] ): """simple docstring""" __UpperCamelCase =str(tmp_path / '''dataset-train.arrow''' ) try: with ArrowWriter(path=__UpperCamelCase ) as writer: if raise_exception: raise pa.lib.ArrowInvalid() else: writer.stream.close() except pa.lib.ArrowInvalid: pass finally: assert writer.stream.closed def lowerCAmelCase (__UpperCamelCase : Optional[int] ): """simple docstring""" __UpperCamelCase ='''mock://dataset-train.arrow''' with ArrowWriter(path=__UpperCamelCase , storage_options=mockfs.storage_options ) as writer: assert isinstance(writer._fs , type(__UpperCamelCase ) ) assert writer._fs.storage_options == mockfs.storage_options writer.write({'''col_1''': '''foo''', '''col_2''': 1} ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} ) __UpperCamelCase , __UpperCamelCase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert mockfs.exists(__UpperCamelCase ) def lowerCAmelCase (): """simple docstring""" __UpperCamelCase =pa.BufferOutputStream() with ParquetWriter(stream=__UpperCamelCase ) as writer: writer.write({'''col_1''': '''foo''', '''col_2''': 1} ) writer.write({'''col_1''': '''bar''', '''col_2''': 2} ) __UpperCamelCase , __UpperCamelCase =writer.finalize() assert num_examples == 2 assert num_bytes > 0 __UpperCamelCase =pa.BufferReader(output.getvalue() ) __UpperCamelCase =pq.read_table(__UpperCamelCase ) assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} @require_pil @pytest.mark.parametrize('''embed_local_files''' , [False, True] ) def lowerCAmelCase (__UpperCamelCase : Optional[int] , __UpperCamelCase : int ): """simple docstring""" import PIL.Image __UpperCamelCase =str(tmp_path / '''test_image_rgb.jpg''' ) PIL.Image.fromarray(np.zeros((5, 5) , dtype=np.uinta ) ).save(__UpperCamelCase , format='''png''' ) __UpperCamelCase =pa.BufferOutputStream() with ParquetWriter( stream=__UpperCamelCase , features=Features({'''image''': Image()} ) , embed_local_files=__UpperCamelCase ) as writer: writer.write({'''image''': image_path} ) writer.finalize() __UpperCamelCase =pa.BufferReader(output.getvalue() ) __UpperCamelCase =pq.read_table(__UpperCamelCase ) __UpperCamelCase =pa_table.to_pydict() if embed_local_files: assert isinstance(out['''image'''][0]['''path'''] , __UpperCamelCase ) with open(__UpperCamelCase , '''rb''' ) as f: assert out["image"][0]["bytes"] == f.read() else: assert out["image"][0]["path"] == image_path assert out["image"][0]["bytes"] is None def lowerCAmelCase (): """simple docstring""" __UpperCamelCase =pa.schema([pa.field('''col_1''' , pa.string() , nullable=__UpperCamelCase )] ) __UpperCamelCase =pa.BufferOutputStream() with ArrowWriter(stream=__UpperCamelCase ) as writer: writer._build_writer(inferred_schema=__UpperCamelCase ) assert writer._schema == pa.schema([pa.field('''col_1''' , pa.string() )] )
85
"""simple docstring""" from queue import PriorityQueue from typing import Any import numpy as np def lowerCAmelCase (__UpperCamelCase : dict , __UpperCamelCase : str , __UpperCamelCase : set , __UpperCamelCase : set , __UpperCamelCase : dict , __UpperCamelCase : dict , __UpperCamelCase : PriorityQueue , __UpperCamelCase : dict , __UpperCamelCase : float | int , ): """simple docstring""" for nxt, d in graph[v]: if nxt in visited_forward: continue __UpperCamelCase =cst_fwd.get(__UpperCamelCase , np.inf ) __UpperCamelCase =cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) __UpperCamelCase =new_cost_f __UpperCamelCase =v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: __UpperCamelCase =cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def lowerCAmelCase (__UpperCamelCase : str , __UpperCamelCase : str , __UpperCamelCase : dict , __UpperCamelCase : dict ): """simple docstring""" __UpperCamelCase =-1 __UpperCamelCase =set() __UpperCamelCase =set() __UpperCamelCase ={source: 0} __UpperCamelCase ={destination: 0} __UpperCamelCase ={source: None} __UpperCamelCase ={destination: None} __UpperCamelCase =PriorityQueue() __UpperCamelCase =PriorityQueue() __UpperCamelCase =np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): __UpperCamelCase , __UpperCamelCase =queue_forward.get() visited_forward.add(__UpperCamelCase ) __UpperCamelCase , __UpperCamelCase =queue_backward.get() visited_backward.add(__UpperCamelCase ) __UpperCamelCase =pass_and_relaxation( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ) __UpperCamelCase =pass_and_relaxation( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: __UpperCamelCase =shortest_distance return shortest_path_distance __lowercase = { '''B''': [['''C''', 1]], '''C''': [['''D''', 1]], '''D''': [['''F''', 1]], '''E''': [['''B''', 1], ['''G''', 2]], '''F''': [], '''G''': [['''F''', 1]], } __lowercase = { '''B''': [['''E''', 1]], '''C''': [['''B''', 1]], '''D''': [['''C''', 1]], '''F''': [['''D''', 1], ['''G''', 1]], '''E''': [[None, np.inf]], '''G''': [['''E''', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
85
1
"""simple docstring""" import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis_float32 (there's also the fix_lavis branch) # also note: to convert Vicuna checkpoints, we had to include /home/niels/python_projects/checkpoints/FastChat/vicuna-7b in lavis/configs/models/blip2/blip2_instruct_vicuna7b.yaml # same for Vicuna-13b from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipImageProcessor, InstructBlipConfig, InstructBlipForConditionalGeneration, InstructBlipProcessor, InstructBlipQFormerConfig, InstructBlipVisionConfig, LlamaConfig, LlamaTokenizerFast, TaConfig, TaTokenizerFast, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def A_ ( ): '''simple docstring''' snake_case_ :Optional[int] = """https://raw.githubusercontent.com/salesforce/LAVIS/main/docs/_static/Confusing-Pictures.jpg""" snake_case_ :Optional[Any] = Image.open(requests.get(_lowercase, stream=_lowercase ).raw ).convert("""RGB""" ) return image def A_ ( _lowercase ): '''simple docstring''' snake_case_ :str = [] # fmt: off # vision encoder rename_keys.append(("""visual_encoder.cls_token""", """vision_model.embeddings.class_embedding""") ) rename_keys.append(("""visual_encoder.pos_embed""", """vision_model.embeddings.position_embedding""") ) rename_keys.append(("""visual_encoder.patch_embed.proj.weight""", """vision_model.embeddings.patch_embedding.weight""") ) rename_keys.append(("""visual_encoder.patch_embed.proj.bias""", """vision_model.embeddings.patch_embedding.bias""") ) rename_keys.append(("""ln_vision.weight""", """vision_model.post_layernorm.weight""") ) rename_keys.append(("""ln_vision.bias""", """vision_model.post_layernorm.bias""") ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((f"""visual_encoder.blocks.{i}.norm1.weight""", f"""vision_model.encoder.layers.{i}.layer_norm1.weight""") ) rename_keys.append((f"""visual_encoder.blocks.{i}.norm1.bias""", f"""vision_model.encoder.layers.{i}.layer_norm1.bias""") ) rename_keys.append((f"""visual_encoder.blocks.{i}.norm2.weight""", f"""vision_model.encoder.layers.{i}.layer_norm2.weight""") ) rename_keys.append((f"""visual_encoder.blocks.{i}.norm2.bias""", f"""vision_model.encoder.layers.{i}.layer_norm2.bias""") ) rename_keys.append((f"""visual_encoder.blocks.{i}.attn.qkv.weight""", f"""vision_model.encoder.layers.{i}.self_attn.qkv.weight""") ) rename_keys.append((f"""visual_encoder.blocks.{i}.attn.proj.weight""", f"""vision_model.encoder.layers.{i}.self_attn.projection.weight""",) ) rename_keys.append((f"""visual_encoder.blocks.{i}.attn.proj.bias""", f"""vision_model.encoder.layers.{i}.self_attn.projection.bias""") ) rename_keys.append((f"""visual_encoder.blocks.{i}.mlp.fc1.weight""", f"""vision_model.encoder.layers.{i}.mlp.fc1.weight""") ) rename_keys.append((f"""visual_encoder.blocks.{i}.mlp.fc1.bias""", f"""vision_model.encoder.layers.{i}.mlp.fc1.bias""") ) rename_keys.append((f"""visual_encoder.blocks.{i}.mlp.fc2.weight""", f"""vision_model.encoder.layers.{i}.mlp.fc2.weight""") ) rename_keys.append((f"""visual_encoder.blocks.{i}.mlp.fc2.bias""", f"""vision_model.encoder.layers.{i}.mlp.fc2.bias""") ) # QFormer rename_keys.append(("""Qformer.bert.embeddings.LayerNorm.weight""", """qformer.embeddings.layernorm.weight""") ) rename_keys.append(("""Qformer.bert.embeddings.LayerNorm.bias""", """qformer.embeddings.layernorm.bias""") ) # fmt: on return rename_keys def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' snake_case_ :str = dct.pop(_lowercase ) snake_case_ :Dict = val def A_ ( _lowercase, _lowercase ): '''simple docstring''' for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases snake_case_ :Dict = state_dict.pop(f"""visual_encoder.blocks.{i}.attn.q_bias""" ) snake_case_ :Tuple = state_dict.pop(f"""visual_encoder.blocks.{i}.attn.v_bias""" ) # next, set bias in the state dict snake_case_ :Union[str, Any] = torch.cat((q_bias, torch.zeros_like(_lowercase, requires_grad=_lowercase ), v_bias) ) snake_case_ :int = qkv_bias def A_ ( _lowercase ): '''simple docstring''' snake_case_ :List[str] = 364 if """coco""" in model_name else 224 snake_case_ :Tuple = InstructBlipVisionConfig(image_size=_lowercase ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "t5-xl" in model_name: snake_case_ :Optional[Any] = TaConfig.from_pretrained("""google/flan-t5-xl""", dense_act_fn="""gelu""", bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: snake_case_ :List[Any] = TaConfig.from_pretrained("""google/flan-t5-xxl""", dense_act_fn="""gelu""", bos_token_id=1 ).to_dict() elif "vicuna-7b" in model_name: snake_case_ :Optional[int] = LlamaConfig.from_pretrained("""decapoda-research/llama-7b-hf""", vocab_size=32001 ).to_dict() elif "vicuna-13b" in model_name: snake_case_ :Optional[Any] = LlamaConfig.from_pretrained("""decapoda-research/llama-13b-hf""", vocab_size=32001 ).to_dict() else: raise ValueError("""Model name not supported""" ) # the authors add one special "[DEC]" token to the vocab of Q-Former, hence vocab size = 30522 + 1 snake_case_ :Optional[int] = InstructBlipQFormerConfig(vocab_size=30523 ).to_dict() snake_case_ :List[str] = InstructBlipConfig(vision_config=_lowercase, text_config=_lowercase, qformer_config=_lowercase ) return config, image_size @torch.no_grad() def A_ ( _lowercase, _lowercase=None, _lowercase=False ): '''simple docstring''' snake_case_ :int = AutoTokenizer.from_pretrained("""bert-base-uncased""", truncation_side="""left""" ) qformer_tokenizer.add_special_tokens({"""bos_token""": """[DEC]"""} ) if "t5" in model_name: snake_case_ :Optional[int] = TaTokenizerFast.from_pretrained("""google/flan-t5-xl""", truncation_side="""left""" ) elif "vicuna" in model_name: # the following was used in the original implementation: # tokenizer = LlamaTokenizer.from_pretrained("huggyllama/llama-7b", use_fast=False, truncation_side="left") # tokenizer.add_special_tokens({"pad_token": "[PAD]"}) # tokenizer.add_special_tokens({"bos_token": "</s>"}) # tokenizer.add_special_tokens({"eos_token": "</s>"}) # tokenizer.add_special_tokens({"unk_token": "</s>"}) snake_case_ :str = LlamaTokenizerFast.from_pretrained( """huggyllama/llama-7b""", truncation_side="""left""", bos_token="""</s>""", unk_token="""</s>""" ) tokenizer.add_special_tokens({"""pad_token""": """[PAD]"""} ) snake_case_, snake_case_ :Union[str, Any] = get_blipa_config(_lowercase ) snake_case_ :Union[str, Any] = InstructBlipForConditionalGeneration(_lowercase ).eval() snake_case_ :Union[str, Any] = { """instructblip-vicuna-7b""": ("""blip2_vicuna_instruct""", """vicuna7b"""), """instructblip-vicuna-13b""": ("""blip2_vicuna_instruct""", """vicuna13b"""), """instructblip-flan-t5-xl""": ("""blip2_t5_instruct""", """flant5xl"""), """instructblip-flan-t5-xxl""": ("""blip2_t5_instruct""", """flant5xxl"""), } snake_case_, snake_case_ :List[Any] = model_name_to_original[model_name] # load original model print("""Loading original model...""" ) snake_case_ :Optional[Any] = """cuda:1""" if torch.cuda.is_available() else """cpu""" snake_case_ :Any = """cuda:2""" if torch.cuda.is_available() else """cpu""" snake_case_, snake_case_, snake_case_ :int = load_model_and_preprocess( name=_lowercase, model_type=_lowercase, is_eval=_lowercase, device=_lowercase ) original_model.eval() print("""Done!""" ) # update state dict keys snake_case_ :int = original_model.state_dict() snake_case_ :int = create_rename_keys(_lowercase ) for src, dest in rename_keys: rename_key(_lowercase, _lowercase, _lowercase ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): snake_case_ :Optional[Any] = state_dict.pop(_lowercase ) if key.startswith("""Qformer.bert""" ): snake_case_ :str = key.replace("""Qformer.bert""", """qformer""" ) if "attention.self" in key: snake_case_ :int = key.replace("""self""", """attention""" ) if "llm_proj" in key: snake_case_ :Dict = key.replace("""llm_proj""", """language_projection""" ) if "t5_proj" in key: snake_case_ :Dict = key.replace("""t5_proj""", """language_projection""" ) if key.startswith("""llm_model""" ): snake_case_ :Union[str, Any] = key.replace("""llm_model""", """language_model""" ) if key.startswith("""t5""" ): snake_case_ :List[str] = key.replace("""t5""", """language""" ) snake_case_ :List[str] = val # read in qv biases read_in_q_v_bias(_lowercase, _lowercase ) # note: weights get loaded in torch.float32 by default hf_model.load_state_dict(_lowercase, strict=_lowercase ) snake_case_ :Optional[Any] = load_demo_image() snake_case_ :Tuple = """What is unusual about this image?""" # create processor snake_case_ :Tuple = BlipImageProcessor( size={"""height""": image_size, """width""": image_size}, image_mean=_lowercase, image_std=_lowercase ) snake_case_ :Tuple = InstructBlipProcessor( image_processor=_lowercase, tokenizer=_lowercase, qformer_tokenizer=_lowercase, ) snake_case_ :Tuple = processor(images=_lowercase, text=_lowercase, return_tensors="""pt""" ).to(_lowercase ) # make sure processor creates exact same pixel values snake_case_ :Optional[Any] = vis_processors["""eval"""](_lowercase ).unsqueeze(0 ).to(_lowercase ) snake_case_ :List[str] = inputs.pixel_values assert torch.allclose(original_pixel_values.to(pixel_values.device ), _lowercase ) original_model.to(_lowercase ) hf_model.to(_lowercase ) with torch.no_grad(): if "vicuna" in model_name: snake_case_ :List[Any] = original_model({"""image""": original_pixel_values, """text_input""": [prompt]} ).logits snake_case_ :Union[str, Any] = hf_model(**_lowercase ).logits else: snake_case_ :Dict = original_model( {"""image""": original_pixel_values, """text_input""": [prompt], """text_output""": ["""\n"""]} ).logits snake_case_ :Dict = tokenizer("""\n""", return_tensors="""pt""" ).input_ids.to(_lowercase ) snake_case_ :List[str] = label_input_ids.masked_fill(label_input_ids == tokenizer.pad_token_id, -100 ) snake_case_ :List[Any] = hf_model(**_lowercase, labels=_lowercase ).logits print("""First values of original logits:""", original_logits[0, :3, :3] ) print("""First values of HF logits:""", logits[0, :3, :3] ) # assert values assert original_logits.shape == logits.shape snake_case_ :Any = 1e-4 if """vicuna""" in model_name else 1e-5 assert torch.allclose(original_logits.to(logits.device ), _lowercase, atol=_lowercase ) print("""Looks ok!""" ) print("""Generating with original model...""" ) snake_case_ :Dict = original_model.generate({"""image""": original_pixel_values, """prompt""": prompt}, num_beams=5 ) # important: we need to cast the weights of the HF model to the appropriate type print("""Generating with HF model...""" ) snake_case_ :str = hf_model.generate( **_lowercase, do_sample=_lowercase, num_beams=5, max_length=256, min_length=1, top_p=0.9, repetition_penalty=1.5, length_penalty=1.0, temperature=1, ) if "vicuna" in model_name: # convert output id 0 to 2 (eos_token_id) # TODO add this in the generate method? snake_case_ :Optional[int] = 2 print("""Original generation:""", _lowercase ) snake_case_ :str = processor.batch_decode(_lowercase, skip_special_tokens=_lowercase ) snake_case_ :Any = [text.strip() for text in output_text] print("""HF generation:""", _lowercase ) if pytorch_dump_folder_path is not None: processor.save_pretrained(_lowercase ) hf_model.save_pretrained(_lowercase ) if push_to_hub: processor.push_to_hub(f"""Salesforce/{model_name}""" ) hf_model.push_to_hub(f"""Salesforce/{model_name}""" ) if __name__ == "__main__": __a = argparse.ArgumentParser() __a = [ "instructblip-vicuna-7b", "instructblip-vicuna-13b", "instructblip-flan-t5-xl", "instructblip-flan-t5-xxl", ] parser.add_argument( "--model_name", default="instructblip-flan-t5-xl", choices=choices, type=str, help="Path to hf config.json of model to convert", ) parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub after converting", ) __a = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
"""simple docstring""" import argparse import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( CLIPTokenizer, CLIPTokenizerFast, VideoMAEImageProcessor, XCLIPConfig, XCLIPModel, XCLIPProcessor, XCLIPTextConfig, XCLIPVisionConfig, ) def A_ ( _lowercase, _lowercase ): '''simple docstring''' snake_case_ :int = XCLIPTextConfig() # derive patch size from model name snake_case_ :Union[str, Any] = model_name.find("""patch""" ) snake_case_ :List[str] = int(model_name[start_idx + len("""patch""" ) : start_idx + len("""patch""" ) + 2] ) snake_case_ :Any = XCLIPVisionConfig(patch_size=_lowercase, num_frames=_lowercase ) if "large" in model_name: snake_case_ :Optional[Any] = 768 snake_case_ :Union[str, Any] = 3072 snake_case_ :Any = 12 snake_case_ :Any = 1024 snake_case_ :str = 4096 snake_case_ :Union[str, Any] = 16 snake_case_ :Union[str, Any] = 24 snake_case_ :Tuple = 768 snake_case_ :Any = 3072 if model_name == "xclip-large-patch14-16-frames": snake_case_ :Any = 336 snake_case_ :Any = XCLIPConfig.from_text_vision_configs(_lowercase, _lowercase ) if "large" in model_name: snake_case_ :List[Any] = 768 return config def A_ ( _lowercase ): '''simple docstring''' if name == "token_embedding.weight": snake_case_ :Optional[Any] = name.replace("""token_embedding.weight""", """text_model.embeddings.token_embedding.weight""" ) if name == "positional_embedding": snake_case_ :Tuple = name.replace("""positional_embedding""", """text_model.embeddings.position_embedding.weight""" ) if "ln_1" in name: snake_case_ :Dict = name.replace("""ln_1""", """layer_norm1""" ) if "ln_2" in name: snake_case_ :str = name.replace("""ln_2""", """layer_norm2""" ) if "c_fc" in name: snake_case_ :str = name.replace("""c_fc""", """fc1""" ) if "c_proj" in name: snake_case_ :int = name.replace("""c_proj""", """fc2""" ) if name.startswith("""transformer.resblocks""" ): snake_case_ :Union[str, Any] = name.replace("""transformer.resblocks""", """text_model.encoder.layers""" ) if "attn.out_proj" in name and "message" not in name: snake_case_ :Union[str, Any] = name.replace("""attn.out_proj""", """self_attn.out_proj""" ) if "ln_final" in name: snake_case_ :Union[str, Any] = name.replace("""ln_final""", """text_model.final_layer_norm""" ) # visual encoder if name == "visual.class_embedding": snake_case_ :Any = name.replace("""visual.class_embedding""", """vision_model.embeddings.class_embedding""" ) if name == "visual.positional_embedding": snake_case_ :Optional[int] = name.replace("""visual.positional_embedding""", """vision_model.embeddings.position_embedding.weight""" ) if name.startswith("""visual.transformer.resblocks""" ): snake_case_ :Union[str, Any] = name.replace("""visual.transformer.resblocks""", """vision_model.encoder.layers""" ) if "visual.conv1" in name: snake_case_ :int = name.replace("""visual.conv1""", """vision_model.embeddings.patch_embedding""" ) if "visual.ln_pre" in name: snake_case_ :Any = name.replace("""visual.ln_pre""", """vision_model.pre_layernorm""" ) if "visual.ln_post" in name: snake_case_ :str = name.replace("""visual.ln_post""", """vision_model.post_layernorm""" ) if "visual.proj" in name: snake_case_ :Union[str, Any] = name.replace("""visual.proj""", """visual_projection.weight""" ) if "text_projection" in name: snake_case_ :Dict = name.replace("""text_projection""", """text_projection.weight""" ) # things on top if "prompts_visual_proj" in name: snake_case_ :List[str] = name.replace("""prompts_visual_proj""", """prompts_visual_projection""" ) if "prompts_visual_ln" in name: snake_case_ :Dict = name.replace("""prompts_visual_ln""", """prompts_visual_layernorm""" ) # mit if name == "mit.positional_embedding": snake_case_ :str = name.replace("""positional""", """position""" ) if name.startswith("""mit.resblocks""" ): snake_case_ :Dict = name.replace("""mit.resblocks""", """mit.encoder.layers""" ) # prompts generator if name.startswith("""prompts_generator.norm""" ): snake_case_ :Union[str, Any] = name.replace("""prompts_generator.norm""", """prompts_generator.layernorm""" ) return name def A_ ( _lowercase, _lowercase ): '''simple docstring''' for key in orig_state_dict.copy().keys(): snake_case_ :Dict = orig_state_dict.pop(_lowercase ) if "attn.in_proj" in key: snake_case_ :Optional[Any] = key.split(""".""" ) if key.startswith("""visual""" ): snake_case_ :Any = key_split[3] snake_case_ :Optional[Any] = config.vision_config.hidden_size if "message_attn" in key: if "weight" in key: snake_case_ :str = val[ :dim, : ] snake_case_ :Optional[int] = val[ dim : dim * 2, : ] snake_case_ :Union[str, Any] = val[ -dim:, : ] else: snake_case_ :Dict = val[ :dim ] snake_case_ :Optional[int] = val[ dim : dim * 2 ] snake_case_ :Optional[int] = val[ -dim: ] else: if "weight" in key: snake_case_ :Optional[Any] = val[ :dim, : ] snake_case_ :List[str] = val[ dim : dim * 2, : ] snake_case_ :Dict = val[ -dim:, : ] else: snake_case_ :Union[str, Any] = val[:dim] snake_case_ :Union[str, Any] = val[ dim : dim * 2 ] snake_case_ :Union[str, Any] = val[-dim:] elif key.startswith("""mit""" ): snake_case_ :Tuple = key_split[2] snake_case_ :Union[str, Any] = config.vision_config.mit_hidden_size if "weight" in key: snake_case_ :Optional[int] = val[:dim, :] snake_case_ :Optional[int] = val[dim : dim * 2, :] snake_case_ :str = val[-dim:, :] else: snake_case_ :str = val[:dim] snake_case_ :Any = val[dim : dim * 2] snake_case_ :int = val[-dim:] else: snake_case_ :Tuple = key_split[2] snake_case_ :Any = config.text_config.hidden_size if "weight" in key: snake_case_ :Dict = val[:dim, :] snake_case_ :Dict = val[ dim : dim * 2, : ] snake_case_ :List[str] = val[-dim:, :] else: snake_case_ :Any = val[:dim] snake_case_ :Tuple = val[ dim : dim * 2 ] snake_case_ :List[str] = val[-dim:] else: snake_case_ :Optional[int] = rename_key(_lowercase ) if new_key_name in ["visual_projection.weight", "text_projection.weight"]: snake_case_ :Optional[Any] = val.T snake_case_ :Tuple = val return orig_state_dict def A_ ( _lowercase ): '''simple docstring''' if num_frames == 8: snake_case_ :str = """eating_spaghetti_8_frames.npy""" elif num_frames == 16: snake_case_ :int = """eating_spaghetti.npy""" elif num_frames == 32: snake_case_ :List[str] = """eating_spaghetti_32_frames.npy""" snake_case_ :int = hf_hub_download( repo_id="""hf-internal-testing/spaghetti-video""", filename=_lowercase, repo_type="""dataset""", ) snake_case_ :Union[str, Any] = np.load(_lowercase ) return list(_lowercase ) def A_ ( _lowercase, _lowercase=None, _lowercase=False ): '''simple docstring''' snake_case_ :List[Any] = { # fully supervised kinetics-400 checkpoints """xclip-base-patch32""": """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_8.pth""", """xclip-base-patch32-16-frames""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_16.pth""" ), """xclip-base-patch16""": """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_8.pth""", """xclip-base-patch16-16-frames""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_16.pth""" ), """xclip-large-patch14""": """https://drive.google.com/u/0/uc?id=1NUOImq0o5DlQTST17iIP3vG7DgmHQuCx&amp;export=download&amp;confirm=t&amp;uuid=b26caedc-88e2-473e-830a-9d158b653cdb""", """xclip-large-patch14-16-frames""": """https://drive.google.com/u/0/uc?id=1FOYgnJc097OJ4lGwtRCCydQyVPJEOH7d&amp;export=download&amp;confirm=t&amp;uuid=538fa810-e671-4050-b385-9a623f89804f""", # fully supervised kinetics-600 checkpoints """xclip-base-patch16-kinetics-600""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_8.pth""" ), """xclip-base-patch16-kinetics-600-16-frames""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_16.pth""" ), """xclip-large-patch14-kinetics-600""": """https://drive.google.com/u/0/uc?id=1FV8C1INuM91sLAN4ImjzePLIlpMSihwV&amp;export=download&amp;confirm=t&amp;uuid=141d4977-4a65-44ae-864f-4b0c19f838be""", # few shot """xclip-base-patch16-hmdb-2-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_2.pth""" ), """xclip-base-patch16-hmdb-4-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_4.pth""" ), """xclip-base-patch16-hmdb-8-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_8.pth""" ), """xclip-base-patch16-hmdb-16-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_16.pth""" ), """xclip-base-patch16-ucf-2-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_2.pth""" ), """xclip-base-patch16-ucf-4-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_4.pth""" ), """xclip-base-patch16-ucf-8-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_8.pth""" ), """xclip-base-patch16-ucf-16-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_16.pth""" ), # zero shot """xclip-base-patch16-zero-shot""": """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/zero.pth""", } snake_case_ :Optional[int] = model_to_url[model_name] snake_case_ :int = 8 if "16-frames" in model_name: snake_case_ :List[Any] = 16 elif "shot" in model_name: snake_case_ :Dict = 32 snake_case_ :Optional[int] = get_xclip_config(_lowercase, _lowercase ) snake_case_ :Optional[Any] = XCLIPModel(_lowercase ) model.eval() if "drive" in checkpoint_url: snake_case_ :List[str] = """pytorch_model.bin""" gdown.cached_download(_lowercase, _lowercase, quiet=_lowercase ) snake_case_ :List[Any] = torch.load(_lowercase, map_location="""cpu""" )["""model"""] else: snake_case_ :Tuple = torch.hub.load_state_dict_from_url(_lowercase )["""model"""] snake_case_ :Union[str, Any] = convert_state_dict(_lowercase, _lowercase ) snake_case_ :str = XCLIPModel(_lowercase ) snake_case_, snake_case_ :Optional[int] = model.load_state_dict(_lowercase, strict=_lowercase ) assert missing_keys == ["text_model.embeddings.position_ids", "vision_model.embeddings.position_ids"] model.eval() snake_case_ :List[str] = 336 if model_name == """xclip-large-patch14-16-frames""" else 224 snake_case_ :List[Any] = VideoMAEImageProcessor(size=_lowercase ) snake_case_ :Any = CLIPTokenizer.from_pretrained("""openai/clip-vit-base-patch32""" ) snake_case_ :str = CLIPTokenizerFast.from_pretrained("""openai/clip-vit-base-patch32""" ) snake_case_ :Optional[Any] = XCLIPProcessor(image_processor=_lowercase, tokenizer=_lowercase ) snake_case_ :Optional[int] = prepare_video(_lowercase ) snake_case_ :Optional[Any] = processor( text=["""playing sports""", """eating spaghetti""", """go shopping"""], videos=_lowercase, return_tensors="""pt""", padding=_lowercase ) print("""Shape of pixel values:""", inputs.pixel_values.shape ) with torch.no_grad(): snake_case_ :List[Any] = model(**_lowercase ) # Verify outputs snake_case_ :List[Any] = outputs.logits_per_video snake_case_ :Any = logits_per_video.softmax(dim=1 ) print("""Probs:""", _lowercase ) # kinetics-400 if model_name == "xclip-base-patch32": snake_case_ :Union[str, Any] = torch.tensor([[0.0019, 0.9951, 0.0030]] ) elif model_name == "xclip-base-patch32-16-frames": snake_case_ :str = torch.tensor([[7.09_99e-04, 9.98_83e-01, 4.55_80e-04]] ) elif model_name == "xclip-base-patch16": snake_case_ :Tuple = torch.tensor([[0.0083, 0.9681, 0.0236]] ) elif model_name == "xclip-base-patch16-16-frames": snake_case_ :Any = torch.tensor([[7.69_37e-04, 9.97_28e-01, 1.94_73e-03]] ) elif model_name == "xclip-large-patch14": snake_case_ :str = torch.tensor([[0.0062, 0.9864, 0.0075]] ) elif model_name == "xclip-large-patch14-16-frames": snake_case_ :Tuple = torch.tensor([[3.38_77e-04, 9.99_37e-01, 2.88_88e-04]] ) # kinetics-600 elif model_name == "xclip-base-patch16-kinetics-600": snake_case_ :List[Any] = torch.tensor([[0.0555, 0.8914, 0.0531]] ) elif model_name == "xclip-base-patch16-kinetics-600-16-frames": snake_case_ :Union[str, Any] = torch.tensor([[3.85_54e-04, 9.99_29e-01, 3.27_54e-04]] ) elif model_name == "xclip-large-patch14-kinetics-600": snake_case_ :List[Any] = torch.tensor([[0.0036, 0.9920, 0.0045]] ) # few shot elif model_name == "xclip-base-patch16-hmdb-2-shot": snake_case_ :Dict = torch.tensor([[7.18_90e-06, 9.99_94e-01, 5.65_59e-05]] ) elif model_name == "xclip-base-patch16-hmdb-4-shot": snake_case_ :Union[str, Any] = torch.tensor([[1.03_20e-05, 9.99_93e-01, 6.24_35e-05]] ) elif model_name == "xclip-base-patch16-hmdb-8-shot": snake_case_ :str = torch.tensor([[4.13_77e-06, 9.99_90e-01, 9.83_86e-05]] ) elif model_name == "xclip-base-patch16-hmdb-16-shot": snake_case_ :str = torch.tensor([[4.13_47e-05, 9.99_62e-01, 3.34_11e-04]] ) elif model_name == "xclip-base-patch16-ucf-2-shot": snake_case_ :int = torch.tensor([[8.58_57e-05, 9.99_28e-01, 6.32_91e-04]] ) elif model_name == "xclip-base-patch16-ucf-4-shot": snake_case_ :Optional[int] = torch.tensor([[8.58_57e-05, 9.99_28e-01, 6.32_91e-04]] ) elif model_name == "xclip-base-patch16-ucf-8-shot": snake_case_ :Any = torch.tensor([[0.0027, 0.9904, 0.0070]] ) elif model_name == "xclip-base-patch16-ucf-16-shot": snake_case_ :Tuple = torch.tensor([[9.82_19e-04, 9.95_93e-01, 3.08_63e-03]] ) # zero shot elif model_name == "xclip-base-patch16-zero-shot": snake_case_ :Union[str, Any] = torch.tensor([[3.50_82e-04, 9.97_85e-01, 1.79_66e-03]] ) else: raise ValueError(f"""Model name {model_name} not supported""" ) assert torch.allclose(_lowercase, _lowercase, atol=1e-3 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowercase ) if push_to_hub: print("""Pushing model, processor and slow tokenizer files to the hub...""" ) model.push_to_hub(_lowercase, organization="""nielsr""" ) processor.push_to_hub(_lowercase, organization="""nielsr""" ) slow_tokenizer.push_to_hub(_lowercase, organization="""nielsr""" ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="xclip-base-patch32", type=str, help="Name of the model.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) __a = parser.parse_args() convert_xclip_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
1
import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType _A = logging.get_logger(__name__) class A ( lowercase__ ): __snake_case = '''vision-encoder-decoder''' __snake_case = True def __init__( self, **UpperCamelCase__ ): """simple docstring""" super().__init__(**_a ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f"A configuraton of type {self.model_type} cannot be instantiated because " f"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) lowerCAmelCase_ = kwargs.pop('''encoder''' ) lowerCAmelCase_ = encoder_config.pop('''model_type''' ) lowerCAmelCase_ = kwargs.pop('''decoder''' ) lowerCAmelCase_ = decoder_config.pop('''model_type''' ) lowerCAmelCase_ = AutoConfig.for_model(_a, **_a ) lowerCAmelCase_ = AutoConfig.for_model(_a, **_a ) lowerCAmelCase_ = True @classmethod def SCREAMING_SNAKE_CASE__ ( cls, UpperCamelCase__, UpperCamelCase__, **UpperCamelCase__ ): """simple docstring""" logger.info('''Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config''' ) lowerCAmelCase_ = True lowerCAmelCase_ = True return cls(encoder=encoder_config.to_dict(), decoder=decoder_config.to_dict(), **_a ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = copy.deepcopy(self.__dict__ ) lowerCAmelCase_ = self.encoder.to_dict() lowerCAmelCase_ = self.decoder.to_dict() lowerCAmelCase_ = self.__class__.model_type return output class A ( lowercase__ ): __snake_case = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" return 1E-4 @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" return OrderedDict({'''last_hidden_state''': {0: '''batch''', 1: '''encoder_sequence'''}} ) class A ( lowercase__ ): @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = OrderedDict() lowerCAmelCase_ = {0: 'batch', 1: 'past_decoder_sequence + sequence'} lowerCAmelCase_ = {0: 'batch', 1: 'past_decoder_sequence + sequence'} lowerCAmelCase_ = {0: 'batch', 1: 'encoder_sequence'} return common_inputs def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__ = -1, UpperCamelCase__ = -1, UpperCamelCase__ = False, UpperCamelCase__ = None, ): """simple docstring""" import torch lowerCAmelCase_ = OrderedDict() lowerCAmelCase_ = super().generate_dummy_inputs( _a, batch_size=_a, seq_length=_a, is_pair=_a, framework=_a ) lowerCAmelCase_ = dummy_input['input_ids'].shape lowerCAmelCase_ = (batch, encoder_sequence, self._config.encoder_hidden_size) lowerCAmelCase_ = dummy_input.pop('''input_ids''' ) lowerCAmelCase_ = dummy_input.pop('''attention_mask''' ) lowerCAmelCase_ = torch.zeros(_a ) return common_inputs class A ( lowercase__ ): @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" pass def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__ ): """simple docstring""" return VisionEncoderDecoderEncoderOnnxConfig(_a ) def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__, UpperCamelCase__ = "default" ): """simple docstring""" lowerCAmelCase_ = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(_a, _a )
355
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal _A = logging.get_logger(__name__) _A = TypeVar('''DatasetType''', Dataset, IterableDataset) def __UpperCamelCase ( _A , _A = None , _A = None , _A = None , _A = None , _A = "first_exhausted" , ): from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError('''Unable to interleave an empty list of datasets.''' ) for i, dataset in enumerate(_A ): if not isinstance(_A , (Dataset, IterableDataset) ): if isinstance(_A , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " '''is an empty dataset dictionary.''' ) raise ValueError( f"Dataset at position {i} has at least one split: {list(_A )}\n" f"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(_A ) )}']" ) raise ValueError( f"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(_A ).__name__}." ) if i == 0: lowerCAmelCase_ , lowerCAmelCase_ = ( (Dataset, IterableDataset) if isinstance(_A , _A ) else (IterableDataset, Dataset) ) elif not isinstance(_A , _A ): raise ValueError( f"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(f"{stopping_strategy} is not supported. Please enter a valid stopping_strategy." ) if dataset_type is Dataset: return _interleave_map_style_datasets( _A , _A , _A , info=_A , split=_A , stopping_strategy=_A ) else: return _interleave_iterable_datasets( _A , _A , _A , info=_A , split=_A , stopping_strategy=_A ) def __UpperCamelCase ( _A , _A = None , _A = None , _A = 0 , ): if not dsets: raise ValueError('''Unable to concatenate an empty list of datasets.''' ) for i, dataset in enumerate(_A ): if not isinstance(_A , (Dataset, IterableDataset) ): if isinstance(_A , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " '''is an empty dataset dictionary.''' ) raise ValueError( f"Dataset at position {i} has at least one split: {list(_A )}\n" f"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(_A ) )}']" ) raise ValueError( f"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(_A ).__name__}." ) if i == 0: lowerCAmelCase_ , lowerCAmelCase_ = ( (Dataset, IterableDataset) if isinstance(_A , _A ) else (IterableDataset, Dataset) ) elif not isinstance(_A , _A ): raise ValueError( f"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if dataset_type is Dataset: return _concatenate_map_style_datasets(_A , info=_A , split=_A , axis=_A ) else: return _concatenate_iterable_datasets(_A , info=_A , split=_A , axis=_A )
167
0
import os import pytest from transformers.dynamic_module_utils import get_imports snake_case : Optional[int] = ''' import os ''' snake_case : Optional[Any] = ''' def foo(): import os return False ''' snake_case : str = ''' def foo(): def bar(): if True: import os return False return bar() ''' snake_case : Optional[Any] = ''' import os try: import bar except ImportError: raise ValueError() ''' snake_case : Optional[Any] = ''' import os def foo(): try: import bar except ImportError: raise ValueError() ''' snake_case : int = ''' import os try: import bar except (ImportError, AttributeError): raise ValueError() ''' snake_case : int = ''' import os try: import bar except ImportError as e: raise ValueError() ''' snake_case : Union[str, Any] = ''' import os try: import bar except: raise ValueError() ''' snake_case : Union[str, Any] = ''' import os try: import bar import baz except ImportError: raise ValueError() ''' snake_case : Union[str, Any] = ''' import os try: import bar import baz except ImportError: x = 1 raise ValueError() ''' snake_case : Any = [ TOP_LEVEL_IMPORT, IMPORT_IN_FUNCTION, DEEPLY_NESTED_IMPORT, TOP_LEVEL_TRY_IMPORT, GENERIC_EXCEPT_IMPORT, MULTILINE_TRY_IMPORT, MULTILINE_BOTH_IMPORT, MULTIPLE_EXCEPTS_IMPORT, EXCEPT_AS_IMPORT, TRY_IMPORT_IN_FUNCTION, ] @pytest.mark.parametrize("case" , _A ) def lowerCAmelCase_ ( _snake_case : int , _snake_case : int ) -> List[str]: '''simple docstring''' __magic_name__ : List[Any] = os.path.join(_A , "test_file.py" ) with open(_A , "w" ) as _tmp_file: _tmp_file.write(_A ) __magic_name__ : str = get_imports(_A ) assert parsed_imports == ["os"]
281
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Union[str, Any] = { '''unc-nlp/lxmert-base-uncased''': '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json''', } class UpperCAmelCase__ ( A__ ): """simple docstring""" a = "lxmert" a = {} def __init__( self : Union[str, Any] , __lowerCamelCase : List[str]=3_0522 , __lowerCamelCase : Union[str, Any]=768 , __lowerCamelCase : Dict=12 , __lowerCamelCase : Union[str, Any]=9500 , __lowerCamelCase : Union[str, Any]=1600 , __lowerCamelCase : Any=400 , __lowerCamelCase : List[str]=3072 , __lowerCamelCase : List[str]="gelu" , __lowerCamelCase : Optional[Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Optional[Any]=512 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Any=0.02 , __lowerCamelCase : Any=1e-12 , __lowerCamelCase : List[Any]=9 , __lowerCamelCase : Any=5 , __lowerCamelCase : List[str]=5 , __lowerCamelCase : Optional[Any]=2048 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : List[str]=6.67 , __lowerCamelCase : Dict=True , __lowerCamelCase : Any=True , __lowerCamelCase : Any=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Any=True , **__lowerCamelCase : Optional[Any] , ) -> Any: SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = num_qa_labels SCREAMING_SNAKE_CASE__ = num_object_labels SCREAMING_SNAKE_CASE__ = num_attr_labels SCREAMING_SNAKE_CASE__ = l_layers SCREAMING_SNAKE_CASE__ = x_layers SCREAMING_SNAKE_CASE__ = r_layers SCREAMING_SNAKE_CASE__ = visual_feat_dim SCREAMING_SNAKE_CASE__ = visual_pos_dim SCREAMING_SNAKE_CASE__ = visual_loss_normalizer SCREAMING_SNAKE_CASE__ = task_matched SCREAMING_SNAKE_CASE__ = task_mask_lm SCREAMING_SNAKE_CASE__ = task_obj_predict SCREAMING_SNAKE_CASE__ = task_qa SCREAMING_SNAKE_CASE__ = visual_obj_loss SCREAMING_SNAKE_CASE__ = visual_attr_loss SCREAMING_SNAKE_CASE__ = visual_feat_loss SCREAMING_SNAKE_CASE__ = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**__lowerCamelCase )
314
0
import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Optional[Any]: # Construct model if openai_config_file == "": _snake_case = OpenAIGPTConfig() else: _snake_case = OpenAIGPTConfig.from_json_file(lowerCamelCase__ ) _snake_case = OpenAIGPTModel(lowerCamelCase__ ) # Load weights from numpy load_tf_weights_in_openai_gpt(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # Save pytorch-model _snake_case = pytorch_dump_folder_path + '/' + WEIGHTS_NAME _snake_case = pytorch_dump_folder_path + '/' + CONFIG_NAME print(F'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(model.state_dict() , lowerCamelCase__ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(lowerCamelCase__ , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowercase : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--openai_checkpoint_folder_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--openai_config_file", default="", type=str, help=( "An optional config json file corresponding to the pre-trained OpenAI model. \n" "This specifies the model architecture." ), ) lowercase : str = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
353
'''simple docstring''' from __future__ import annotations from collections import namedtuple def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> tuple: _snake_case = namedtuple('result' , 'name value' ) if (voltage, current, power).count(0 ) != 1: raise ValueError('Only one argument must be 0' ) elif power < 0: raise ValueError( 'Power cannot be negative in any electrical/electronics system' ) elif voltage == 0: return result('voltage' , power / current ) elif current == 0: return result('current' , power / voltage ) elif power == 0: return result('power' , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError('Exactly one argument must be 0' ) if __name__ == "__main__": import doctest doctest.testmod()
160
0
import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin __A =get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __A =2_5_6_0_4_7 __A =2_5_6_1_4_5 @require_sentencepiece @require_tokenizers class _SCREAMING_SNAKE_CASE ( snake_case_ , unittest.TestCase ): lowerCAmelCase__ = NllbTokenizer lowerCAmelCase__ = NllbTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True lowerCAmelCase__ = {} def SCREAMING_SNAKE_CASE_( self ) -> Dict: super().setUp() # We have a SentencePiece fixture for testing lowerCamelCase_ = NllbTokenizer(lowercase , keep_accents=lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = NllbTokenizer(lowercase , keep_accents=lowercase ) lowerCamelCase_ = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowercase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) lowerCamelCase_ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowercase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) lowerCamelCase_ = tokenizer.convert_tokens_to_ids(lowercase ) self.assertListEqual( lowercase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) lowerCamelCase_ = tokenizer.convert_ids_to_tokens(lowercase ) self.assertListEqual( lowercase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-nllb", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})' ): lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(lowercase , **lowercase ) lowerCamelCase_ = self.tokenizer_class.from_pretrained(lowercase , **lowercase ) lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = tokenizer_r.save_pretrained(lowercase ) lowerCamelCase_ = tokenizer_p.save_pretrained(lowercase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) lowerCamelCase_ = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f ) self.assertSequenceEqual(lowercase , lowercase ) # Checks everything loads correctly in the same way lowerCamelCase_ = tokenizer_r.from_pretrained(lowercase ) lowerCamelCase_ = tokenizer_p.from_pretrained(lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowercase , lowercase ) ) shutil.rmtree(lowercase ) # Save tokenizer rust, legacy_format=True lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = tokenizer_r.save_pretrained(lowercase , legacy_format=lowercase ) lowerCamelCase_ = tokenizer_p.save_pretrained(lowercase ) # Checks it save with the same files self.assertSequenceEqual(lowercase , lowercase ) # Checks everything loads correctly in the same way lowerCamelCase_ = tokenizer_r.from_pretrained(lowercase ) lowerCamelCase_ = tokenizer_p.from_pretrained(lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowercase , lowercase ) ) shutil.rmtree(lowercase ) # Save tokenizer rust, legacy_format=False lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = tokenizer_r.save_pretrained(lowercase , legacy_format=lowercase ) lowerCamelCase_ = tokenizer_p.save_pretrained(lowercase ) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCamelCase_ = tokenizer_r.from_pretrained(lowercase ) lowerCamelCase_ = tokenizer_p.from_pretrained(lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowercase , lowercase ) ) shutil.rmtree(lowercase ) @require_torch def SCREAMING_SNAKE_CASE_( self ) -> str: if not self.test_seqaseq: return lowerCamelCase_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'{tokenizer.__class__.__name__}' ): # Longer text that will definitely require truncation. lowerCamelCase_ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for" " Syria is that 'there is no military solution' to the nearly five-year conflict and more weapons" " will only worsen the violence and misery for millions of people.", ] lowerCamelCase_ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al" " Rusiei pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi" " că noi arme nu vor face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] try: lowerCamelCase_ = tokenizer.prepare_seqaseq_batch( src_texts=lowercase , tgt_texts=lowercase , max_length=3 , max_target_length=10 , return_tensors="pt" , src_lang="eng_Latn" , tgt_lang="ron_Latn" , ) except NotImplementedError: return self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 10 ) # max_target_length will default to max_length if not specified lowerCamelCase_ = tokenizer.prepare_seqaseq_batch( lowercase , tgt_texts=lowercase , max_length=3 , return_tensors="pt" ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 3 ) lowerCamelCase_ = tokenizer.prepare_seqaseq_batch( src_texts=lowercase , max_length=3 , max_target_length=10 , return_tensors="pt" ) self.assertEqual(batch_encoder_only.input_ids.shape[1] , 3 ) self.assertEqual(batch_encoder_only.attention_mask.shape[1] , 3 ) self.assertNotIn("decoder_input_ids" , lowercase ) @unittest.skip("Unfortunately way too slow to build a BPE with SentencePiece." ) def SCREAMING_SNAKE_CASE_( self ) -> Any: pass def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})' ): lowerCamelCase_ = [AddedToken("<special>" , lstrip=lowercase )] lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained( lowercase , additional_special_tokens=lowercase , **lowercase ) lowerCamelCase_ = tokenizer_r.encode("Hey this is a <special> token" ) lowerCamelCase_ = tokenizer_r.encode("<special>" , add_special_tokens=lowercase )[0] self.assertTrue(special_token_id in r_output ) if self.test_slow_tokenizer: lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained( lowercase , additional_special_tokens=lowercase , **lowercase , ) lowerCamelCase_ = self.tokenizer_class.from_pretrained( lowercase , additional_special_tokens=lowercase , **lowercase ) lowerCamelCase_ = tokenizer_p.encode("Hey this is a <special> token" ) lowerCamelCase_ = tokenizer_cr.encode("Hey this is a <special> token" ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) self.assertTrue(special_token_id in p_output ) self.assertTrue(special_token_id in cr_output ) @require_torch @require_sentencepiece @require_tokenizers class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): lowerCAmelCase__ = 'facebook/nllb-200-distilled-600M' lowerCAmelCase__ = [ ' UN Chief Says There Is No Military Solution in Syria', ' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.', ] lowerCAmelCase__ = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', 'Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei' ' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor' ' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.', ] lowerCAmelCase__ = [ 25_60_47, 1_62_97, 13_44_08, 81_65, 24_80_66, 1_47_34, 9_50, 11_35, 10_57_21, 35_73, 83, 2_73_52, 1_08, 4_94_86, 2, ] @classmethod def SCREAMING_SNAKE_CASE_( cls ) -> int: lowerCamelCase_ = NllbTokenizer.from_pretrained( cls.checkpoint_name , src_lang="eng_Latn" , tgt_lang="ron_Latn" ) lowerCamelCase_ = 1 return cls def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ace_Arab"] , 256001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ace_Latn"] , 256002 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["fra_Latn"] , 256057 ) def SCREAMING_SNAKE_CASE_( self ) -> Any: lowerCamelCase_ = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: self.assertIn(lowercase , self.tokenizer.all_special_ids ) # fmt: off lowerCamelCase_ = [RO_CODE, 4254, 98068, 112923, 39072, 3909, 713, 102767, 26, 17314, 35642, 14683, 33118, 2022, 66987, 2, 256047] # fmt: on lowerCamelCase_ = self.tokenizer.decode(lowercase , skip_special_tokens=lowercase ) lowerCamelCase_ = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=lowercase ) self.assertEqual(lowercase , lowercase ) self.assertNotIn(self.tokenizer.eos_token , lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Any: lowerCamelCase_ = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , lowercase ) lowerCamelCase_ = 10 lowerCamelCase_ = self.tokenizer(lowercase , max_length=lowercase , truncation=lowercase ).input_ids[0] self.assertEqual(ids[-1] , 2 ) self.assertEqual(ids[0] , lowercase ) self.assertEqual(len(lowercase ) , lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Any: self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"] ) , [256203, 3] ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(lowercase ) lowerCamelCase_ = NllbTokenizer.from_pretrained(lowercase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , lowercase ) @require_torch def SCREAMING_SNAKE_CASE_( self ) -> int: lowerCamelCase_ = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=lowercase , truncation=lowercase , max_length=len(self.expected_src_tokens ) , return_tensors="pt" , ) lowerCamelCase_ = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.lang_code_to_id["ron_Latn"] ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual((2, 15) , batch.input_ids.shape ) self.assertEqual((2, 15) , batch.attention_mask.shape ) lowerCamelCase_ = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , lowercase ) self.assertEqual(lowercase , batch.decoder_input_ids[0, 0] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def SCREAMING_SNAKE_CASE_( self ) -> int: lowerCamelCase_ = self.tokenizer(self.src_text , padding=lowercase , truncation=lowercase , max_length=3 , return_tensors="pt" ) lowerCamelCase_ = self.tokenizer( text_target=self.tgt_text , padding=lowercase , truncation=lowercase , max_length=10 , return_tensors="pt" ) lowerCamelCase_ = targets["input_ids"] lowerCamelCase_ = shift_tokens_right( lowercase , self.tokenizer.pad_token_id , decoder_start_token_id=self.tokenizer.lang_code_to_id[self.tokenizer.tgt_lang] , ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="eng_Latn" , tgt_lang="fra_Latn" ) self.assertEqual( nested_simplify(lowercase ) , { # A, test, EOS, en_XX "input_ids": [[256047, 70, 7356, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 256057, } , ) @require_torch def SCREAMING_SNAKE_CASE_( self ) -> Tuple: lowerCamelCase_ = True lowerCamelCase_ = self.tokenizer( "UN Chief says there is no military solution in Syria" , src_lang="eng_Latn" , tgt_lang="fra_Latn" ) self.assertEqual( inputs.input_ids , [16297, 134408, 25653, 6370, 248, 254, 103929, 94995, 108, 49486, 2, 256047] ) lowerCamelCase_ = False lowerCamelCase_ = self.tokenizer( "UN Chief says there is no military solution in Syria" , src_lang="eng_Latn" , tgt_lang="fra_Latn" ) self.assertEqual( inputs.input_ids , [256047, 16297, 134408, 25653, 6370, 248, 254, 103929, 94995, 108, 49486, 2] )
19
'''simple docstring''' from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder lowercase_ = datasets.utils.logging.get_logger(__name__) class __A ( folder_based_builder.FolderBasedBuilderConfig ): '''simple docstring''' __lowerCamelCase : bool = None __lowerCamelCase : bool = None class __A ( folder_based_builder.FolderBasedBuilder ): '''simple docstring''' __lowerCamelCase : int = datasets.Audio() __lowerCamelCase : str = 'audio' __lowerCamelCase : Optional[Any] = AudioFolderConfig __lowerCamelCase : List[str] # definition at the bottom of the script __lowerCamelCase : Union[str, Any] = AudioClassification(audio_column='audio' , label_column='label' ) lowercase_ = [ ".aiff", ".au", ".avr", ".caf", ".flac", ".htk", ".svx", ".mat4", ".mat5", ".mpc2k", ".ogg", ".paf", ".pvf", ".raw", ".rf64", ".sd2", ".sds", ".ircam", ".voc", ".w64", ".wav", ".nist", ".wavex", ".wve", ".xi", ".mp3", ".opus", ] lowercase_ = AUDIO_EXTENSIONS
211
0
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=lowerCamelCase__ ) class SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ): """simple docstring""" lowercase : str = field(default='audio-classification' , metadata={'include_in_asdict_even_if_is_default': True} ) lowercase : ClassVar[Features] = Features({'audio': Audio()} ) lowercase : ClassVar[Features] = Features({'labels': ClassLabel} ) lowercase : str = "audio" lowercase : str = "labels" def __lowerCamelCase ( self , __UpperCamelCase ) -> Tuple: '''simple docstring''' if self.label_column not in features: raise ValueError(f'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] , __UpperCamelCase ): raise ValueError(f'''Column {self.label_column} is not a ClassLabel.''' ) __UpperCamelCase : Dict = copy.deepcopy(self ) __UpperCamelCase : Tuple = self.label_schema.copy() __UpperCamelCase : Any = features[self.label_column] __UpperCamelCase : List[str] = label_schema return task_template @property def __lowerCamelCase ( self ) -> Dict[str, str]: '''simple docstring''' return { self.audio_column: "audio", self.label_column: "labels", }
351
import warnings from ...utils import logging from .image_processing_deformable_detr import DeformableDetrImageProcessor lowercase : Optional[Any] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ): """simple docstring""" def __init__( self , *__UpperCamelCase , **__UpperCamelCase ) -> None: '''simple docstring''' warnings.warn( "The class DeformableDetrFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use DeformableDetrImageProcessor instead." , __UpperCamelCase , ) super().__init__(*__UpperCamelCase , **__UpperCamelCase )
171
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase_ = {'configuration_plbart': ['PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PLBartConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['PLBartTokenizer'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ 'PLBART_PRETRAINED_MODEL_ARCHIVE_LIST', 'PLBartForCausalLM', 'PLBartForConditionalGeneration', 'PLBartForSequenceClassification', 'PLBartModel', 'PLBartPreTrainedModel', ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure)
16
"""simple docstring""" def lowercase ( __snake_case : Optional[int] ): lowercase_ : int = 0 lowercase_ : Optional[Any] = len(__snake_case ) for i in range(n - 1 ): for j in range(i + 1 , __snake_case ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def lowercase ( __snake_case : str ): if len(__snake_case ) <= 1: return arr, 0 lowercase_ : Optional[Any] = len(__snake_case ) // 2 lowercase_ : List[Any] = arr[0:mid] lowercase_ : Union[str, Any] = arr[mid:] lowercase_ , lowercase_ : Tuple = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = _count_cross_inversions(__snake_case , __snake_case ) lowercase_ : List[Any] = inversion_p + inversions_q + cross_inversions return c, num_inversions def lowercase ( __snake_case : str , __snake_case : Optional[int] ): lowercase_ : Optional[Any] = [] lowercase_ : Any = 0 while i < len(__snake_case ) and j < len(__snake_case ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(__snake_case ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(__snake_case ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def lowercase ( ): lowercase_ : Union[str, Any] = [1_0, 2, 1, 5, 5, 2, 1_1] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) lowercase_ : int = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 8 print('''number of inversions = ''' , __snake_case ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() lowercase_ : Dict = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) # an empty list should also have zero inversions lowercase_ : List[Any] = [] lowercase_ : Any = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : List[str] = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) if __name__ == "__main__": main()
33
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class _a : _lowercase : Optional[Any] = PegasusConfig _lowercase : int = {} _lowercase : List[str] = '''gelu''' def __init__( self: List[str] , UpperCamelCase_: Dict , UpperCamelCase_: List[str]=13 , UpperCamelCase_: Union[str, Any]=7 , UpperCamelCase_: str=True , UpperCamelCase_: List[Any]=False , UpperCamelCase_: List[str]=99 , UpperCamelCase_: List[Any]=32 , UpperCamelCase_: Any=2 , UpperCamelCase_: Optional[Any]=4 , UpperCamelCase_: Optional[int]=37 , UpperCamelCase_: List[str]=0.1 , UpperCamelCase_: Optional[Any]=0.1 , UpperCamelCase_: List[str]=40 , UpperCamelCase_: List[str]=2 , UpperCamelCase_: Optional[int]=1 , UpperCamelCase_: Optional[int]=0 , ) -> Dict: """simple docstring""" lowercase__ = parent lowercase__ = batch_size lowercase__ = seq_length lowercase__ = is_training lowercase__ = use_labels lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = eos_token_id lowercase__ = pad_token_id lowercase__ = bos_token_id def lowerCamelCase_ ( self: Tuple ) -> int: """simple docstring""" lowercase__ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowercase__ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowercase__ = tf.concat([input_ids, eos_tensor] , axis=1 ) lowercase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowercase__ = prepare_pegasus_inputs_dict(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) return config, inputs_dict def lowerCamelCase_ ( self: int , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Union[str, Any] ) -> Optional[int]: """simple docstring""" lowercase__ = TFPegasusModel(config=UpperCamelCase_ ).get_decoder() lowercase__ = inputs_dict["input_ids"] lowercase__ = input_ids[:1, :] lowercase__ = inputs_dict["attention_mask"][:1, :] lowercase__ = inputs_dict["head_mask"] lowercase__ = 1 # first forward pass lowercase__ = model(UpperCamelCase_ , attention_mask=UpperCamelCase_ , head_mask=UpperCamelCase_ , use_cache=UpperCamelCase_ ) lowercase__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowercase__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowercase__ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and lowercase__ = tf.concat([input_ids, next_tokens] , axis=-1 ) lowercase__ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) lowercase__ = model(UpperCamelCase_ , attention_mask=UpperCamelCase_ )[0] lowercase__ = model(UpperCamelCase_ , attention_mask=UpperCamelCase_ , past_key_values=UpperCamelCase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice lowercase__ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) lowercase__ = output_from_no_past[:, -3:, random_slice_idx] lowercase__ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(UpperCamelCase_ , UpperCamelCase_ , rtol=1E-3 ) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , ): """simple docstring""" if attention_mask is None: lowercase__ = tf.cast(tf.math.not_equal(_lowerCamelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowercase__ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowercase__ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowercase__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowercase__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _a ( a__ , a__ , unittest.TestCase ): _lowercase : str = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () _lowercase : Optional[int] = (TFPegasusForConditionalGeneration,) if is_tf_available() else () _lowercase : int = ( { '''conversational''': TFPegasusForConditionalGeneration, '''feature-extraction''': TFPegasusModel, '''summarization''': TFPegasusForConditionalGeneration, '''text2text-generation''': TFPegasusForConditionalGeneration, '''translation''': TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) _lowercase : Optional[int] = True _lowercase : Dict = False _lowercase : Optional[int] = False def lowerCamelCase_ ( self: str ) -> Any: """simple docstring""" lowercase__ = TFPegasusModelTester(self ) lowercase__ = ConfigTester(self , config_class=UpperCamelCase_ ) def lowerCamelCase_ ( self: Optional[int] ) -> str: """simple docstring""" self.config_tester.run_common_tests() def lowerCamelCase_ ( self: str ) -> List[str]: """simple docstring""" lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*UpperCamelCase_ ) @require_sentencepiece @require_tokenizers @require_tf class _a ( unittest.TestCase ): _lowercase : Dict = [ ''' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.''', ''' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning \'Oh I think you\'re nominated\'\", said Dappy.\"And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around.\"At the end of the day we\'re grateful to be where we are in our careers.\"If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" ''', ] _lowercase : Tuple = [ '''California\'s largest electricity provider has cut power to hundreds of thousands of customers in an effort to''' ''' reduce the risk of wildfires.''', '''N-Dubz have revealed they\'re \"grateful\" to have been nominated for four Mobo Awards.''', ] # differs slightly from pytorch, likely due to numerical differences in linear layers _lowercase : List[str] = '''google/pegasus-xsum''' @cached_property def lowerCamelCase_ ( self: Optional[int] ) -> Optional[Any]: """simple docstring""" return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def lowerCamelCase_ ( self: Union[str, Any] ) -> Any: """simple docstring""" lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def lowerCamelCase_ ( self: Optional[int] , **UpperCamelCase_: Any ) -> Union[str, Any]: """simple docstring""" lowercase__ = self.translate_src_text(**UpperCamelCase_ ) assert self.expected_text == generated_words def lowerCamelCase_ ( self: List[Any] , **UpperCamelCase_: Tuple ) -> List[str]: """simple docstring""" lowercase__ = self.tokenizer(self.src_text , **UpperCamelCase_ , padding=UpperCamelCase_ , return_tensors='''tf''' ) lowercase__ = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=UpperCamelCase_ , ) lowercase__ = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=UpperCamelCase_ ) return generated_words @slow def lowerCamelCase_ ( self: Dict ) -> Union[str, Any]: """simple docstring""" self._assert_generated_batch_equal_expected()
370
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class _a ( unittest.TestCase ): def lowerCamelCase_ ( self: Dict ) -> Any: """simple docstring""" lowercase__ = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) lowercase__ = get_activation('''gelu''' ) self.assertTrue(torch.allclose(gelu_python(UpperCamelCase_ ) , torch_builtin(UpperCamelCase_ ) ) ) self.assertFalse(torch.allclose(gelu_python(UpperCamelCase_ ) , gelu_new(UpperCamelCase_ ) ) ) def lowerCamelCase_ ( self: Optional[int] ) -> str: """simple docstring""" lowercase__ = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) lowercase__ = get_activation('''gelu''' ) lowercase__ = get_activation('''gelu_10''' ) lowercase__ = torch_builtin(UpperCamelCase_ ) lowercase__ = geluaa(UpperCamelCase_ ) lowercase__ = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(UpperCamelCase_ ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def lowerCamelCase_ ( self: Union[str, Any] ) -> Tuple: """simple docstring""" get_activation('''gelu''' ) get_activation('''gelu_10''' ) get_activation('''gelu_fast''' ) get_activation('''gelu_new''' ) get_activation('''gelu_python''' ) get_activation('''gelu_pytorch_tanh''' ) get_activation('''linear''' ) get_activation('''mish''' ) get_activation('''quick_gelu''' ) get_activation('''relu''' ) get_activation('''sigmoid''' ) get_activation('''silu''' ) get_activation('''swish''' ) get_activation('''tanh''' ) with self.assertRaises(UpperCamelCase_ ): get_activation('''bogus''' ) with self.assertRaises(UpperCamelCase_ ): get_activation(UpperCamelCase_ ) def lowerCamelCase_ ( self: Tuple ) -> int: """simple docstring""" lowercase__ = get_activation('''gelu''' ) lowercase__ = 1 lowercase__ = get_activation('''gelu''' ) self.assertEqual(acta.a , 1 ) with self.assertRaises(UpperCamelCase_ ): lowercase__ = acta.a
93
0
'''simple docstring''' import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort UpperCamelCase__: List[str] = logging.get_logger(__name__) UpperCamelCase__: Any = { "tensor(bool)": np.bool_, "tensor(int8)": np.inta, "tensor(uint8)": np.uinta, "tensor(int16)": np.intaa, "tensor(uint16)": np.uintaa, "tensor(int32)": np.intaa, "tensor(uint32)": np.uintaa, "tensor(int64)": np.intaa, "tensor(uint64)": np.uintaa, "tensor(float16)": np.floataa, "tensor(float)": np.floataa, "tensor(double)": np.floataa, } class SCREAMING_SNAKE_CASE: """simple docstring""" def __init__( self : Optional[int] , __snake_case : Optional[int]=None , **__snake_case : str ) -> Union[str, Any]: logger.info('''`diffusers.OnnxRuntimeModel` is experimental and might change in the future.''' ) UpperCAmelCase : Union[str, Any] = model UpperCAmelCase : Optional[Any] = kwargs.get('''model_save_dir''' , __snake_case ) UpperCAmelCase : Optional[int] = kwargs.get('''latest_model_name''' , __snake_case ) def __call__( self : Optional[Any] , **__snake_case : Optional[Any] ) -> List[Any]: UpperCAmelCase : Optional[int] = {k: np.array(__snake_case ) for k, v in kwargs.items()} return self.model.run(__snake_case , __snake_case ) @staticmethod def A ( __snake_case : Union[str, Path] , __snake_case : List[str]=None , __snake_case : Optional[int]=None ) -> Optional[int]: if provider is None: logger.info('''No onnxruntime provider specified, using CPUExecutionProvider''' ) UpperCAmelCase : List[Any] = '''CPUExecutionProvider''' return ort.InferenceSession(__snake_case , providers=[provider] , sess_options=__snake_case ) def A ( self : Optional[Any] , __snake_case : Union[str, Path] , __snake_case : Optional[str] = None , **__snake_case : int ) -> str: UpperCAmelCase : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME UpperCAmelCase : Optional[int] = self.model_save_dir.joinpath(self.latest_model_name ) UpperCAmelCase : int = Path(__snake_case ).joinpath(__snake_case ) try: shutil.copyfile(__snake_case , __snake_case ) except shutil.SameFileError: pass # copy external weights (for models >2GB) UpperCAmelCase : Tuple = self.model_save_dir.joinpath(__snake_case ) if src_path.exists(): UpperCAmelCase : List[Any] = Path(__snake_case ).joinpath(__snake_case ) try: shutil.copyfile(__snake_case , __snake_case ) except shutil.SameFileError: pass def A ( self : Optional[int] , __snake_case : Union[str, os.PathLike] , **__snake_case : int , ) -> List[Any]: if os.path.isfile(__snake_case ): logger.error(F"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(__snake_case , exist_ok=__snake_case ) # saving model weights/files self._save_pretrained(__snake_case , **__snake_case ) @classmethod def A ( cls : Optional[int] , __snake_case : Union[str, Path] , __snake_case : Optional[Union[bool, str, None]] = None , __snake_case : Optional[Union[str, None]] = None , __snake_case : bool = False , __snake_case : Optional[str] = None , __snake_case : Optional[str] = None , __snake_case : Optional[str] = None , __snake_case : Optional["ort.SessionOptions"] = None , **__snake_case : Any , ) -> List[str]: UpperCAmelCase : Dict = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(__snake_case ): UpperCAmelCase : Optional[Any] = OnnxRuntimeModel.load_model( os.path.join(__snake_case , __snake_case ) , provider=__snake_case , sess_options=__snake_case ) UpperCAmelCase : List[Any] = Path(__snake_case ) # load model from hub else: # download model UpperCAmelCase : List[str] = hf_hub_download( repo_id=__snake_case , filename=__snake_case , use_auth_token=__snake_case , revision=__snake_case , cache_dir=__snake_case , force_download=__snake_case , ) UpperCAmelCase : int = Path(__snake_case ).parent UpperCAmelCase : Union[str, Any] = Path(__snake_case ).name UpperCAmelCase : Optional[Any] = OnnxRuntimeModel.load_model(__snake_case , provider=__snake_case , sess_options=__snake_case ) return cls(model=__snake_case , **__snake_case ) @classmethod def A ( cls : Any , __snake_case : Union[str, Path] , __snake_case : bool = True , __snake_case : Optional[str] = None , __snake_case : Optional[str] = None , **__snake_case : Optional[int] , ) -> str: UpperCAmelCase : str = None if len(str(__snake_case ).split('''@''' ) ) == 2: UpperCAmelCase , UpperCAmelCase : Optional[Any] = model_id.split('''@''' ) return cls._from_pretrained( model_id=__snake_case , revision=__snake_case , cache_dir=__snake_case , force_download=__snake_case , use_auth_token=__snake_case , **__snake_case , )
23
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import Callable, Dict, List, Tuple import timm import torch import torch.nn as nn from classy_vision.models.regnet import RegNet, RegNetParams, RegNetYaagf, RegNetYaagf, RegNetYaaagf from huggingface_hub import cached_download, hf_hub_url from torch import Tensor from vissl.models.model_helpers import get_trunk_forward_outputs from transformers import AutoImageProcessor, RegNetConfig, RegNetForImageClassification, RegNetModel from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase =logging.get_logger() @dataclass class a_ : """simple docstring""" __UpperCAmelCase = 42 __UpperCAmelCase = field(default_factory=lowerCamelCase_ ) __UpperCAmelCase = field(default_factory=lowerCamelCase_ ) def _lowerCAmelCase ( self : List[Any] ,snake_case : Dict ,snake_case : Tensor ,snake_case : Tensor ): SCREAMING_SNAKE_CASE =len(list(m.modules() ) ) == 1 or isinstance(snake_case ,nn.Convad ) or isinstance(snake_case ,nn.BatchNormad ) if has_not_submodules: self.traced.append(snake_case ) def __call__( self : List[str] ,snake_case : Tensor ): for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(snake_case ) [x.remove() for x in self.handles] return self @property def _lowerCAmelCase ( self : Optional[Any] ): # check the len of the state_dict keys to see if we have learnable params return list(filter(lambda snake_case : len(list(x.state_dict().keys() ) ) > 0 ,self.traced ) ) @dataclass class a_ : """simple docstring""" __UpperCAmelCase = 42 __UpperCAmelCase = 42 __UpperCAmelCase = 1 __UpperCAmelCase = field(default_factory=lowerCamelCase_ ) __UpperCAmelCase = field(default_factory=lowerCamelCase_ ) __UpperCAmelCase = True def __call__( self : str ,snake_case : Tensor ): SCREAMING_SNAKE_CASE =Tracker(self.dest )(snake_case ).parametrized SCREAMING_SNAKE_CASE =Tracker(self.src )(snake_case ).parametrized SCREAMING_SNAKE_CASE =list(filter(lambda snake_case : type(snake_case ) not in self.src_skip ,snake_case ) ) SCREAMING_SNAKE_CASE =list(filter(lambda snake_case : type(snake_case ) not in self.dest_skip ,snake_case ) ) if len(snake_case ) != len(snake_case ) and self.raise_if_mismatch: raise Exception( f'Numbers of operations are different. Source module has {len(snake_case )} operations while' f' destination module has {len(snake_case )}.' ) for dest_m, src_m in zip(snake_case ,snake_case ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(f'Transfered from={src_m} to={dest_m}' ) class a_ ( nn.Module ): """simple docstring""" def __init__( self : Any ,snake_case : nn.Module ): super().__init__() SCREAMING_SNAKE_CASE =[] # - get the stem feature_blocks.append(('conv1', model.stem) ) # - get all the feature blocks for k, v in model.trunk_output.named_children(): assert k.startswith('block' ), f'Unexpected layer name {k}' SCREAMING_SNAKE_CASE =len(snake_case ) + 1 feature_blocks.append((f'res{block_index}', v) ) SCREAMING_SNAKE_CASE =nn.ModuleDict(snake_case ) def _lowerCAmelCase ( self : Dict ,snake_case : Tensor ): return get_trunk_forward_outputs( snake_case ,out_feat_keys=snake_case ,feature_blocks=self._feature_blocks ,) class a_ ( lowerCamelCase_ ): """simple docstring""" def _lowerCAmelCase ( self : Optional[int] ,snake_case : str ): SCREAMING_SNAKE_CASE =x.split('-' ) return x_split[0] + x_split[1] + "_" + "".join(x_split[2:] ) def __getitem__( self : Optional[Any] ,snake_case : str ): # default to timm! if x not in self: SCREAMING_SNAKE_CASE =self.convert_name_to_timm(snake_case ) SCREAMING_SNAKE_CASE =partial(lambda: (timm.create_model(snake_case ,pretrained=snake_case ).eval(), None) ) else: SCREAMING_SNAKE_CASE =super().__getitem__(snake_case ) return val class a_ ( lowerCamelCase_ ): """simple docstring""" def __getitem__( self : int ,snake_case : str ): if "seer" in x and "in1k" not in x: SCREAMING_SNAKE_CASE =RegNetModel else: SCREAMING_SNAKE_CASE =RegNetForImageClassification return val def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ): """simple docstring""" for from_key, to_key in keys: SCREAMING_SNAKE_CASE =from_state_dict[from_key].clone() print(F'Copied key={from_key} to={to_key}' ) return to_state_dict def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ = True, ): """simple docstring""" print(F'Converting {name}...' ) with torch.no_grad(): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =from_model_func() SCREAMING_SNAKE_CASE =our_model_func(lowerCAmelCase_ ).eval() SCREAMING_SNAKE_CASE =ModuleTransfer(src=lowerCAmelCase_, dest=lowerCAmelCase_, raise_if_mismatch=lowerCAmelCase_ ) SCREAMING_SNAKE_CASE =torch.randn((1, 3, 224, 224) ) module_transfer(lowerCAmelCase_ ) if from_state_dict is not None: SCREAMING_SNAKE_CASE =[] # for seer - in1k finetuned we have to manually copy the head if "seer" in name and "in1k" in name: SCREAMING_SNAKE_CASE =[('0.clf.0.weight', 'classifier.1.weight'), ('0.clf.0.bias', 'classifier.1.bias')] SCREAMING_SNAKE_CASE =manually_copy_vissl_head(lowerCAmelCase_, our_model.state_dict(), lowerCAmelCase_ ) our_model.load_state_dict(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE =our_model(lowerCAmelCase_, output_hidden_states=lowerCAmelCase_ ) SCREAMING_SNAKE_CASE =( our_outputs.logits if isinstance(lowerCAmelCase_, lowerCAmelCase_ ) else our_outputs.last_hidden_state ) SCREAMING_SNAKE_CASE =from_model(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE =from_output[-1] if type(lowerCAmelCase_ ) is list else from_output # now since I don't want to use any config files, vissl seer model doesn't actually have an head, so let's just check the last hidden state if "seer" in name and "in1k" in name: SCREAMING_SNAKE_CASE =our_outputs.hidden_states[-1] assert torch.allclose(lowerCAmelCase_, lowerCAmelCase_ ), "The model logits don't match the original one." if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / name, commit_message='Add model', use_temp_dir=lowerCAmelCase_, ) SCREAMING_SNAKE_CASE =224 if 'seer' not in name else 384 # we can use the convnext one SCREAMING_SNAKE_CASE =AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k', size=lowerCAmelCase_ ) image_processor.push_to_hub( repo_path_or_name=save_directory / name, commit_message='Add image processor', use_temp_dir=lowerCAmelCase_, ) print(F'Pushed {name}' ) def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_ = None, lowerCAmelCase_ = True ): """simple docstring""" SCREAMING_SNAKE_CASE ='imagenet-1k-id2label.json' SCREAMING_SNAKE_CASE =1000 SCREAMING_SNAKE_CASE =(1, num_labels) SCREAMING_SNAKE_CASE ='huggingface/label-files' SCREAMING_SNAKE_CASE =num_labels SCREAMING_SNAKE_CASE =json.load(open(cached_download(hf_hub_url(lowerCAmelCase_, lowerCAmelCase_, repo_type='dataset' ) ), 'r' ) ) SCREAMING_SNAKE_CASE ={int(lowerCAmelCase_ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE =idalabel SCREAMING_SNAKE_CASE ={v: k for k, v in idalabel.items()} SCREAMING_SNAKE_CASE =partial(lowerCAmelCase_, num_labels=lowerCAmelCase_, idalabel=lowerCAmelCase_, labelaid=lowerCAmelCase_ ) SCREAMING_SNAKE_CASE ={ 'regnet-x-002': ImageNetPreTrainedConfig( depths=[1, 1, 4, 7], hidden_sizes=[24, 56, 152, 368], groups_width=8, layer_type='x' ), 'regnet-x-004': ImageNetPreTrainedConfig( depths=[1, 2, 7, 12], hidden_sizes=[32, 64, 160, 384], groups_width=16, layer_type='x' ), 'regnet-x-006': ImageNetPreTrainedConfig( depths=[1, 3, 5, 7], hidden_sizes=[48, 96, 240, 528], groups_width=24, layer_type='x' ), 'regnet-x-008': ImageNetPreTrainedConfig( depths=[1, 3, 7, 5], hidden_sizes=[64, 128, 288, 672], groups_width=16, layer_type='x' ), 'regnet-x-016': ImageNetPreTrainedConfig( depths=[2, 4, 10, 2], hidden_sizes=[72, 168, 408, 912], groups_width=24, layer_type='x' ), 'regnet-x-032': ImageNetPreTrainedConfig( depths=[2, 6, 15, 2], hidden_sizes=[96, 192, 432, 1008], groups_width=48, layer_type='x' ), 'regnet-x-040': ImageNetPreTrainedConfig( depths=[2, 5, 14, 2], hidden_sizes=[80, 240, 560, 1360], groups_width=40, layer_type='x' ), 'regnet-x-064': ImageNetPreTrainedConfig( depths=[2, 4, 10, 1], hidden_sizes=[168, 392, 784, 1624], groups_width=56, layer_type='x' ), 'regnet-x-080': ImageNetPreTrainedConfig( depths=[2, 5, 15, 1], hidden_sizes=[80, 240, 720, 1920], groups_width=120, layer_type='x' ), 'regnet-x-120': ImageNetPreTrainedConfig( depths=[2, 5, 11, 1], hidden_sizes=[224, 448, 896, 2240], groups_width=112, layer_type='x' ), 'regnet-x-160': ImageNetPreTrainedConfig( depths=[2, 6, 13, 1], hidden_sizes=[256, 512, 896, 2048], groups_width=128, layer_type='x' ), 'regnet-x-320': ImageNetPreTrainedConfig( depths=[2, 7, 13, 1], hidden_sizes=[336, 672, 1344, 2520], groups_width=168, layer_type='x' ), # y variant 'regnet-y-002': ImageNetPreTrainedConfig(depths=[1, 1, 4, 7], hidden_sizes=[24, 56, 152, 368], groups_width=8 ), 'regnet-y-004': ImageNetPreTrainedConfig( depths=[1, 3, 6, 6], hidden_sizes=[48, 104, 208, 440], groups_width=8 ), 'regnet-y-006': ImageNetPreTrainedConfig( depths=[1, 3, 7, 4], hidden_sizes=[48, 112, 256, 608], groups_width=16 ), 'regnet-y-008': ImageNetPreTrainedConfig( depths=[1, 3, 8, 2], hidden_sizes=[64, 128, 320, 768], groups_width=16 ), 'regnet-y-016': ImageNetPreTrainedConfig( depths=[2, 6, 17, 2], hidden_sizes=[48, 120, 336, 888], groups_width=24 ), 'regnet-y-032': ImageNetPreTrainedConfig( depths=[2, 5, 13, 1], hidden_sizes=[72, 216, 576, 1512], groups_width=24 ), 'regnet-y-040': ImageNetPreTrainedConfig( depths=[2, 6, 12, 2], hidden_sizes=[128, 192, 512, 1088], groups_width=64 ), 'regnet-y-064': ImageNetPreTrainedConfig( depths=[2, 7, 14, 2], hidden_sizes=[144, 288, 576, 1296], groups_width=72 ), 'regnet-y-080': ImageNetPreTrainedConfig( depths=[2, 4, 10, 1], hidden_sizes=[168, 448, 896, 2016], groups_width=56 ), 'regnet-y-120': ImageNetPreTrainedConfig( depths=[2, 5, 11, 1], hidden_sizes=[224, 448, 896, 2240], groups_width=112 ), 'regnet-y-160': ImageNetPreTrainedConfig( depths=[2, 4, 11, 1], hidden_sizes=[224, 448, 1232, 3024], groups_width=112 ), 'regnet-y-320': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1], hidden_sizes=[232, 696, 1392, 3712], groups_width=232 ), # models created by SEER -> https://arxiv.org/abs/2202.08360 'regnet-y-320-seer': RegNetConfig(depths=[2, 5, 12, 1], hidden_sizes=[232, 696, 1392, 3712], groups_width=232 ), 'regnet-y-640-seer': RegNetConfig(depths=[2, 5, 12, 1], hidden_sizes=[328, 984, 1968, 4920], groups_width=328 ), 'regnet-y-1280-seer': RegNetConfig( depths=[2, 7, 17, 1], hidden_sizes=[528, 1056, 2904, 7392], groups_width=264 ), 'regnet-y-2560-seer': RegNetConfig( depths=[3, 7, 16, 1], hidden_sizes=[640, 1696, 2544, 5088], groups_width=640 ), 'regnet-y-10b-seer': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1], hidden_sizes=[2020, 4040, 11110, 28280], groups_width=1010 ), # finetuned on imagenet 'regnet-y-320-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1], hidden_sizes=[232, 696, 1392, 3712], groups_width=232 ), 'regnet-y-640-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1], hidden_sizes=[328, 984, 1968, 4920], groups_width=328 ), 'regnet-y-1280-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1], hidden_sizes=[528, 1056, 2904, 7392], groups_width=264 ), 'regnet-y-2560-seer-in1k': ImageNetPreTrainedConfig( depths=[3, 7, 16, 1], hidden_sizes=[640, 1696, 2544, 5088], groups_width=640 ), 'regnet-y-10b-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1], hidden_sizes=[2020, 4040, 11110, 28280], groups_width=1010 ), } SCREAMING_SNAKE_CASE =NameToOurModelFuncMap() SCREAMING_SNAKE_CASE =NameToFromModelFuncMap() # add seer weights logic def load_using_classy_vision(lowerCAmelCase_, lowerCAmelCase_ ) -> Tuple[nn.Module, Dict]: SCREAMING_SNAKE_CASE =torch.hub.load_state_dict_from_url(lowerCAmelCase_, model_dir=str(lowerCAmelCase_ ), map_location='cpu' ) SCREAMING_SNAKE_CASE =model_func() # check if we have a head, if yes add it SCREAMING_SNAKE_CASE =files['classy_state_dict']['base_model']['model'] SCREAMING_SNAKE_CASE =model_state_dict['trunk'] model.load_state_dict(lowerCAmelCase_ ) return model.eval(), model_state_dict["heads"] # pretrained SCREAMING_SNAKE_CASE =partial( lowerCAmelCase_, 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet32d/seer_regnet32gf_model_iteration244000.torch', lambda: FakeRegNetVisslWrapper(RegNetYaagf() ), ) SCREAMING_SNAKE_CASE =partial( lowerCAmelCase_, 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet64/seer_regnet64gf_model_final_checkpoint_phase0.torch', lambda: FakeRegNetVisslWrapper(RegNetYaagf() ), ) SCREAMING_SNAKE_CASE =partial( lowerCAmelCase_, 'https://dl.fbaipublicfiles.com/vissl/model_zoo/swav_ig1b_regnet128Gf_cnstant_bs32_node16_sinkhorn10_proto16k_syncBN64_warmup8k/model_final_checkpoint_phase0.torch', lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ), ) SCREAMING_SNAKE_CASE =partial( lowerCAmelCase_, 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet10B/model_iteration124500_conso.torch', lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27, group_width=1010, w_a=1744, w_a=620.83, w_m=2.52 ) ) ), ) # IN1K finetuned SCREAMING_SNAKE_CASE =partial( lowerCAmelCase_, 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet32_finetuned_in1k_model_final_checkpoint_phase78.torch', lambda: FakeRegNetVisslWrapper(RegNetYaagf() ), ) SCREAMING_SNAKE_CASE =partial( lowerCAmelCase_, 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet64_finetuned_in1k_model_final_checkpoint_phase78.torch', lambda: FakeRegNetVisslWrapper(RegNetYaagf() ), ) SCREAMING_SNAKE_CASE =partial( lowerCAmelCase_, 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet128_finetuned_in1k_model_final_checkpoint_phase78.torch', lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ), ) SCREAMING_SNAKE_CASE =partial( lowerCAmelCase_, 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_10b_finetuned_in1k_model_phase28_conso.torch', lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27, group_width=1010, w_a=1744, w_a=620.83, w_m=2.52 ) ) ), ) if model_name: convert_weight_and_push( lowerCAmelCase_, names_to_from_model_map[model_name], names_to_ours_model_map[model_name], names_to_config[model_name], lowerCAmelCase_, lowerCAmelCase_, ) else: for model_name, config in names_to_config.items(): convert_weight_and_push( lowerCAmelCase_, names_to_from_model_map[model_name], names_to_ours_model_map[model_name], lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, ) return config, expected_shape if __name__ == "__main__": _lowerCamelCase =argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default=None, type=str, help=( "The name of the model you wish to convert, it must be one of the supported regnet* architecture," " currently: regnetx-*, regnety-*. If `None`, all of them will the converted." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=Path, required=True, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", default=True, type=bool, required=False, help="If True, push model and image processor to the hub.", ) _lowerCamelCase =parser.parse_args() _lowerCamelCase =args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
334
0
def __UpperCamelCase ( _A ): lowerCAmelCase_ = generate_pascal_triangle(_A ) for row_idx in range(_A ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=''' ''' ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=''' ''' ) else: print(triangle[row_idx][col_idx] , end='''''' ) print() def __UpperCamelCase ( _A ): if not isinstance(_A , _A ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) lowerCAmelCase_ = [] for current_row_idx in range(_A ): lowerCAmelCase_ = populate_current_row(_A , _A ) triangle.append(_A ) return triangle def __UpperCamelCase ( _A , _A ): lowerCAmelCase_ = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 lowerCAmelCase_ , lowerCAmelCase_ = 1, 1 for current_col_idx in range(1 , _A ): calculate_current_element( _A , _A , _A , _A ) return current_row def __UpperCamelCase ( _A , _A , _A , _A , ): lowerCAmelCase_ = triangle[current_row_idx - 1][current_col_idx - 1] lowerCAmelCase_ = triangle[current_row_idx - 1][current_col_idx] lowerCAmelCase_ = above_to_left_elt + above_to_right_elt def __UpperCamelCase ( _A ): if not isinstance(_A , _A ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) lowerCAmelCase_ = [[1]] for row_index in range(1 , _A ): lowerCAmelCase_ = [0] + result[-1] + [0] lowerCAmelCase_ = row_index + 1 # Calculate the number of distinct elements in a row lowerCAmelCase_ = sum(divmod(_A , 2 ) ) lowerCAmelCase_ = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] lowerCAmelCase_ = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() lowerCAmelCase_ = row_first_half + row_second_half result.append(_A ) return result def __UpperCamelCase ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(_A , _A ) -> None: lowerCAmelCase_ = f"{func.__name__}({value})" lowerCAmelCase_ = timeit(f"__main__.{call}" , setup='''import __main__''' ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(f"{call:38} -- {timing:.4f} seconds" ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(_A , _A ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
354
def __UpperCamelCase ( _A = 4000000 ): lowerCAmelCase_ = [0, 1] lowerCAmelCase_ = 0 while fib[i] <= n: fib.append(fib[i] + fib[i + 1] ) if fib[i + 2] > n: break i += 1 lowerCAmelCase_ = 0 for j in range(len(_A ) - 1 ): if fib[j] % 2 == 0: total += fib[j] return total if __name__ == "__main__": print(f"{solution() = }")
167
0
'''simple docstring''' from collections import defaultdict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> bool: lowerCamelCase__ : List[str] = first_str.lower().strip() lowerCamelCase__ : Dict = second_str.lower().strip() # Remove whitespace lowerCamelCase__ : int = first_str.replace(""" """ , """""" ) lowerCamelCase__ : Optional[int] = second_str.replace(""" """ , """""" ) # Strings of different lengths are not anagrams if len(UpperCamelCase ) != len(UpperCamelCase ): return False # Default values for count should be 0 lowerCamelCase__ : defaultdict[str, int] = defaultdict(UpperCamelCase ) # For each character in input strings, # increment count in the corresponding for i in range(len(UpperCamelCase ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() _A : Optional[Any] =input('''Enter the first string ''').strip() _A : str =input('''Enter the second string ''').strip() _A : Dict =check_anagrams(input_a, input_b) print(F'{input_a} and {input_b} are {"" if status else "not "}anagrams.')
41
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __A = logging.get_logger(__name__) __A = {"vocab_file": "spm_char.model"} __A = { "vocab_file": { "microsoft/speecht5_asr": "https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model", "microsoft/speecht5_tts": "https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model", "microsoft/speecht5_vc": "https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model", } } __A = { "microsoft/speecht5_asr": 1_024, "microsoft/speecht5_tts": 1_024, "microsoft/speecht5_vc": 1_024, } class A ( __UpperCAmelCase ): lowerCamelCase : List[str] = VOCAB_FILES_NAMES lowerCamelCase : Dict = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : List[Any] = ["""input_ids""", """attention_mask"""] def __init__( self , lowerCamelCase__ , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<pad>" , lowerCamelCase__ = None , **lowerCamelCase__ , ) -> None: '''simple docstring''' lowercase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase__ , ) lowercase__ = vocab_file lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase__ ) @property def A__ ( self ) -> str: '''simple docstring''' return self.sp_model.get_piece_size() def A__ ( self ) -> int: '''simple docstring''' lowercase__ = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Dict: '''simple docstring''' lowercase__ = self.__dict__.copy() lowercase__ = None return state def __setstate__( self , lowerCamelCase__ ) -> Union[str, Any]: '''simple docstring''' lowercase__ = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): lowercase__ = {} lowercase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A__ ( self , lowerCamelCase__ ) -> List[str]: '''simple docstring''' return self.sp_model.encode(lowerCamelCase__ , out_type=lowerCamelCase__ ) def A__ ( self , lowerCamelCase__ ) -> Optional[int]: '''simple docstring''' return self.sp_model.piece_to_id(lowerCamelCase__ ) def A__ ( self , lowerCamelCase__ ) -> int: '''simple docstring''' lowercase__ = self.sp_model.IdToPiece(lowerCamelCase__ ) return token def A__ ( self , lowerCamelCase__ ) -> str: '''simple docstring''' lowercase__ = [] lowercase__ = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowerCamelCase__ ) + token lowercase__ = [] else: current_sub_tokens.append(lowerCamelCase__ ) out_string += self.sp_model.decode(lowerCamelCase__ ) return out_string.strip() def A__ ( self , lowerCamelCase__ , lowerCamelCase__=None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def A__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) lowercase__ = [1] if token_ids_a is None: return ([0] * len(lowerCamelCase__ )) + suffix_ones return ([0] * len(lowerCamelCase__ )) + ([0] * len(lowerCamelCase__ )) + suffix_ones def A__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase__ = os.path.join( lowerCamelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ , """wb""" ) as fi: lowercase__ = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,)
164
0
'''simple docstring''' import math import sys def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Tuple = '' try: with open(_A , 'rb' ) as binary_file: _lowerCAmelCase : Dict = binary_file.read() for dat in data: _lowerCAmelCase : Any = f'{dat:08b}' result += curr_byte return result except OSError: print('File not accessible' ) sys.exit() def lowercase (_A ): """simple docstring""" _lowerCAmelCase : List[str] = {'0': '0', '1': '1'} _lowerCAmelCase , _lowerCAmelCase : Optional[int] = '', '' _lowerCAmelCase : Optional[Any] = len(_A ) for i in range(len(_A ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue _lowerCAmelCase : Any = lexicon[curr_string] result += last_match_id _lowerCAmelCase : Dict = last_match_id + '0' if math.loga(_A ).is_integer(): _lowerCAmelCase : str = {} for curr_key in list(_A ): _lowerCAmelCase : Any = lexicon.pop(_A ) _lowerCAmelCase : Tuple = new_lex _lowerCAmelCase : List[str] = last_match_id + '1' index += 1 _lowerCAmelCase : Dict = '' return result def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : Union[str, Any] = 8 try: with open(_A , 'wb' ) as opened_file: _lowerCAmelCase : Tuple = [ to_write[i : i + byte_length] for i in range(0 , len(_A ) , _A ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('10000000' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(_A , 2 ).to_bytes(1 , byteorder='big' ) ) except OSError: print('File not accessible' ) sys.exit() def lowercase (_A ): """simple docstring""" _lowerCAmelCase : Union[str, Any] = 0 for letter in data_bits: if letter == "1": break counter += 1 _lowerCAmelCase : Optional[Any] = data_bits[counter:] _lowerCAmelCase : str = data_bits[counter + 1 :] return data_bits def lowercase (_A , _A ): """simple docstring""" _lowerCAmelCase : List[str] = read_file_binary(_A ) _lowerCAmelCase : List[str] = remove_prefix(_A ) _lowerCAmelCase : str = decompress_data(_A ) write_file_binary(_A , _A ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
25
'''simple docstring''' def lowercase (_A = 1_0_0_0_0_0_0 ): """simple docstring""" _lowerCAmelCase : Any = set(range(3 , _A , 2 ) ) primes.add(2 ) for p in range(3 , _A , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , _A , _A ) ) ) _lowerCAmelCase : Union[str, Any] = [float(_A ) for n in range(limit + 1 )] for p in primes: for n in range(_A , limit + 1 , _A ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F'''{solution() = }''')
25
1
"""simple docstring""" def snake_case_ ( A_ : int ): '''simple docstring''' assert isinstance(A_, A_ ), F'''The input value of [n={number}] is not an integer''' if number == 1: return 2 elif number < 1: _lowerCamelCase : List[Any] = F'''The input value of [n={number}] has to be > 0''' raise ValueError(A_ ) else: _lowerCamelCase : Dict = sylvester(number - 1 ) _lowerCamelCase : Dict = num - 1 _lowerCamelCase : Optional[int] = num return lower * upper + 1 if __name__ == "__main__": print(F"""The 8th number in Sylvester's sequence: {sylvester(8)}""")
72
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __UpperCAmelCase ( lowerCamelCase__ , unittest.TestCase ): UpperCamelCase = LayoutLMTokenizer UpperCamelCase = LayoutLMTokenizerFast UpperCamelCase = True UpperCamelCase = True def __magic_name__ ( self : Any ): super().setUp() UpperCAmelCase : Dict = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] UpperCAmelCase : int = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file, '''w''', encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __magic_name__ ( self : Union[str, Any], **__A : List[str] ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname, **__A ) def __magic_name__ ( self : Optional[int], __A : int ): UpperCAmelCase : Optional[Any] = '''UNwant\u00E9d,running''' UpperCAmelCase : Optional[int] = '''unwanted, running''' return input_text, output_text def __magic_name__ ( self : Any ): UpperCAmelCase : Union[str, Any] = self.tokenizer_class(self.vocab_file ) UpperCAmelCase : Optional[Any] = tokenizer.tokenize('''UNwant\u00E9d,running''' ) self.assertListEqual(__A, ['''un''', '''##want''', '''##ed''', ''',''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ), [7, 4, 5, 1_0, 8, 9] ) def __magic_name__ ( self : Optional[int] ): pass
336
0
import math def lowerCAmelCase_ ( __UpperCAmelCase: float , __UpperCAmelCase: float ) -> float: if ( not isinstance(__UpperCAmelCase , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * power_factor def lowerCAmelCase_ ( __UpperCAmelCase: float , __UpperCAmelCase: float ) -> float: if ( not isinstance(__UpperCAmelCase , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
357
from __future__ import annotations import bisect def lowerCAmelCase_ ( __UpperCAmelCase: list[int] , __UpperCAmelCase: int , __UpperCAmelCase: int = 0 , __UpperCAmelCase: int = -1 ) -> int: if hi < 0: UpperCamelCase__ : Union[str, Any] = len(__UpperCAmelCase ) while lo < hi: UpperCamelCase__ : Optional[Any] = lo + (hi - lo) // 2 if sorted_collection[mid] < item: UpperCamelCase__ : Optional[int] = mid + 1 else: UpperCamelCase__ : Tuple = mid return lo def lowerCAmelCase_ ( __UpperCAmelCase: list[int] , __UpperCAmelCase: int , __UpperCAmelCase: int = 0 , __UpperCAmelCase: int = -1 ) -> int: if hi < 0: UpperCamelCase__ : int = len(__UpperCAmelCase ) while lo < hi: UpperCamelCase__ : List[Any] = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: UpperCamelCase__ : Optional[int] = mid + 1 else: UpperCamelCase__ : int = mid return lo def lowerCAmelCase_ ( __UpperCAmelCase: list[int] , __UpperCAmelCase: int , __UpperCAmelCase: int = 0 , __UpperCAmelCase: int = -1 ) -> None: sorted_collection.insert(bisect_left(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) def lowerCAmelCase_ ( __UpperCAmelCase: list[int] , __UpperCAmelCase: int , __UpperCAmelCase: int = 0 , __UpperCAmelCase: int = -1 ) -> None: sorted_collection.insert(bisect_right(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) , __UpperCAmelCase ) def lowerCAmelCase_ ( __UpperCAmelCase: list[int] , __UpperCAmelCase: int ) -> int | None: UpperCamelCase__ : Optional[int] = 0 UpperCamelCase__ : List[str] = len(__UpperCAmelCase ) - 1 while left <= right: UpperCamelCase__ : List[str] = left + (right - left) // 2 UpperCamelCase__ : Union[str, Any] = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: UpperCamelCase__ : List[str] = midpoint - 1 else: UpperCamelCase__ : List[str] = midpoint + 1 return None def lowerCAmelCase_ ( __UpperCAmelCase: list[int] , __UpperCAmelCase: int ) -> int | None: UpperCamelCase__ : Union[str, Any] = bisect.bisect_left(__UpperCAmelCase , __UpperCAmelCase ) if index != len(__UpperCAmelCase ) and sorted_collection[index] == item: return index return None def lowerCAmelCase_ ( __UpperCAmelCase: list[int] , __UpperCAmelCase: int , __UpperCAmelCase: int , __UpperCAmelCase: int ) -> int | None: if right < left: return None UpperCamelCase__ : Optional[int] = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , midpoint - 1 ) else: return binary_search_by_recursion(__UpperCAmelCase , __UpperCAmelCase , midpoint + 1 , __UpperCAmelCase ) if __name__ == "__main__": UpperCAmelCase_ = input('Enter numbers separated by comma:\n').strip() UpperCAmelCase_ = sorted(int(item) for item in user_input.split(',')) UpperCAmelCase_ = int(input('Enter a single number to be found in the list:\n')) UpperCAmelCase_ = binary_search(collection, target) if result is None: print(F'''{target} was not found in {collection}.''') else: print(F'''{target} was found at position {result} in {collection}.''')
247
0
import math class UpperCAmelCase : '''simple docstring''' def __init__( self : Optional[int] ,A : Dict=0 ): # a graph with Node 0,1,...,N-1 __A = n __A = [ [math.inf for j in range(0 ,A )] for i in range(0 ,A ) ] # adjacency matrix for weight __A = [ [math.inf for j in range(0 ,A )] for i in range(0 ,A ) ] # dp[i][j] stores minimum distance from i to j def UpperCamelCase_ ( self : List[str] ,A : Tuple ,A : Union[str, Any] ,A : Optional[int] ): __A = w def UpperCamelCase_ ( self : int ): for k in range(0 ,self.n ): for i in range(0 ,self.n ): for j in range(0 ,self.n ): __A = min(self.dp[i][j] ,self.dp[i][k] + self.dp[k][j] ) def UpperCamelCase_ ( self : Optional[int] ,A : str ,A : Union[str, Any] ): return self.dp[u][v] if __name__ == "__main__": SCREAMING_SNAKE_CASE :Optional[int] = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 10) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 10) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
15
from math import sqrt def _snake_case( SCREAMING_SNAKE_CASE__ ) -> bool: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number >= 0 ), "'number' must been an int and positive" lowercase : Union[str, Any] = True # 0 and 1 are none primes. if number <= 1: lowercase : str = False for divisor in range(2 , int(round(sqrt(SCREAMING_SNAKE_CASE__ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowercase : Any = False break # precondition assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), "'status' must been from type bool" return status def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowercase : str = list(range(2 , n + 1 ) ) lowercase : Tuple = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(SCREAMING_SNAKE_CASE__ ) ): for j in range(i + 1 , len(SCREAMING_SNAKE_CASE__ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowercase : Tuple = 0 # filters actual prime numbers. lowercase : int = [x for x in begin_list if x != 0] # precondition assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), "'ans' must been from type list" return ans def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (n > 2), "'N' must been an int and > 2" lowercase : Dict = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(SCREAMING_SNAKE_CASE__ ): ans.append(SCREAMING_SNAKE_CASE__ ) # precondition assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), "'ans' must been from type list" return ans def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Tuple: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and number >= 0, "'number' must been an int and >= 0" lowercase : Tuple = [] # this list will be returns of the function. # potential prime number factors. lowercase : Optional[Any] = 2 lowercase : Any = number if number == 0 or number == 1: ans.append(SCREAMING_SNAKE_CASE__ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(SCREAMING_SNAKE_CASE__ ): while quotient != 1: if is_prime(SCREAMING_SNAKE_CASE__ ) and (quotient % factor == 0): ans.append(SCREAMING_SNAKE_CASE__ ) quotient /= factor else: factor += 1 else: ans.append(SCREAMING_SNAKE_CASE__ ) # precondition assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), "'ans' must been from type list" return ans def _snake_case( SCREAMING_SNAKE_CASE__ ) -> List[Any]: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowercase : Tuple = 0 # prime factorization of 'number' lowercase : Optional[int] = prime_factorization(SCREAMING_SNAKE_CASE__ ) lowercase : Union[str, Any] = max(SCREAMING_SNAKE_CASE__ ) # precondition assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), "'ans' must been from type int" return ans def _snake_case( SCREAMING_SNAKE_CASE__ ) -> str: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowercase : Union[str, Any] = 0 # prime factorization of 'number' lowercase : Tuple = prime_factorization(SCREAMING_SNAKE_CASE__ ) lowercase : Union[str, Any] = min(SCREAMING_SNAKE_CASE__ ) # precondition assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), "'ans' must been from type int" return ans def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), "'number' must been an int" assert isinstance(number % 2 == 0 , SCREAMING_SNAKE_CASE__ ), "compare bust been from type bool" return number % 2 == 0 def _snake_case( SCREAMING_SNAKE_CASE__ ) -> int: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), "'number' must been an int" assert isinstance(number % 2 != 0 , SCREAMING_SNAKE_CASE__ ), "compare bust been from type bool" return number % 2 != 0 def _snake_case( SCREAMING_SNAKE_CASE__ ) -> int: assert ( isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (number > 2) and is_even(SCREAMING_SNAKE_CASE__ ) ), "'number' must been an int, even and > 2" lowercase : Union[str, Any] = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowercase : str = get_prime_numbers(SCREAMING_SNAKE_CASE__ ) lowercase : Any = len(SCREAMING_SNAKE_CASE__ ) # run variable for while-loops. lowercase : Optional[Any] = 0 lowercase : List[Any] = None # exit variable. for break up the loops lowercase : Any = True while i < len_pn and loop: lowercase : str = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowercase : Union[str, Any] = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (len(SCREAMING_SNAKE_CASE__ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> str: assert ( isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowercase : Union[str, Any] = 0 while numbera != 0: lowercase : Optional[int] = numbera % numbera lowercase : Optional[int] = numbera lowercase : Dict = rest # precondition assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Tuple: assert ( isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowercase : Dict = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowercase : Optional[Any] = prime_factorization(SCREAMING_SNAKE_CASE__ ) lowercase : Optional[int] = prime_factorization(SCREAMING_SNAKE_CASE__ ) elif numbera == 1 or numbera == 1: lowercase : Union[str, Any] = [] lowercase : List[str] = [] lowercase : Dict = max(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase : Union[str, Any] = 0 lowercase : Optional[Any] = 0 lowercase : List[str] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowercase : Dict = prime_fac_a.count(SCREAMING_SNAKE_CASE__ ) lowercase : Optional[Any] = prime_fac_a.count(SCREAMING_SNAKE_CASE__ ) for _ in range(max(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ): ans *= n else: lowercase : List[Any] = prime_fac_a.count(SCREAMING_SNAKE_CASE__ ) for _ in range(SCREAMING_SNAKE_CASE__ ): ans *= n done.append(SCREAMING_SNAKE_CASE__ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowercase : Optional[int] = prime_fac_a.count(SCREAMING_SNAKE_CASE__ ) for _ in range(SCREAMING_SNAKE_CASE__ ): ans *= n done.append(SCREAMING_SNAKE_CASE__ ) # precondition assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Any: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (n >= 0), "'number' must been a positive int" lowercase : Dict = 0 lowercase : List[str] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(SCREAMING_SNAKE_CASE__ ): ans += 1 # precondition assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and is_prime( SCREAMING_SNAKE_CASE__ ), "'ans' must been a prime number and from type int" return ans def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[Any]: assert ( is_prime(SCREAMING_SNAKE_CASE__ ) and is_prime(SCREAMING_SNAKE_CASE__ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowercase : List[str] = p_number_a + 1 # jump to the next number lowercase : List[Any] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(SCREAMING_SNAKE_CASE__ ): number += 1 while number < p_number_a: ans.append(SCREAMING_SNAKE_CASE__ ) number += 1 # fetch the next prime number. while not is_prime(SCREAMING_SNAKE_CASE__ ): number += 1 # precondition assert ( isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ans[0] != p_number_a and ans[len(SCREAMING_SNAKE_CASE__ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Any: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (n >= 1), "'n' must been int and >= 1" lowercase : Optional[Any] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(SCREAMING_SNAKE_CASE__ ) # precondition assert ans[0] == 1 and ans[len(SCREAMING_SNAKE_CASE__ ) - 1] == n, "Error in function getDivisiors(...)" return ans def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Any: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and ( number > 1 ), "'number' must been an int and >= 1" lowercase : str = get_divisors(SCREAMING_SNAKE_CASE__ ) # precondition assert ( isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (divisors[0] == 1) and (divisors[len(SCREAMING_SNAKE_CASE__ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Dict: assert ( isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowercase : Tuple = gcd(abs(SCREAMING_SNAKE_CASE__ ) , abs(SCREAMING_SNAKE_CASE__ ) ) # precondition assert ( isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def _snake_case( SCREAMING_SNAKE_CASE__ ) -> int: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (n >= 0), "'n' must been a int and >= 0" lowercase : List[str] = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: assert isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and (n >= 0), "'n' must been an int and >= 0" lowercase : int = 0 lowercase : Union[str, Any] = 1 lowercase : int = 1 # this will be return for _ in range(n - 1 ): lowercase : Optional[int] = ans ans += fiba lowercase : Optional[int] = tmp return ans
20
0
"""simple docstring""" from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class snake_case ( UpperCAmelCase ): def lowerCamelCase__ ( self : Union[str, Any] , A : float ): '''simple docstring''' return 0.0 def snake_case (A_ :np.ndarray , A_ :int ): '''simple docstring''' a : Any = min([-2_0, np.min(fft_results[1 : samplerate // 2 - 1] )] ) a : Optional[int] = max([2_0, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def snake_case (A_ :FilterType , A_ :int ): '''simple docstring''' a : List[Any] = 5_1_2 a : Any = [1] + [0] * (size - 1) a : int = [filter_type.process(A_ ) for item in inputs] a : Dict = [0] * (samplerate - size) # zero-padding outputs += filler a : List[str] = np.abs(np.fft.fft(A_ ) ) a : Optional[int] = 2_0 * np.logaa(A_ ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(2_4 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) # Display within reasonable bounds a : List[Any] = get_bounds(A_ , A_ ) plt.ylim(max([-8_0, bounds[0]] ) , min([8_0, bounds[1]] ) ) plt.ylabel('Gain (dB)' ) plt.plot(A_ ) plt.show() def snake_case (A_ :FilterType , A_ :int ): '''simple docstring''' a : Any = 5_1_2 a : Tuple = [1] + [0] * (size - 1) a : List[Any] = [filter_type.process(A_ ) for item in inputs] a : int = [0] * (samplerate - size) # zero-padding outputs += filler a : Union[str, Any] = np.angle(np.fft.fft(A_ ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(2_4 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('Phase shift (Radians)' ) plt.plot(np.unwrap(A_ , -2 * pi ) ) plt.show()
369
"""simple docstring""" def snake_case (A_ :str , A_ :bool = False ): '''simple docstring''' if not isinstance(A_ , A_ ): a : Union[str, Any] = f'''Expected string as input, found {type(A_ )}''' raise ValueError(A_ ) if not isinstance(A_ , A_ ): a : Optional[int] = f'''Expected boolean as use_pascal parameter, found {type(A_ )}''' raise ValueError(A_ ) a : Tuple = input_str.split('_' ) a : Dict = 0 if use_pascal else 1 a : int = words[start_index:] a : int = [word[0].upper() + word[1:] for word in words_to_capitalize] a : List[str] = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
186
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __A : Dict = { '''configuration_nezha''': ['''NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''NezhaConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ '''NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''NezhaForNextSentencePrediction''', '''NezhaForMaskedLM''', '''NezhaForPreTraining''', '''NezhaForMultipleChoice''', '''NezhaForQuestionAnswering''', '''NezhaForSequenceClassification''', '''NezhaForTokenClassification''', '''NezhaModel''', '''NezhaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
"""simple docstring""" def lowercase ( __snake_case : int = 1_0_0 ): lowercase_ : str = 0 lowercase_ : List[Any] = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
33
1
import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class snake_case__ ( unittest.TestCase ): """simple docstring""" def __init__( self : List[str] , UpperCamelCase__ : str , UpperCamelCase__ : List[str]=3 , UpperCamelCase__ : List[Any]=32 , UpperCamelCase__ : Optional[int]=3 , UpperCamelCase__ : Dict=10 , UpperCamelCase__ : Any=[10, 20, 30, 40] , UpperCamelCase__ : Any=[1, 1, 2, 1] , UpperCamelCase__ : List[str]=True , UpperCamelCase__ : List[Any]=True , UpperCamelCase__ : str="relu" , UpperCamelCase__ : Union[str, Any]=3 , UpperCamelCase__ : Tuple=None , ) -> List[str]: """simple docstring""" snake_case : List[str] = parent snake_case : Tuple = batch_size snake_case : int = image_size snake_case : Any = num_channels snake_case : Optional[int] = embeddings_size snake_case : Optional[int] = hidden_sizes snake_case : str = depths snake_case : Tuple = is_training snake_case : List[str] = use_labels snake_case : List[str] = hidden_act snake_case : Tuple = num_labels snake_case : Tuple = scope snake_case : List[str] = len(UpperCamelCase__ ) def lowerCAmelCase ( self : Dict ) -> int: """simple docstring""" snake_case : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case : Any = self.get_config() return config, pixel_values def lowerCAmelCase ( self : List[str] ) -> str: """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def lowerCAmelCase ( self : str , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int ) -> Tuple: """simple docstring""" snake_case : List[str] = FlaxRegNetModel(config=UpperCamelCase__ ) snake_case : str = model(UpperCamelCase__ ) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowerCAmelCase ( self : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Dict ) -> Dict: """simple docstring""" snake_case : int = self.num_labels snake_case : List[str] = FlaxRegNetForImageClassification(config=UpperCamelCase__ ) snake_case : Any = model(UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase ( self : str ) -> Dict: """simple docstring""" snake_case : str = self.prepare_config_and_inputs() snake_case : Tuple = config_and_inputs snake_case : int = {'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class snake_case__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" lowerCamelCase = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def lowerCAmelCase ( self : List[str] ) -> None: """simple docstring""" snake_case : List[str] = FlaxRegNetModelTester(self ) snake_case : Union[str, Any] = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def lowerCAmelCase ( self : Any ) -> str: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCAmelCase ( self : Tuple ) -> Any: """simple docstring""" return def lowerCAmelCase ( self : Optional[int] ) -> str: """simple docstring""" snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCAmelCase ( self : str ) -> Dict: """simple docstring""" snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def lowerCAmelCase ( self : List[str] ) -> int: """simple docstring""" pass @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def lowerCAmelCase ( self : Union[str, Any] ) -> Any: """simple docstring""" pass def lowerCAmelCase ( self : Optional[int] ) -> str: """simple docstring""" snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Union[str, Any] = model_class(UpperCamelCase__ ) snake_case : Union[str, Any] = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case : int = [*signature.parameters.keys()] snake_case : Optional[Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCAmelCase ( self : List[Any] ) -> List[str]: """simple docstring""" def check_hidden_states_output(UpperCamelCase__ : List[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Tuple ): snake_case : Union[str, Any] = model_class(UpperCamelCase__ ) snake_case : Any = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) snake_case : Optional[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states snake_case : Optional[int] = self.model_tester.num_stages self.assertEqual(len(UpperCamelCase__ ) , expected_num_stages + 1 ) snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Tuple = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case : List[str] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCAmelCase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" snake_case : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): snake_case : Any = self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) snake_case : Optional[Any] = model_class(UpperCamelCase__ ) @jax.jit def model_jitted(UpperCamelCase__ : Dict , **UpperCamelCase__ : Tuple ): return model(pixel_values=UpperCamelCase__ , **UpperCamelCase__ ) with self.subTest('''JIT Enabled''' ): snake_case : Optional[int] = model_jitted(**UpperCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): snake_case : Tuple = model_jitted(**UpperCamelCase__ ).to_tuple() self.assertEqual(len(UpperCamelCase__ ) , len(UpperCamelCase__ ) ) for jitted_output, output in zip(UpperCamelCase__ , UpperCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def _UpperCamelCase ( ) -> Union[str, Any]: '''simple docstring''' snake_case : List[str] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_flax class snake_case__ ( unittest.TestCase ): """simple docstring""" @cached_property def lowerCAmelCase ( self : Optional[Any] ) -> Tuple: """simple docstring""" return AutoImageProcessor.from_pretrained('''facebook/regnet-y-040''' ) if is_vision_available() else None @slow def lowerCAmelCase ( self : int ) -> int: """simple docstring""" snake_case : str = FlaxRegNetForImageClassification.from_pretrained('''facebook/regnet-y-040''' ) snake_case : Any = self.default_image_processor snake_case : Any = prepare_img() snake_case : Union[str, Any] = image_processor(images=UpperCamelCase__ , return_tensors='''np''' ) snake_case : List[str] = model(**UpperCamelCase__ ) # verify the logits snake_case : Optional[int] = (1, 1000) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) snake_case : Dict = jnp.array([-0.4_180, -1.5_051, -3.4_836] ) self.assertTrue(jnp.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) )
364
'''simple docstring''' from collections.abc import Generator def _UpperCamelCase ( ) -> Generator[int, None, None]: '''simple docstring''' snake_case ,snake_case : Tuple = 0, 1 while True: snake_case ,snake_case : List[Any] = b, a + b yield b def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ = 1000 ) -> int: '''simple docstring''' snake_case : Optional[int] = 1 snake_case : List[Any] = fibonacci_generator() while len(str(next(SCREAMING_SNAKE_CASE__ ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
83
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=99 , _lowerCamelCase=[1, 1, 2] , _lowerCamelCase=1 , _lowerCamelCase=32 , _lowerCamelCase=4 , _lowerCamelCase=8 , _lowerCamelCase=37 , _lowerCamelCase="gelu_new" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=0.0 , _lowerCamelCase=512 , _lowerCamelCase=3 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=4 , _lowerCamelCase=None , _lowerCamelCase=False , ) -> Optional[Any]: A_ : Dict = parent A_ : Optional[Any] = batch_size A_ : Optional[int] = seq_length A_ : Union[str, Any] = is_training A_ : str = use_input_mask A_ : Optional[int] = use_token_type_ids A_ : Tuple = use_labels A_ : Any = vocab_size A_ : Dict = block_sizes A_ : List[Any] = num_decoder_layers A_ : Union[str, Any] = d_model A_ : str = n_head A_ : List[Any] = d_head A_ : Union[str, Any] = d_inner A_ : List[str] = hidden_act A_ : List[str] = hidden_dropout A_ : Tuple = attention_dropout A_ : Dict = activation_dropout A_ : Tuple = max_position_embeddings A_ : List[str] = type_vocab_size A_ : List[Any] = 2 A_ : str = num_labels A_ : int = num_choices A_ : Optional[int] = scope A_ : Any = initializer_std # Used in the tests to check the size of the first attention layer A_ : int = n_head # Used in the tests to check the size of the first hidden state A_ : Optional[int] = self.d_model # Used in the tests to check the number of output hidden states/attentions A_ : Dict = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: A_ : List[Any] = self.num_hidden_layers + 2 def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : Tuple = None if self.use_input_mask: A_ : Any = random_attention_mask([self.batch_size, self.seq_length] ) A_ : Any = None if self.use_token_type_ids: A_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A_ : str = None A_ : Optional[Any] = None A_ : List[Any] = None if self.use_labels: A_ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : Tuple = ids_tensor([self.batch_size] , self.num_choices ) A_ : int = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) -> Optional[Any]: A_ : Any = TFFunnelModel(config=__UpperCAmelCase ) A_ : Tuple = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} A_ : Tuple = model(__UpperCAmelCase ) A_ : List[str] = [input_ids, input_mask] A_ : Any = model(__UpperCAmelCase ) A_ : Dict = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) A_ : List[str] = False A_ : Tuple = TFFunnelModel(config=__UpperCAmelCase ) A_ : List[str] = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) A_ : Optional[Any] = False A_ : str = TFFunnelModel(config=__UpperCAmelCase ) A_ : Optional[int] = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) -> Any: A_ : int = TFFunnelBaseModel(config=__UpperCAmelCase ) A_ : int = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} A_ : Optional[int] = model(__UpperCAmelCase ) A_ : List[str] = [input_ids, input_mask] A_ : List[Any] = model(__UpperCAmelCase ) A_ : Any = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) A_ : int = False A_ : Tuple = TFFunnelBaseModel(config=__UpperCAmelCase ) A_ : Any = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) A_ : int = False A_ : Any = TFFunnelBaseModel(config=__UpperCAmelCase ) A_ : Optional[int] = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) -> int: A_ : Optional[Any] = TFFunnelForPreTraining(config=__UpperCAmelCase ) A_ : int = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} A_ : List[Any] = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) -> Dict: A_ : List[Any] = TFFunnelForMaskedLM(config=__UpperCAmelCase ) A_ : List[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} A_ : int = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) -> int: A_ : str = self.num_labels A_ : Union[str, Any] = TFFunnelForSequenceClassification(config=__UpperCAmelCase ) A_ : int = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} A_ : int = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) -> List[str]: A_ : Tuple = self.num_choices A_ : Any = TFFunnelForMultipleChoice(config=__UpperCAmelCase ) A_ : List[str] = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) A_ : str = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) A_ : str = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) A_ : List[Any] = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } A_ : List[str] = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) -> Any: A_ : Optional[Any] = self.num_labels A_ : Optional[int] = TFFunnelForTokenClassification(config=__UpperCAmelCase ) A_ : str = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} A_ : Union[str, Any] = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ) -> Union[str, Any]: A_ : Optional[Any] = TFFunnelForQuestionAnswering(config=__UpperCAmelCase ) A_ : Any = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} A_ : Any = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : str = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ) : List[str] = config_and_inputs A_ : List[str] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class _lowerCAmelCase ( a__, a__, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) lowerCamelCase = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> int: A_ : str = TFFunnelModelTester(self ) A_ : Any = ConfigTester(self , config_class=__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> List[Any]: self.config_tester.run_common_tests() def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> List[Any]: A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) @require_tf class _lowerCAmelCase ( a__, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> str: A_ : List[str] = TFFunnelModelTester(self , base=__UpperCAmelCase ) A_ : List[Any] = ConfigTester(self , config_class=__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Dict: self.config_tester.run_common_tests() def UpperCAmelCase_ ( self ) -> str: A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase )
344
"""simple docstring""" from ..utils import DummyObject, requires_backends class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Optional[Any] = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> int: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[Any]: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> str: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : str = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> str: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Any = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[str]: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[Any]: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Union[str, Any] = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[str]: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Dict: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Dict = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Optional[Any] = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[Any]: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> int: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Union[str, Any] = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Any: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Any: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Any: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Union[str, Any] = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Union[str, Any] = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[Any]: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Union[str, Any] = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> str: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[Any]: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Tuple = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Dict: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> str: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Optional[Any] = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Any: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> str: requires_backends(cls , ['''flax'''] ) class __lowerCamelCase ( metaclass=a__ ): '''simple docstring''' A_ : Any = ['flax'] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[Any]: requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Any: requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Dict: requires_backends(cls , ['''flax'''] )
320
0
"""simple docstring""" # flake8: noqa # Lint as: python3 from typing import Dict, List, Optional, Type from .. import config from ..utils import logging from .formatting import ( ArrowFormatter, CustomFormatter, Formatter, PandasFormatter, PythonFormatter, TensorFormatter, format_table, query_table, ) from .np_formatter import NumpyFormatter _a : List[Any]= logging.get_logger(__name__) _a : Dict[Optional[str], Type[Formatter]]= {} _a : Dict[Optional[str], str]= {} _a : Dict[Optional[str], Exception]= {} def __UpperCAmelCase ( UpperCAmelCase_ : type , UpperCAmelCase_ : Optional[str] , UpperCAmelCase_ : Optional[List[str]] = None , ) -> Optional[int]: '''simple docstring''' __snake_case : Optional[Any] = aliases if aliases is not None else [] if format_type in _FORMAT_TYPES: logger.warning( F"Overwriting format type '{format_type}' ({_FORMAT_TYPES[format_type].__name__} -> {formatter_cls.__name__})" ) __snake_case : List[Any] = formatter_cls for alias in set(aliases + [format_type] ): if alias in _FORMAT_TYPES_ALIASES: logger.warning( F"Overwriting format type alias '{alias}' ({_FORMAT_TYPES_ALIASES[alias]} -> {format_type})" ) __snake_case : Tuple = format_type def __UpperCAmelCase ( UpperCAmelCase_ : Exception , UpperCAmelCase_ : Optional[str] , UpperCAmelCase_ : Optional[List[str]] = None ) -> Tuple: '''simple docstring''' __snake_case : Optional[int] = aliases if aliases is not None else [] for alias in set(aliases + [format_type] ): __snake_case : str = unavailable_error # Here we define all the available formatting functions that can be used by `Dataset.set_format` _register_formatter(PythonFormatter, None, aliases=["python"]) _register_formatter(ArrowFormatter, "arrow", aliases=["pa", "pyarrow"]) _register_formatter(NumpyFormatter, "numpy", aliases=["np"]) _register_formatter(PandasFormatter, "pandas", aliases=["pd"]) _register_formatter(CustomFormatter, "custom") if config.TORCH_AVAILABLE: from .torch_formatter import TorchFormatter _register_formatter(TorchFormatter, "torch", aliases=["pt", "pytorch"]) else: _a : Union[str, Any]= ValueError("PyTorch needs to be installed to be able to return PyTorch tensors.") _register_unavailable_formatter(_torch_error, "torch", aliases=["pt", "pytorch"]) if config.TF_AVAILABLE: from .tf_formatter import TFFormatter _register_formatter(TFFormatter, "tensorflow", aliases=["tf"]) else: _a : Optional[Any]= ValueError("Tensorflow needs to be installed to be able to return Tensorflow tensors.") _register_unavailable_formatter(_tf_error, "tensorflow", aliases=["tf"]) if config.JAX_AVAILABLE: from .jax_formatter import JaxFormatter _register_formatter(JaxFormatter, "jax", aliases=[]) else: _a : Tuple= ValueError("JAX needs to be installed to be able to return JAX arrays.") _register_unavailable_formatter(_jax_error, "jax", aliases=[]) def __UpperCAmelCase ( UpperCAmelCase_ : Optional[str] ) -> Optional[str]: '''simple docstring''' if format_type in _FORMAT_TYPES_ALIASES: return _FORMAT_TYPES_ALIASES[format_type] else: return format_type def __UpperCAmelCase ( UpperCAmelCase_ : Optional[str] , **UpperCAmelCase_ : Dict ) -> Formatter: '''simple docstring''' __snake_case : Union[str, Any] = get_format_type_from_alias(UpperCAmelCase_ ) if format_type in _FORMAT_TYPES: return _FORMAT_TYPES[format_type](**UpperCAmelCase_ ) if format_type in _FORMAT_TYPES_ALIASES_UNAVAILABLE: raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type] else: raise ValueError( F"Return type should be None or selected in {list(type for type in _FORMAT_TYPES.keys() if type != None )}, but got '{format_type}'" )
95
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig _a : Tuple= { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/config.json", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/config.json", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/config.json", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/config.json", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/config.json", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/config.json", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/config.json", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/config.json", } class UpperCamelCase ( lowercase ): UpperCAmelCase : str = """albert""" def __init__(self : Union[str, Any] , _A : int=3_00_00 , _A : Any=1_28 , _A : Tuple=40_96 , _A : int=12 , _A : Tuple=1 , _A : int=64 , _A : Optional[Any]=1_63_84 , _A : Optional[Any]=1 , _A : List[str]="gelu_new" , _A : Any=0 , _A : Optional[Any]=0 , _A : List[Any]=5_12 , _A : List[Any]=2 , _A : Dict=0.02 , _A : Union[str, Any]=1E-12 , _A : Tuple=0.1 , _A : int="absolute" , _A : List[Any]=0 , _A : str=2 , _A : int=3 , **_A : Tuple , ) -> Union[str, Any]: super().__init__(pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , **_A) __snake_case : str = vocab_size __snake_case : List[Any] = embedding_size __snake_case : Tuple = hidden_size __snake_case : Optional[Any] = num_hidden_layers __snake_case : List[str] = num_hidden_groups __snake_case : Dict = num_attention_heads __snake_case : Any = inner_group_num __snake_case : Union[str, Any] = hidden_act __snake_case : List[Any] = intermediate_size __snake_case : List[str] = hidden_dropout_prob __snake_case : int = attention_probs_dropout_prob __snake_case : Any = max_position_embeddings __snake_case : Optional[int] = type_vocab_size __snake_case : Optional[Any] = initializer_range __snake_case : List[Any] = layer_norm_eps __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : List[str] = position_embedding_type class UpperCamelCase ( lowercase ): @property def _lowercase (self : Tuple) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": __snake_case : int = {0: 'batch', 1: 'choice', 2: 'sequence'} else: __snake_case : Optional[Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis), ])
95
1
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging A =logging.get_logger(__name__) A ={ 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class _a ( __a ): __a : Optional[int] = """mvp""" __a : Dict = ["""past_key_values"""] __a : int = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[str] , lowercase : List[str]=50_267 , lowercase : List[str]=1_024 , lowercase : Any=12 , lowercase : Optional[int]=4_096 , lowercase : Tuple=16 , lowercase : Dict=12 , lowercase : List[str]=4_096 , lowercase : Union[str, Any]=16 , lowercase : str=0.0 , lowercase : List[str]=0.0 , lowercase : Union[str, Any]="gelu" , lowercase : int=1_024 , lowercase : Union[str, Any]=0.1 , lowercase : Union[str, Any]=0.0 , lowercase : Optional[Any]=0.0 , lowercase : Union[str, Any]=0.02 , lowercase : str=0.0 , lowercase : Optional[Any]=False , lowercase : Tuple=True , lowercase : Tuple=1 , lowercase : Optional[Any]=0 , lowercase : List[Any]=2 , lowercase : Dict=True , lowercase : int=2 , lowercase : Optional[int]=2 , lowercase : Any=False , lowercase : Tuple=100 , lowercase : int=800 , **lowercase : str , ): '''simple docstring''' UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = classifier_dropout UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = use_prompt UpperCAmelCase = prompt_length UpperCAmelCase = prompt_mid_dim super().__init__( pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , is_encoder_decoder=lowercase , decoder_start_token_id=lowercase , forced_eos_token_id=lowercase , **lowercase , ) if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' , lowercase ): UpperCAmelCase = self.bos_token_id warnings.warn( f"Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. " '''The config can simply be saved and uploaded again to be fixed.''' )
34
'''simple docstring''' from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal A =logging.get_logger(__name__) A =TypeVar('DatasetType', Dataset, IterableDataset) def snake_case_ (_a : List[DatasetType] , _a : Optional[List[float]] = None , _a : Optional[int] = None , _a : Optional[DatasetInfo] = None , _a : Optional[NamedSplit] = None , _a : Literal["first_exhausted", "all_exhausted"] = "first_exhausted" , ): from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError('''Unable to interleave an empty list of datasets.''' ) for i, dataset in enumerate(_a ): if not isinstance(_a , (Dataset, IterableDataset) ): if isinstance(_a , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " '''is an empty dataset dictionary.''' ) raise ValueError( F"Dataset at position {i} has at least one split: {list(_a )}\n" F"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(_a ) )}']" ) raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(_a ).__name__}." ) if i == 0: UpperCAmelCase , UpperCAmelCase = ( (Dataset, IterableDataset) if isinstance(_a , _a ) else (IterableDataset, Dataset) ) elif not isinstance(_a , _a ): raise ValueError( F"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(F"{stopping_strategy} is not supported. Please enter a valid stopping_strategy." ) if dataset_type is Dataset: return _interleave_map_style_datasets( _a , _a , _a , info=_a , split=_a , stopping_strategy=_a ) else: return _interleave_iterable_datasets( _a , _a , _a , info=_a , split=_a , stopping_strategy=_a ) def snake_case_ (_a : List[DatasetType] , _a : Optional[DatasetInfo] = None , _a : Optional[NamedSplit] = None , _a : int = 0 , ): if not dsets: raise ValueError('''Unable to concatenate an empty list of datasets.''' ) for i, dataset in enumerate(_a ): if not isinstance(_a , (Dataset, IterableDataset) ): if isinstance(_a , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " '''is an empty dataset dictionary.''' ) raise ValueError( F"Dataset at position {i} has at least one split: {list(_a )}\n" F"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(_a ) )}']" ) raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(_a ).__name__}." ) if i == 0: UpperCAmelCase , UpperCAmelCase = ( (Dataset, IterableDataset) if isinstance(_a , _a ) else (IterableDataset, Dataset) ) elif not isinstance(_a , _a ): raise ValueError( F"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if dataset_type is Dataset: return _concatenate_map_style_datasets(_a , info=_a , split=_a , axis=_a ) else: return _concatenate_iterable_datasets(_a , info=_a , split=_a , axis=_a )
34
1
"""simple docstring""" import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def _snake_case ( UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : int=1 ): if n_shave_prefix_segments >= 0: return ".".join(path.split(""".""" )[n_shave_prefix_segments:] ) else: return ".".join(path.split(""".""" )[:n_shave_prefix_segments] ) def _snake_case ( UpperCAmelCase_ : Dict , UpperCAmelCase_ : List[Any]=0 ): A__ = [] for old_item in old_list: A__ = old_item.replace("""in_layers.0""" , """norm1""" ) A__ = new_item.replace("""in_layers.2""" , """conv1""" ) A__ = new_item.replace("""out_layers.0""" , """norm2""" ) A__ = new_item.replace("""out_layers.3""" , """conv2""" ) A__ = new_item.replace("""emb_layers.1""" , """time_emb_proj""" ) A__ = new_item.replace("""skip_connection""" , """conv_shortcut""" ) A__ = shave_segments(UpperCAmelCase_ , n_shave_prefix_segments=UpperCAmelCase_ ) mapping.append({"""old""": old_item, """new""": new_item} ) return mapping def _snake_case ( UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any]=0 ): A__ = [] for old_item in old_list: A__ = old_item A__ = new_item.replace("""norm.weight""" , """group_norm.weight""" ) A__ = new_item.replace("""norm.bias""" , """group_norm.bias""" ) A__ = new_item.replace("""proj_out.weight""" , """proj_attn.weight""" ) A__ = new_item.replace("""proj_out.bias""" , """proj_attn.bias""" ) A__ = shave_segments(UpperCAmelCase_ , n_shave_prefix_segments=UpperCAmelCase_ ) mapping.append({"""old""": old_item, """new""": new_item} ) return mapping def _snake_case ( UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Tuple=None , UpperCAmelCase_ : List[Any]=None , UpperCAmelCase_ : Optional[Any]=None ): assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ), "Paths should be a list of dicts containing 'old' and 'new' keys." # Splits the attention layers into three variables. if attention_paths_to_split is not None: for path, path_map in attention_paths_to_split.items(): A__ = old_checkpoint[path] A__ = old_tensor.shape[0] // 3 A__ = (-1, channels) if len(old_tensor.shape ) == 3 else (-1) A__ = old_tensor.shape[0] // config["""num_head_channels"""] // 3 A__ = old_tensor.reshape((num_heads, 3 * channels // num_heads) + old_tensor.shape[1:] ) A__ , A__ , A__ = old_tensor.split(channels // num_heads , dim=1 ) A__ = query.reshape(UpperCAmelCase_ ) A__ = key.reshape(UpperCAmelCase_ ) A__ = value.reshape(UpperCAmelCase_ ) for path in paths: A__ = path["""new"""] # These have already been assigned if attention_paths_to_split is not None and new_path in attention_paths_to_split: continue # Global renaming happens here A__ = new_path.replace("""middle_block.0""" , """mid_block.resnets.0""" ) A__ = new_path.replace("""middle_block.1""" , """mid_block.attentions.0""" ) A__ = new_path.replace("""middle_block.2""" , """mid_block.resnets.1""" ) if additional_replacements is not None: for replacement in additional_replacements: A__ = new_path.replace(replacement["""old"""] , replacement["""new"""] ) # proj_attn.weight has to be converted from conv 1D to linear if "proj_attn.weight" in new_path: A__ = old_checkpoint[path["""old"""]][:, :, 0] else: A__ = old_checkpoint[path["""old"""]] def _snake_case ( UpperCAmelCase_ : Any , UpperCAmelCase_ : List[str] ): A__ = {} A__ = checkpoint["""time_embed.0.weight"""] A__ = checkpoint["""time_embed.0.bias"""] A__ = checkpoint["""time_embed.2.weight"""] A__ = checkpoint["""time_embed.2.bias"""] A__ = checkpoint["""input_blocks.0.0.weight"""] A__ = checkpoint["""input_blocks.0.0.bias"""] A__ = checkpoint["""out.0.weight"""] A__ = checkpoint["""out.0.bias"""] A__ = checkpoint["""out.2.weight"""] A__ = checkpoint["""out.2.bias"""] # Retrieves the keys for the input blocks only A__ = len({""".""".join(layer.split(""".""" )[:2] ) for layer in checkpoint if """input_blocks""" in layer} ) A__ = { layer_id: [key for key in checkpoint if F"""input_blocks.{layer_id}""" in key] for layer_id in range(UpperCAmelCase_ ) } # Retrieves the keys for the middle blocks only A__ = len({""".""".join(layer.split(""".""" )[:2] ) for layer in checkpoint if """middle_block""" in layer} ) A__ = { layer_id: [key for key in checkpoint if F"""middle_block.{layer_id}""" in key] for layer_id in range(UpperCAmelCase_ ) } # Retrieves the keys for the output blocks only A__ = len({""".""".join(layer.split(""".""" )[:2] ) for layer in checkpoint if """output_blocks""" in layer} ) A__ = { layer_id: [key for key in checkpoint if F"""output_blocks.{layer_id}""" in key] for layer_id in range(UpperCAmelCase_ ) } for i in range(1 , UpperCAmelCase_ ): A__ = (i - 1) // (config["""num_res_blocks"""] + 1) A__ = (i - 1) % (config["""num_res_blocks"""] + 1) A__ = [key for key in input_blocks[i] if F"""input_blocks.{i}.0""" in key] A__ = [key for key in input_blocks[i] if F"""input_blocks.{i}.1""" in key] if F"""input_blocks.{i}.0.op.weight""" in checkpoint: A__ = checkpoint[ F"""input_blocks.{i}.0.op.weight""" ] A__ = checkpoint[ F"""input_blocks.{i}.0.op.bias""" ] continue A__ = renew_resnet_paths(UpperCAmelCase_ ) A__ = {"""old""": F"""input_blocks.{i}.0""", """new""": F"""down_blocks.{block_id}.resnets.{layer_in_block_id}"""} A__ = {"""old""": """resnets.2.op""", """new""": """downsamplers.0.op"""} assign_to_checkpoint( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , additional_replacements=[meta_path, resnet_op] , config=UpperCAmelCase_ ) if len(UpperCAmelCase_ ): A__ = renew_attention_paths(UpperCAmelCase_ ) A__ = { """old""": F"""input_blocks.{i}.1""", """new""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}""", } A__ = { F"""input_blocks.{i}.1.qkv.bias""": { """key""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", """query""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", """value""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, F"""input_blocks.{i}.1.qkv.weight""": { """key""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", """query""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", """value""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , additional_replacements=[meta_path] , attention_paths_to_split=UpperCAmelCase_ , config=UpperCAmelCase_ , ) A__ = middle_blocks[0] A__ = middle_blocks[1] A__ = middle_blocks[2] A__ = renew_resnet_paths(UpperCAmelCase_ ) assign_to_checkpoint(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , config=UpperCAmelCase_ ) A__ = renew_resnet_paths(UpperCAmelCase_ ) assign_to_checkpoint(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , config=UpperCAmelCase_ ) A__ = renew_attention_paths(UpperCAmelCase_ ) A__ = { """middle_block.1.qkv.bias""": { """key""": """mid_block.attentions.0.key.bias""", """query""": """mid_block.attentions.0.query.bias""", """value""": """mid_block.attentions.0.value.bias""", }, """middle_block.1.qkv.weight""": { """key""": """mid_block.attentions.0.key.weight""", """query""": """mid_block.attentions.0.query.weight""", """value""": """mid_block.attentions.0.value.weight""", }, } assign_to_checkpoint( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , attention_paths_to_split=UpperCAmelCase_ , config=UpperCAmelCase_ ) for i in range(UpperCAmelCase_ ): A__ = i // (config["""num_res_blocks"""] + 1) A__ = i % (config["""num_res_blocks"""] + 1) A__ = [shave_segments(UpperCAmelCase_ , 2 ) for name in output_blocks[i]] A__ = {} for layer in output_block_layers: A__ , A__ = layer.split(""".""" )[0], shave_segments(UpperCAmelCase_ , 1 ) if layer_id in output_block_list: output_block_list[layer_id].append(UpperCAmelCase_ ) else: A__ = [layer_name] if len(UpperCAmelCase_ ) > 1: A__ = [key for key in output_blocks[i] if F"""output_blocks.{i}.0""" in key] A__ = [key for key in output_blocks[i] if F"""output_blocks.{i}.1""" in key] A__ = renew_resnet_paths(UpperCAmelCase_ ) A__ = renew_resnet_paths(UpperCAmelCase_ ) A__ = {"""old""": F"""output_blocks.{i}.0""", """new""": F"""up_blocks.{block_id}.resnets.{layer_in_block_id}"""} assign_to_checkpoint(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , additional_replacements=[meta_path] , config=UpperCAmelCase_ ) if ["conv.weight", "conv.bias"] in output_block_list.values(): A__ = list(output_block_list.values() ).index(["""conv.weight""", """conv.bias"""] ) A__ = checkpoint[ F"""output_blocks.{i}.{index}.conv.weight""" ] A__ = checkpoint[ F"""output_blocks.{i}.{index}.conv.bias""" ] # Clear attentions as they have been attributed above. if len(UpperCAmelCase_ ) == 2: A__ = [] if len(UpperCAmelCase_ ): A__ = renew_attention_paths(UpperCAmelCase_ ) A__ = { """old""": F"""output_blocks.{i}.1""", """new""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}""", } A__ = { F"""output_blocks.{i}.1.qkv.bias""": { """key""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", """query""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", """value""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, F"""output_blocks.{i}.1.qkv.weight""": { """key""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", """query""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", """value""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , additional_replacements=[meta_path] , attention_paths_to_split=to_split if any("""qkv""" in key for key in attentions ) else None , config=UpperCAmelCase_ , ) else: A__ = renew_resnet_paths(UpperCAmelCase_ , n_shave_prefix_segments=1 ) for path in resnet_0_paths: A__ = """.""".join(["""output_blocks""", str(UpperCAmelCase_ ), path["""old"""]] ) A__ = """.""".join(["""up_blocks""", str(UpperCAmelCase_ ), """resnets""", str(UpperCAmelCase_ ), path["""new"""]] ) A__ = checkpoint[old_path] return new_checkpoint if __name__ == "__main__": SCREAMING_SNAKE_CASE_ : int = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the architecture.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') SCREAMING_SNAKE_CASE_ : List[Any] = parser.parse_args() SCREAMING_SNAKE_CASE_ : Any = torch.load(args.checkpoint_path) with open(args.config_file) as f: SCREAMING_SNAKE_CASE_ : Union[str, Any] = json.loads(f.read()) SCREAMING_SNAKE_CASE_ : Tuple = convert_ldm_checkpoint(checkpoint, config) if "ldm" in config: del config["ldm"] SCREAMING_SNAKE_CASE_ : Optional[int] = UNetaDModel(**config) model.load_state_dict(converted_checkpoint) try: SCREAMING_SNAKE_CASE_ : Any = DDPMScheduler.from_config('/'.join(args.checkpoint_path.split('/')[:-1])) SCREAMING_SNAKE_CASE_ : Union[str, Any] = VQModel.from_pretrained('/'.join(args.checkpoint_path.split('/')[:-1])) SCREAMING_SNAKE_CASE_ : Optional[int] = LDMPipeline(unet=model, scheduler=scheduler, vae=vqvae) pipe.save_pretrained(args.dump_path) except: # noqa: E722 model.save_pretrained(args.dump_path)
360
"""simple docstring""" from collections import UserDict from typing import Union import numpy as np import requests from ..utils import ( add_end_docstrings, logging, ) from .audio_classification import ffmpeg_read from .base import PIPELINE_INIT_ARGS, Pipeline SCREAMING_SNAKE_CASE_ : Tuple = logging.get_logger(__name__) @add_end_docstrings(_lowerCamelCase ) class a ( _lowerCamelCase ): """simple docstring""" def __init__( self: Any , **UpperCamelCase: int ): """simple docstring""" super().__init__(**UpperCamelCase ) if self.framework != "pt": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) # No specific FOR_XXX available yet def __call__( self: Tuple , UpperCamelCase: Union[np.ndarray, bytes, str] , **UpperCamelCase: Tuple ): """simple docstring""" return super().__call__(UpperCamelCase , **UpperCamelCase ) def UpperCamelCase ( self: Tuple , **UpperCamelCase: Union[str, Any] ): """simple docstring""" A__ = {} if "candidate_labels" in kwargs: A__ = kwargs["""candidate_labels"""] if "hypothesis_template" in kwargs: A__ = kwargs["""hypothesis_template"""] return preprocess_params, {}, {} def UpperCamelCase ( self: Optional[int] , UpperCamelCase: Tuple , UpperCamelCase: Tuple=None , UpperCamelCase: Optional[Any]="This is a sound of {}." ): """simple docstring""" if isinstance(UpperCamelCase , UpperCamelCase ): if audio.startswith("""http://""" ) or audio.startswith("""https://""" ): # We need to actually check for a real protocol, otherwise it's impossible to use a local file # like http_huggingface_co.png A__ = requests.get(UpperCamelCase ).content else: with open(UpperCamelCase , """rb""" ) as f: A__ = f.read() if isinstance(UpperCamelCase , UpperCamelCase ): A__ = ffmpeg_read(UpperCamelCase , self.feature_extractor.sampling_rate ) if not isinstance(UpperCamelCase , np.ndarray ): raise ValueError("""We expect a numpy ndarray as input""" ) if len(audio.shape ) != 1: raise ValueError("""We expect a single channel audio input for ZeroShotAudioClassificationPipeline""" ) A__ = self.feature_extractor( [audio] , sampling_rate=self.feature_extractor.sampling_rate , return_tensors="""pt""" ) A__ = candidate_labels A__ = [hypothesis_template.format(UpperCamelCase ) for x in candidate_labels] A__ = self.tokenizer(UpperCamelCase , return_tensors=self.framework , padding=UpperCamelCase ) A__ = [text_inputs] return inputs def UpperCamelCase ( self: Any , UpperCamelCase: Union[str, Any] ): """simple docstring""" A__ = model_inputs.pop("""candidate_labels""" ) A__ = model_inputs.pop("""text_inputs""" ) if isinstance(text_inputs[0] , UpperCamelCase ): A__ = text_inputs[0] else: # Batching case. A__ = text_inputs[0][0] A__ = self.model(**UpperCamelCase , **UpperCamelCase ) A__ = { """candidate_labels""": candidate_labels, """logits""": outputs.logits_per_audio, } return model_outputs def UpperCamelCase ( self: Any , UpperCamelCase: Optional[int] ): """simple docstring""" A__ = model_outputs.pop("""candidate_labels""" ) A__ = model_outputs["""logits"""][0] if self.framework == "pt": A__ = logits.softmax(dim=0 ) A__ = probs.tolist() else: raise ValueError("""`tf` framework not supported.""" ) A__ = [ {"""score""": score, """label""": candidate_label} for score, candidate_label in sorted(zip(UpperCamelCase , UpperCamelCase ) , key=lambda UpperCamelCase : -x[0] ) ] return result
69
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING _snake_case = logging.get_logger(__name__) class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' __A : List[str] = "upernet" def __init__( self , __A=None , __A=512 , __A=0.02 , __A=[1, 2, 3, 6] , __A=True , __A=0.4 , __A=384 , __A=256 , __A=1 , __A=False , __A=255 , **__A , ): """simple docstring""" super().__init__(**__A ) if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) lowerCamelCase : Dict = CONFIG_MAPPING["resnet"](out_features=["stage1", "stage2", "stage3", "stage4"] ) elif isinstance(__A , __A ): lowerCamelCase : Tuple = backbone_config.get("model_type" ) lowerCamelCase : List[str] = CONFIG_MAPPING[backbone_model_type] lowerCamelCase : Tuple = config_class.from_dict(__A ) lowerCamelCase : Any = backbone_config lowerCamelCase : Optional[Any] = hidden_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : Dict = pool_scales lowerCamelCase : List[str] = use_auxiliary_head lowerCamelCase : List[str] = auxiliary_loss_weight lowerCamelCase : List[Any] = auxiliary_in_channels lowerCamelCase : str = auxiliary_channels lowerCamelCase : str = auxiliary_num_convs lowerCamelCase : Optional[int] = auxiliary_concat_input lowerCamelCase : str = loss_ignore_index def _snake_case ( self ): """simple docstring""" lowerCamelCase : Any = copy.deepcopy(self.__dict__ ) lowerCamelCase : Union[str, Any] = self.backbone_config.to_dict() lowerCamelCase : Any = self.__class__.model_type return output
283
import argparse _snake_case = '''docs/source/_static/js/custom.js''' def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE_ , encoding="utf-8" , newline="\n" ) as f: lowerCamelCase : List[str] = f.readlines() lowerCamelCase : int = 0 # First let's put the right version while not lines[index].startswith("const stableVersion =" ): index += 1 lowerCamelCase : str = f"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("const versionMapping = {" ): index += 1 # We go until the end while not lines[index].startswith("}" ): index += 1 # We add the new version at the end lines[index - 1] += f""" \"v{version}\": \"v{version}\",\n""" with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument('''--version''', help='''Release version.''') _snake_case = parser.parse_args() update_custom_js(args.version)
283
1
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": _lowerCAmelCase : Optional[Any] = pd.read_csv("sample_data.csv", header=None) _lowerCAmelCase : Optional[Any] = df.shape[:1][0] # If you're using some other dataset input the target column _lowerCAmelCase : Any = df.iloc[:, 1:2] _lowerCAmelCase : Union[str, Any] = actual_data.values.reshape(len_data, 1) _lowerCAmelCase : Optional[Any] = MinMaxScaler().fit_transform(actual_data) _lowerCAmelCase : Tuple = 10 _lowerCAmelCase : List[str] = 5 _lowerCAmelCase : int = 20 _lowerCAmelCase : Union[str, Any] = len_data - periods * look_back _lowerCAmelCase : int = actual_data[:division] _lowerCAmelCase : str = actual_data[division - look_back :] _lowerCAmelCase , _lowerCAmelCase : Dict = [], [] _lowerCAmelCase , _lowerCAmelCase : Any = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) _lowerCAmelCase : List[str] = np.array(train_x) _lowerCAmelCase : str = np.array(test_x) _lowerCAmelCase : str = np.array([list(i.ravel()) for i in train_y]) _lowerCAmelCase : str = np.array([list(i.ravel()) for i in test_y]) _lowerCAmelCase : List[Any] = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="mean_squared_error", optimizer="adam") _lowerCAmelCase : List[str] = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) _lowerCAmelCase : Dict = model.predict(x_test)
362
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _lowerCAmelCase : int = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Dict = ["BartphoTokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys _lowerCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
308
0
import warnings from ...utils import logging from .image_processing_deformable_detr import DeformableDetrImageProcessor SCREAMING_SNAKE_CASE :List[Any] = logging.get_logger(__name__) class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Optional[Any] ,*A : List[Any] ,**A : int ): warnings.warn( "The class DeformableDetrFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use DeformableDetrImageProcessor instead." ,A ,) super().__init__(*A ,**A )
15
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowerCAmelCase :str = { 'configuration_squeezebert': [ 'SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SqueezeBertConfig', 'SqueezeBertOnnxConfig', ], 'tokenization_squeezebert': ['SqueezeBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase :Optional[int] = ['SqueezeBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase :str = [ 'SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'SqueezeBertForMaskedLM', 'SqueezeBertForMultipleChoice', 'SqueezeBertForQuestionAnswering', 'SqueezeBertForSequenceClassification', 'SqueezeBertForTokenClassification', 'SqueezeBertModel', 'SqueezeBertModule', 'SqueezeBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys _lowerCAmelCase :Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
263
0
"""simple docstring""" import argparse import datetime def _A ( lowercase ): """simple docstring""" a ={ '''0''': '''Sunday''', '''1''': '''Monday''', '''2''': '''Tuesday''', '''3''': '''Wednesday''', '''4''': '''Thursday''', '''5''': '''Friday''', '''6''': '''Saturday''', } a ={0: 1, 1: 2, 2: 3, 3: 4, 4: 5, 5: 6, 6: 0} # Validate if not 0 < len(lowercase ) < 11: raise ValueError('''Must be 10 characters long''' ) # Get month a =int(date_input[0] + date_input[1] ) # Validate if not 0 < m < 13: raise ValueError('''Month must be between 1 - 12''' ) a =date_input[2] # Validate if sep_a not in ["-", "/"]: raise ValueError('''Date separator must be \'-\' or \'/\'''' ) # Get day a =int(date_input[3] + date_input[4] ) # Validate if not 0 < d < 32: raise ValueError('''Date must be between 1 - 31''' ) # Get second separator a =date_input[5] # Validate if sep_a not in ["-", "/"]: raise ValueError('''Date separator must be \'-\' or \'/\'''' ) # Get year a =int(date_input[6] + date_input[7] + date_input[8] + date_input[9] ) # Arbitrary year range if not 45 < y < 85_00: raise ValueError( '''Year out of range. There has to be some sort of limit...right?''' ) # Get datetime obj for validation a =datetime.date(int(lowercase ) , int(lowercase ) , int(lowercase ) ) # Start math if m <= 2: a =y - 1 a =m + 12 # maths var a =int(str(lowercase )[:2] ) a =int(str(lowercase )[2:] ) a =int(2.6 * m - 5.39 ) a =int(c / 4 ) a =int(k / 4 ) a =int(d + k ) a =int(t + u + v + x ) a =int(z - (2 * c) ) a =round(w % 7 ) # End math # Validate math if f != convert_datetime_days[dt_ck.weekday()]: raise AssertionError('''The date was evaluated incorrectly. Contact developer.''' ) # Response a =f'''Your date {date_input}, is a {days[str(lowercase )]}!''' return response if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase_ : Optional[Any] = argparse.ArgumentParser( description=( """Find out what day of the week nearly any date is or was. Enter """ """date as a string in the mm-dd-yyyy or mm/dd/yyyy format""" ) ) parser.add_argument( """date_input""", type=str, help="""Date as a string (mm-dd-yyyy or mm/dd/yyyy)""" ) lowerCamelCase_ : int = parser.parse_args() zeller(args.date_input)
356
"""simple docstring""" # DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class __A : """simple docstring""" __lowerCAmelCase = 42 # setable values __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = None @classmethod def SCREAMING_SNAKE_CASE ( cls , __A , __A , __A ) -> List[str]: return cls(common=__A , init_noise_sigma=__A , timesteps=__A ) @dataclass class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" __lowerCAmelCase = 42 class __A ( _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE ): """simple docstring""" __lowerCAmelCase = [e.name for e in FlaxKarrasDiffusionSchedulers] __lowerCAmelCase = 42 @property def SCREAMING_SNAKE_CASE ( self ) -> str: return True @register_to_config def __init__( self , __A = 1000 , __A = 0.0_001 , __A = 0.02 , __A = "linear" , __A = None , __A = "fixed_small" , __A = True , __A = "epsilon" , __A = jnp.floataa , ) -> List[Any]: a =dtype def SCREAMING_SNAKE_CASE ( self , __A = None ) -> DDPMSchedulerState: if common is None: a =CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution a =jnp.array(1.0 , dtype=self.dtype ) a =jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=__A , init_noise_sigma=__A , timesteps=__A , ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = None ) -> jnp.ndarray: return sample def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = () ) -> DDPMSchedulerState: a =self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 a =(jnp.arange(0 , __A ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=__A , timesteps=__A , ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A=None , __A=None ) -> str: a =state.common.alphas_cumprod[t] a =jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample a =(1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: a =self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": a =jnp.clip(__A , a_min=1E-2_0 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": a =jnp.log(jnp.clip(__A , a_min=1E-2_0 ) ) elif variance_type == "fixed_large": a =state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log a =jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": a =variance a =state.common.betas[t] a =(predicted_variance + 1) / 2 a =frac * max_log + (1 - frac) * min_log return variance def SCREAMING_SNAKE_CASE ( self , __A , __A , __A , __A , __A = None , __A = True , ) -> Union[FlaxDDPMSchedulerOutput, Tuple]: a =timestep if key is None: a =jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: a , a =jnp.split(__A , sample.shape[1] , axis=1 ) else: a =None # 1. compute alphas, betas a =state.common.alphas_cumprod[t] a =jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) a =1 - alpha_prod_t a =1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": a =(sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": a =model_output elif self.config.prediction_type == "v_prediction": a =(alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' ''' for the FlaxDDPMScheduler.''' ) # 3. Clip "predicted x_0" if self.config.clip_sample: a =jnp.clip(__A , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf a =(alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t a =state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf a =pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): a =jax.random.split(__A , num=1 ) a =jax.random.normal(__A , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(__A , __A , predicted_variance=__A ) ** 0.5) * noise a =jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) a =pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=__A , state=__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A , __A , ) -> jnp.ndarray: return add_noise_common(state.common , __A , __A , __A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A , __A , ) -> jnp.ndarray: return get_velocity_common(state.common , __A , __A , __A ) def __len__( self ) -> Optional[int]: return self.config.num_train_timesteps
215
0
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def _A ( ): UpperCamelCase :Optional[int] = ArgumentParser( description=( '''PyTorch TPU distributed training launch ''' '''helper utility that will spawn up ''' '''multiple distributed processes''' ) ) # Optional arguments for the launch helper parser.add_argument('''--num_cores''' , type=SCREAMING_SNAKE_CASE__ , default=1 , help='''Number of TPU cores to use (1 or 8).''' ) # positional parser.add_argument( '''training_script''' , type=SCREAMING_SNAKE_CASE__ , help=( '''The full path to the single TPU training ''' '''program/script to be launched in parallel, ''' '''followed by all the arguments for the ''' '''training script''' ) , ) # rest from the training program parser.add_argument('''training_script_args''' , nargs=SCREAMING_SNAKE_CASE__ ) return parser.parse_args() def _A ( ): UpperCamelCase :Union[str, Any] = parse_args() # Import training_script as a module. UpperCamelCase :int = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) UpperCamelCase :Tuple = script_fpath.stem UpperCamelCase :Any = importlib.import_module(SCREAMING_SNAKE_CASE__ ) # Patch sys.argv UpperCamelCase :Optional[int] = [args.training_script] + args.training_script_args + ['''--tpu_num_cores''', str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
259
import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin __snake_case = 1E-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class UpperCAmelCase_ : """simple docstring""" def __init__( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=16 , SCREAMING_SNAKE_CASE_=13 , SCREAMING_SNAKE_CASE_=7 , SCREAMING_SNAKE_CASE_=14 , SCREAMING_SNAKE_CASE_=10 , SCREAMING_SNAKE_CASE_=19 , SCREAMING_SNAKE_CASE_=5 , SCREAMING_SNAKE_CASE_=4 , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=16 , SCREAMING_SNAKE_CASE_=2 , SCREAMING_SNAKE_CASE_=4 , SCREAMING_SNAKE_CASE_=4 , SCREAMING_SNAKE_CASE_="gelu" , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=[1, 2, 3, 4, 5] , SCREAMING_SNAKE_CASE_=25 , SCREAMING_SNAKE_CASE_=5 , ) -> str: UpperCamelCase :Any = d_model UpperCamelCase :List[str] = parent UpperCamelCase :List[Any] = batch_size UpperCamelCase :str = prediction_length UpperCamelCase :str = context_length UpperCamelCase :int = cardinality UpperCamelCase :Optional[Any] = num_time_features UpperCamelCase :Optional[Any] = lags_sequence UpperCamelCase :str = embedding_dimension UpperCamelCase :str = is_training UpperCamelCase :Optional[int] = hidden_size UpperCamelCase :List[Any] = num_hidden_layers UpperCamelCase :int = num_attention_heads UpperCamelCase :Tuple = intermediate_size UpperCamelCase :List[str] = hidden_act UpperCamelCase :List[str] = hidden_dropout_prob UpperCamelCase :List[Any] = attention_probs_dropout_prob UpperCamelCase :Optional[int] = context_length UpperCamelCase :Tuple = prediction_length + label_length UpperCamelCase :Optional[Any] = label_length UpperCamelCase :Optional[int] = moving_average UpperCamelCase :Union[str, Any] = autocorrelation_factor def UpperCAmelCase ( self ) -> Optional[int]: return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ ) -> List[str]: UpperCamelCase :Optional[Any] = config.context_length + max(config.lags_sequence ) UpperCamelCase :Union[str, Any] = ids_tensor([self.batch_size, 1] , config.cardinality[0] ) UpperCamelCase :List[str] = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) UpperCamelCase :Union[str, Any] = floats_tensor([self.batch_size, _past_length] ) UpperCamelCase :Any = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs UpperCamelCase :Tuple = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) UpperCamelCase :int = floats_tensor([self.batch_size, config.prediction_length] ) UpperCamelCase :Union[str, Any] = { '''past_values''': past_values, '''static_categorical_features''': static_categorical_features, '''past_time_features''': past_time_features, '''past_observed_mask''': past_observed_mask, '''future_time_features''': future_time_features, '''future_values''': future_values, } return inputs_dict def UpperCAmelCase ( self ) -> Tuple: UpperCamelCase :int = self.get_config() UpperCamelCase :Union[str, Any] = self.prepare_autoformer_inputs_dict(SCREAMING_SNAKE_CASE_ ) return config, inputs_dict def UpperCAmelCase ( self ) -> Any: UpperCamelCase , UpperCamelCase :Optional[int] = self.prepare_config_and_inputs() return config, inputs_dict def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: UpperCamelCase :int = AutoformerModel(config=SCREAMING_SNAKE_CASE_ ).to(SCREAMING_SNAKE_CASE_ ).eval() UpperCamelCase :Any = model(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase :str = outputs.encoder_last_hidden_state UpperCamelCase :str = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: UpperCamelCase :Any = model.get_encoder() encoder.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Any = AutoformerEncoder.from_pretrained(SCREAMING_SNAKE_CASE_ ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase :int = model.create_network_inputs(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase , UpperCamelCase :Tuple = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) UpperCamelCase :Tuple = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) UpperCamelCase :Optional[Any] = encoder(inputs_embeds=SCREAMING_SNAKE_CASE_ )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) UpperCamelCase :Optional[Any] = ( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) UpperCamelCase :Union[str, Any] = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) UpperCamelCase :Tuple = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) UpperCamelCase :Optional[Any] = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCamelCase :Union[str, Any] = model.get_decoder() decoder.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Optional[Any] = AutoformerDecoder.from_pretrained(SCREAMING_SNAKE_CASE_ ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :str = decoder( trend=SCREAMING_SNAKE_CASE_ , inputs_embeds=SCREAMING_SNAKE_CASE_ , encoder_hidden_states=SCREAMING_SNAKE_CASE_ , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class UpperCAmelCase_ ( lowercase, lowercase, unittest.TestCase ): """simple docstring""" UpperCamelCase_ : List[str] =(AutoformerModel, AutoformerForPrediction) if is_torch_available() else () UpperCamelCase_ : List[str] =(AutoformerForPrediction,) if is_torch_available() else () UpperCamelCase_ : Optional[Any] ={'feature-extraction': AutoformerModel} if is_torch_available() else {} UpperCamelCase_ : Any =False UpperCamelCase_ : List[str] =False UpperCamelCase_ : Dict =False UpperCamelCase_ : Dict =False UpperCamelCase_ : int =False UpperCamelCase_ : Optional[int] =False def UpperCAmelCase ( self ) -> Tuple: UpperCamelCase :str = AutoformerModelTester(self ) UpperCamelCase :int = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE_ , has_text_modality=SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self ) -> Union[str, Any]: self.config_tester.run_common_tests() def UpperCAmelCase ( self ) -> Optional[Any]: UpperCamelCase , UpperCamelCase :str = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: UpperCamelCase :Optional[int] = model_class(SCREAMING_SNAKE_CASE_ ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCamelCase , UpperCamelCase :List[str] = model_class.from_pretrained(SCREAMING_SNAKE_CASE_ , output_loading_info=SCREAMING_SNAKE_CASE_ ) self.assertEqual(info['''missing_keys'''] , [] ) def UpperCAmelCase ( self ) -> Dict: UpperCamelCase :Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*SCREAMING_SNAKE_CASE_ ) @unittest.skip(reason='''Model has no tokens embeddings''' ) def UpperCAmelCase ( self ) -> int: pass def UpperCAmelCase ( self ) -> Tuple: UpperCamelCase :str = inspect.signature(getattr(SCREAMING_SNAKE_CASE_ , '''forward''' ) ) # The main input is the name of the argument after `self` UpperCamelCase :List[str] = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self ) -> Tuple: UpperCamelCase , UpperCamelCase :Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase :List[Any] = model_class(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase :Tuple = [*signature.parameters.keys()] UpperCamelCase :Optional[Any] = [ '''past_values''', '''past_time_features''', '''past_observed_mask''', '''static_categorical_features''', '''static_real_features''', '''future_values''', '''future_time_features''', ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append('''future_observed_mask''' ) expected_arg_names.extend( [ '''decoder_attention_mask''', '''head_mask''', '''decoder_head_mask''', '''cross_attn_head_mask''', '''encoder_outputs''', '''past_key_values''', '''output_hidden_states''', '''output_attentions''', '''use_cache''', '''return_dict''', ] ) self.assertListEqual(arg_names[: len(SCREAMING_SNAKE_CASE_ )] , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self ) -> Dict: UpperCamelCase , UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase :Dict = True UpperCamelCase :Dict = getattr(self.model_tester , '''seq_length''' , SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Any = getattr(self.model_tester , '''decoder_seq_length''' , SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Union[str, Any] = getattr(self.model_tester , '''encoder_seq_length''' , SCREAMING_SNAKE_CASE_ ) UpperCamelCase :int = getattr(self.model_tester , '''d_model''' , SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Tuple = getattr(self.model_tester , '''num_attention_heads''' , SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Any = d_model // num_attention_heads for model_class in self.all_model_classes: UpperCamelCase :Tuple = True UpperCamelCase :Tuple = False UpperCamelCase :Any = True UpperCamelCase :List[Any] = model_class(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() with torch.no_grad(): UpperCamelCase :int = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) UpperCamelCase :Union[str, Any] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCamelCase :Dict = True UpperCamelCase :Union[str, Any] = model_class(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() with torch.no_grad(): UpperCamelCase :Optional[Any] = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) UpperCamelCase :List[str] = outputs.encoder_attentions self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) UpperCamelCase :List[str] = len(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :List[Any] = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # decoder attentions UpperCamelCase :Union[str, Any] = outputs.decoder_attentions self.assertIsInstance(SCREAMING_SNAKE_CASE_ , (list, tuple) ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions UpperCamelCase :Union[str, Any] = outputs.cross_attentions self.assertIsInstance(SCREAMING_SNAKE_CASE_ , (list, tuple) ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine UpperCamelCase :Any = True UpperCamelCase :int = True UpperCamelCase :Any = model_class(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() with torch.no_grad(): UpperCamelCase :Optional[Any] = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) self.assertEqual(out_len + 2 , len(SCREAMING_SNAKE_CASE_ ) ) UpperCamelCase :List[str] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def UpperCAmelCase ( self ) -> List[Any]: super().test_retain_grad_hidden_states_attentions() def _A ( SCREAMING_SNAKE_CASE__ : int="train-batch.pt" ): UpperCamelCase :Union[str, Any] = hf_hub_download(repo_id='''hf-internal-testing/tourism-monthly-batch''' , filename=SCREAMING_SNAKE_CASE__ , repo_type='''dataset''' ) UpperCamelCase :Tuple = torch.load(SCREAMING_SNAKE_CASE__ , map_location=SCREAMING_SNAKE_CASE__ ) return batch @require_torch @slow class UpperCAmelCase_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> List[Any]: UpperCamelCase :int = AutoformerModel.from_pretrained('''huggingface/autoformer-tourism-monthly''' ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Dict = prepare_batch() with torch.no_grad(): UpperCamelCase :Optional[Any] = model( past_values=batch['''past_values'''] , past_time_features=batch['''past_time_features'''] , past_observed_mask=batch['''past_observed_mask'''] , static_categorical_features=batch['''static_categorical_features'''] , future_values=batch['''future_values'''] , future_time_features=batch['''future_time_features'''] , )[0] UpperCamelCase :Union[str, Any] = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Tuple = torch.tensor( [[0.3593, -1.3398, 0.6330], [0.2279, 1.5396, -0.1792], [0.0450, 1.3225, -0.2335]] , device=SCREAMING_SNAKE_CASE_ ) self.assertTrue(torch.allclose(output[0, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=SCREAMING_SNAKE_CASE_ ) ) def UpperCAmelCase ( self ) -> Union[str, Any]: UpperCamelCase :Any = AutoformerForPrediction.from_pretrained('''huggingface/autoformer-tourism-monthly''' ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Union[str, Any] = prepare_batch('''val-batch.pt''' ) with torch.no_grad(): UpperCamelCase :Dict = model( past_values=batch['''past_values'''] , past_time_features=batch['''past_time_features'''] , past_observed_mask=batch['''past_observed_mask'''] , static_categorical_features=batch['''static_categorical_features'''] , ).encoder_last_hidden_state UpperCamelCase :Union[str, Any] = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Any = torch.tensor( [[-0.0734, -0.9036, 0.8358], [4.7186, 2.4113, 1.9581], [1.7953, 2.3558, 1.2970]] , device=SCREAMING_SNAKE_CASE_ ) self.assertTrue(torch.allclose(output[0, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=SCREAMING_SNAKE_CASE_ ) ) def UpperCAmelCase ( self ) -> Tuple: UpperCamelCase :Optional[int] = AutoformerForPrediction.from_pretrained('''huggingface/autoformer-tourism-monthly''' ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Optional[int] = prepare_batch('''val-batch.pt''' ) with torch.no_grad(): UpperCamelCase :Tuple = model.generate( static_categorical_features=batch['''static_categorical_features'''] , past_time_features=batch['''past_time_features'''] , past_values=batch['''past_values'''] , future_time_features=batch['''future_time_features'''] , past_observed_mask=batch['''past_observed_mask'''] , ) UpperCamelCase :Optional[int] = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , SCREAMING_SNAKE_CASE_ ) UpperCamelCase :str = torch.tensor([3130.6763, 4056.5293, 7053.0786] , device=SCREAMING_SNAKE_CASE_ ) UpperCamelCase :int = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , SCREAMING_SNAKE_CASE_ , rtol=1e-1 ) )
259
1
'''simple docstring''' import argparse import json import re from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileNetVaConfig, MobileNetVaForImageClassification, MobileNetVaImageProcessor, load_tf_weights_in_mobilenet_va, ) from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase : Tuple =logging.get_logger(__name__) def UpperCAmelCase_ ( __lowerCamelCase : List[Any] ): lowercase_ :List[Any] = MobileNetVaConfig(layer_norm_eps=0.001 ) if "_quant" in model_name: raise ValueError("Quantized models are not supported." ) lowercase_ :int = re.match(r"^mobilenet_v1_([^_]*)_([^_]*)$" ,__lowerCamelCase ) if matches: lowercase_ :int = float(matches[1] ) lowercase_ :Union[str, Any] = int(matches[2] ) # The TensorFlow version of MobileNetV1 predicts 1001 classes instead of # the usual 1000. The first class (index 0) is "background". lowercase_ :Optional[int] = 10_01 lowercase_ :Union[str, Any] = "imagenet-1k-id2label.json" lowercase_ :Any = "huggingface/label-files" lowercase_ :str = json.load(open(hf_hub_download(__lowerCamelCase ,__lowerCamelCase ,repo_type="dataset" ) ,"r" ) ) lowercase_ :List[str] = {int(__lowerCamelCase ) + 1: v for k, v in idalabel.items()} lowercase_ :Optional[int] = "background" lowercase_ :List[Any] = idalabel lowercase_ :int = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase_ ( ): lowercase_ :Any = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase_ :Optional[int] = Image.open(requests.get(__lowerCamelCase ,stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def UpperCAmelCase_ ( __lowerCamelCase : Tuple ,__lowerCamelCase : str ,__lowerCamelCase : Dict ,__lowerCamelCase : Tuple=False ): lowercase_ :Optional[Any] = get_mobilenet_va_config(__lowerCamelCase ) # Load 🤗 model lowercase_ :Any = MobileNetVaForImageClassification(__lowerCamelCase ).eval() # Load weights from TensorFlow checkpoint load_tf_weights_in_mobilenet_va(__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ) # Check outputs on an image, prepared by MobileNetV1ImageProcessor lowercase_ :Union[str, Any] = MobileNetVaImageProcessor( crop_size={"width": config.image_size, "height": config.image_size} ,size={"shortest_edge": config.image_size + 32} ,) lowercase_ :Optional[int] = image_processor(images=prepare_img() ,return_tensors="pt" ) lowercase_ :Optional[int] = model(**__lowerCamelCase ) lowercase_ :Optional[int] = outputs.logits assert logits.shape == (1, 10_01) if model_name == "mobilenet_v1_1.0_224": lowercase_ :str = torch.tensor([-4.1_739, -1.1_233, 3.1_205] ) elif model_name == "mobilenet_v1_0.75_192": lowercase_ :str = torch.tensor([-3.9_440, -2.3_141, -0.3_333] ) else: lowercase_ :Optional[Any] = None if expected_logits is not None: assert torch.allclose(logits[0, :3] ,__lowerCamelCase ,atol=1e-4 ) Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__lowerCamelCase ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__lowerCamelCase ) if push_to_hub: print("Pushing to the hub..." ) lowercase_ :Optional[Any] = "google/" + model_name image_processor.push_to_hub(__lowerCamelCase ) model.push_to_hub(__lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase : Optional[Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''mobilenet_v1_1.0_224''', type=str, help='''Name of the MobileNetV1 model you\'d like to convert. Should in the form \'mobilenet_v1_<depth>_<size>\'.''', ) parser.add_argument( '''--checkpoint_path''', required=True, type=str, help='''Path to the original TensorFlow checkpoint (.ckpt file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) lowerCAmelCase : Optional[Any] =parser.parse_args() convert_movilevit_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
147
'''simple docstring''' import numpy as np def UpperCAmelCase_ ( __lowerCamelCase : str ,__lowerCamelCase : Optional[Any] ,__lowerCamelCase : int ,__lowerCamelCase : List[str] ,__lowerCamelCase : List[Any] ): lowercase_ :Dict = int(np.ceil((x_end - xa) / h ) ) lowercase_ :Optional[int] = np.zeros((n + 1,) ) lowercase_ :Tuple = ya lowercase_ :List[str] = xa for k in range(__lowerCamelCase ): lowercase_ :Any = f(__lowerCamelCase ,y[k] ) lowercase_ :Optional[Any] = f(x + 0.5 * h ,y[k] + 0.5 * h * ka ) lowercase_ :Optional[Any] = f(x + 0.5 * h ,y[k] + 0.5 * h * ka ) lowercase_ :Any = f(x + h ,y[k] + h * ka ) lowercase_ :Tuple = y[k] + (1 / 6) * h * (ka + 2 * ka + 2 * ka + ka) x += h return y if __name__ == "__main__": import doctest doctest.testmod()
147
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class __a ( unittest.TestCase ): def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Optional[Any]=7 , SCREAMING_SNAKE_CASE : Dict=3 , SCREAMING_SNAKE_CASE : Tuple=30 , SCREAMING_SNAKE_CASE : Tuple=4_00 , SCREAMING_SNAKE_CASE : Optional[int]=True , SCREAMING_SNAKE_CASE : Dict=None , SCREAMING_SNAKE_CASE : Optional[Any]=True , SCREAMING_SNAKE_CASE : int=[0.5, 0.5, 0.5] , SCREAMING_SNAKE_CASE : Dict=[0.5, 0.5, 0.5] , SCREAMING_SNAKE_CASE : List[str]=True , SCREAMING_SNAKE_CASE : List[str]=1 / 2_55 , SCREAMING_SNAKE_CASE : List[Any]=True , ): '''simple docstring''' UpperCamelCase__ : List[str] = size if size is not None else {"shortest_edge": 18, "longest_edge": 13_33} UpperCamelCase__ : List[Any] = parent UpperCamelCase__ : Optional[int] = batch_size UpperCamelCase__ : Union[str, Any] = num_channels UpperCamelCase__ : Optional[int] = min_resolution UpperCamelCase__ : int = max_resolution UpperCamelCase__ : Any = do_resize UpperCamelCase__ : Any = size UpperCamelCase__ : Any = do_normalize UpperCamelCase__ : Union[str, Any] = image_mean UpperCamelCase__ : str = image_std UpperCamelCase__ : Dict = do_rescale UpperCamelCase__ : Any = rescale_factor UpperCamelCase__ : List[Any] = do_pad def __lowercase ( self : List[str] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __lowercase ( self : Tuple , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Union[str, Any]=False ): '''simple docstring''' if not batched: UpperCamelCase__ : str = image_inputs[0] if isinstance(SCREAMING_SNAKE_CASE , Image.Image ): UpperCamelCase__ , UpperCamelCase__ : Optional[int] = image.size else: UpperCamelCase__ , UpperCamelCase__ : List[str] = image.shape[1], image.shape[2] if w < h: UpperCamelCase__ : Dict = int(self.size["shortest_edge"] * h / w ) UpperCamelCase__ : Dict = self.size["shortest_edge"] elif w > h: UpperCamelCase__ : List[str] = self.size["shortest_edge"] UpperCamelCase__ : List[str] = int(self.size["shortest_edge"] * w / h ) else: UpperCamelCase__ : Any = self.size["shortest_edge"] UpperCamelCase__ : Tuple = self.size["shortest_edge"] else: UpperCamelCase__ : List[Any] = [] for image in image_inputs: UpperCamelCase__ , UpperCamelCase__ : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCamelCase__ : Optional[int] = max(SCREAMING_SNAKE_CASE , key=lambda SCREAMING_SNAKE_CASE : item[0] )[0] UpperCamelCase__ : Optional[int] = max(SCREAMING_SNAKE_CASE , key=lambda SCREAMING_SNAKE_CASE : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class __a ( A__ , unittest.TestCase ): _lowerCAmelCase : str = YolosImageProcessor if is_vision_available() else None def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCamelCase__ : str = YolosImageProcessingTester(self ) @property def __lowercase ( self : str ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __lowercase ( self : Tuple ): '''simple docstring''' UpperCamelCase__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE , "image_mean" ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE , "image_std" ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE , "do_normalize" ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE , "do_resize" ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE , "size" ) ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCamelCase__ : str = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 13_33} ) self.assertEqual(image_processor.do_pad , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : List[str] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=SCREAMING_SNAKE_CASE ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , SCREAMING_SNAKE_CASE ) def __lowercase ( self : str ): '''simple docstring''' pass def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCamelCase__ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE , Image.Image ) # Test not batched input UpperCamelCase__ : Optional[int] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values UpperCamelCase__ , UpperCamelCase__ : Tuple = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase__ , UpperCamelCase__ : List[str] = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE , batched=SCREAMING_SNAKE_CASE ) UpperCamelCase__ : List[str] = image_processing(SCREAMING_SNAKE_CASE , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCamelCase__ : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase__ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE , numpify=SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE , np.ndarray ) # Test not batched input UpperCamelCase__ : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values UpperCamelCase__ , UpperCamelCase__ : Dict = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase__ : Tuple = image_processing(SCREAMING_SNAKE_CASE , return_tensors="pt" ).pixel_values UpperCamelCase__ , UpperCamelCase__ : Tuple = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE , batched=SCREAMING_SNAKE_CASE ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __lowercase ( self : Optional[int] ): '''simple docstring''' UpperCamelCase__ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase__ : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE , torchify=SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE , torch.Tensor ) # Test not batched input UpperCamelCase__ : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values UpperCamelCase__ , UpperCamelCase__ : List[Any] = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase__ : str = image_processing(SCREAMING_SNAKE_CASE , return_tensors="pt" ).pixel_values UpperCamelCase__ , UpperCamelCase__ : Tuple = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE , batched=SCREAMING_SNAKE_CASE ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' UpperCamelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) UpperCamelCase__ : Dict = self.image_processing_class(do_resize=SCREAMING_SNAKE_CASE , do_normalize=SCREAMING_SNAKE_CASE , do_rescale=SCREAMING_SNAKE_CASE ) # create random PyTorch tensors UpperCamelCase__ : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE , torchify=SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors UpperCamelCase__ : Dict = image_processing_a.pad(SCREAMING_SNAKE_CASE , return_tensors="pt" ) UpperCamelCase__ : int = image_processing_a(SCREAMING_SNAKE_CASE , return_tensors="pt" ) self.assertTrue( torch.allclose(encoded_images_with_method["pixel_values"] , encoded_images["pixel_values"] , atol=1e-4 ) ) @slow def __lowercase ( self : int ): '''simple docstring''' UpperCamelCase__ : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: UpperCamelCase__ : str = json.loads(f.read() ) UpperCamelCase__ : Dict = {"image_id": 3_97_69, "annotations": target} # encode them UpperCamelCase__ : Any = YolosImageProcessor.from_pretrained("hustvl/yolos-small" ) UpperCamelCase__ : int = image_processing(images=SCREAMING_SNAKE_CASE , annotations=SCREAMING_SNAKE_CASE , return_tensors="pt" ) # verify pixel values UpperCamelCase__ : List[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Dict = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) ) # verify area UpperCamelCase__ : Union[str, Any] = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , SCREAMING_SNAKE_CASE ) ) # verify boxes UpperCamelCase__ : Any = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Any = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , SCREAMING_SNAKE_CASE , atol=1e-3 ) ) # verify image_id UpperCamelCase__ : Any = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , SCREAMING_SNAKE_CASE ) ) # verify is_crowd UpperCamelCase__ : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , SCREAMING_SNAKE_CASE ) ) # verify class_labels UpperCamelCase__ : Tuple = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , SCREAMING_SNAKE_CASE ) ) # verify orig_size UpperCamelCase__ : Tuple = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , SCREAMING_SNAKE_CASE ) ) # verify size UpperCamelCase__ : int = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , SCREAMING_SNAKE_CASE ) ) @slow def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCamelCase__ : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: UpperCamelCase__ : Union[str, Any] = json.loads(f.read() ) UpperCamelCase__ : Union[str, Any] = {"file_name": "000000039769.png", "image_id": 3_97_69, "segments_info": target} UpperCamelCase__ : int = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them UpperCamelCase__ : Dict = YolosImageProcessor(format="coco_panoptic" ) UpperCamelCase__ : str = image_processing(images=SCREAMING_SNAKE_CASE , annotations=SCREAMING_SNAKE_CASE , masks_path=SCREAMING_SNAKE_CASE , return_tensors="pt" ) # verify pixel values UpperCamelCase__ : Optional[Any] = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["pixel_values"].shape , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : List[str] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) ) # verify area UpperCamelCase__ : int = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , SCREAMING_SNAKE_CASE ) ) # verify boxes UpperCamelCase__ : Union[str, Any] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Dict = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , SCREAMING_SNAKE_CASE , atol=1e-3 ) ) # verify image_id UpperCamelCase__ : Any = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , SCREAMING_SNAKE_CASE ) ) # verify is_crowd UpperCamelCase__ : Optional[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , SCREAMING_SNAKE_CASE ) ) # verify class_labels UpperCamelCase__ : Tuple = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , SCREAMING_SNAKE_CASE ) ) # verify masks UpperCamelCase__ : Any = 82_28_73 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , SCREAMING_SNAKE_CASE ) # verify orig_size UpperCamelCase__ : Tuple = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , SCREAMING_SNAKE_CASE ) ) # verify size UpperCamelCase__ : List[Any] = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , SCREAMING_SNAKE_CASE ) )
189
from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar lowerCamelCase : int =TypeVar('''T''') class __a ( Generic[T] ): _lowerCAmelCase : deque[T] # Cache store of keys _lowerCAmelCase : set[T] # References of the keys in cache _lowerCAmelCase : int = 1_0 # Maximum capacity of cache def __init__( self : List[str] , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' UpperCamelCase__ : Tuple = deque() UpperCamelCase__ : Optional[int] = set() if not n: UpperCamelCase__ : Tuple = sys.maxsize elif n < 0: raise ValueError("n should be an integer greater than 0." ) else: UpperCamelCase__ : Dict = n def __lowercase ( self : List[Any] , SCREAMING_SNAKE_CASE : T ): '''simple docstring''' if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: UpperCamelCase__ : int = self.dq_store.pop() self.key_reference.remove(SCREAMING_SNAKE_CASE ) else: self.dq_store.remove(SCREAMING_SNAKE_CASE ) self.dq_store.appendleft(SCREAMING_SNAKE_CASE ) self.key_reference.add(SCREAMING_SNAKE_CASE ) def __lowercase ( self : Dict ): '''simple docstring''' for k in self.dq_store: print(SCREAMING_SNAKE_CASE ) def __repr__( self : Union[str, Any] ): '''simple docstring''' return F'LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}' if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase : LRUCache[str | int] =LRUCache(4) lru_cache.refer('''A''') lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer('''A''') lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
189
1
lowercase = { "joule": 1.0, "kilojoule": 1_0_0_0, "megajoule": 1_0_0_0_0_0_0, "gigajoule": 1_0_0_0_0_0_0_0_0_0, "wattsecond": 1.0, "watthour": 3_6_0_0, "kilowatthour": 3_6_0_0_0_0_0, "newtonmeter": 1.0, "calorie_nutr": 4_1_8_6.8, "kilocalorie_nutr": 4_1_8_6_8_0_0.0_0, "electronvolt": 1.6_0217_6634E-19, "britishthermalunit_it": 1_0_5_5.0_5_5_8_5, "footpound": 1.35_58_18, } def lowerCamelCase_ ( UpperCamelCase__ : str, UpperCamelCase__ : str, UpperCamelCase__ : float ): '''simple docstring''' if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: UpperCamelCase__ = ( F"""Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n""" F"""Valid values are: {", ".join(UpperCamelCase__ )}""" ) raise ValueError(UpperCamelCase__ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
35
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput lowercase = logging.get_logger(__name__) # pylint: disable=invalid-name def lowerCamelCase_ ( UpperCamelCase__ : Union[List, PIL.Image.Image, torch.Tensor] ): '''simple docstring''' warnings.warn( '''The preprocess method is deprecated and will be removed in a future version. Please''' ''' use VaeImageProcessor.preprocess instead''', UpperCamelCase__, ) if isinstance(UpperCamelCase__, torch.Tensor ): return image elif isinstance(UpperCamelCase__, PIL.Image.Image ): UpperCamelCase__ = [image] if isinstance(image[0], PIL.Image.Image ): UpperCamelCase__ , UpperCamelCase__ = image[0].size UpperCamelCase__ , UpperCamelCase__ = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 UpperCamelCase__ = [np.array(i.resize((w, h), resample=PIL_INTERPOLATION['''lanczos'''] ) )[None, :] for i in image] UpperCamelCase__ = np.concatenate(UpperCamelCase__, axis=0 ) UpperCamelCase__ = np.array(UpperCamelCase__ ).astype(np.floataa ) / 255.0 UpperCamelCase__ = image.transpose(0, 3, 1, 2 ) UpperCamelCase__ = 2.0 * image - 1.0 UpperCamelCase__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(image[0], torch.Tensor ): UpperCamelCase__ = torch.cat(UpperCamelCase__, dim=0 ) return image def lowerCamelCase_ ( UpperCamelCase__ : Union[List, PIL.Image.Image, torch.Tensor] ): '''simple docstring''' if isinstance(UpperCamelCase__, torch.Tensor ): return mask elif isinstance(UpperCamelCase__, PIL.Image.Image ): UpperCamelCase__ = [mask] if isinstance(mask[0], PIL.Image.Image ): UpperCamelCase__ , UpperCamelCase__ = mask[0].size UpperCamelCase__ , UpperCamelCase__ = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 UpperCamelCase__ = [np.array(m.convert('''L''' ).resize((w, h), resample=PIL_INTERPOLATION['''nearest'''] ) )[None, :] for m in mask] UpperCamelCase__ = np.concatenate(UpperCamelCase__, axis=0 ) UpperCamelCase__ = mask.astype(np.floataa ) / 255.0 UpperCamelCase__ = 0 UpperCamelCase__ = 1 UpperCamelCase__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(mask[0], torch.Tensor ): UpperCamelCase__ = torch.cat(UpperCamelCase__, dim=0 ) return mask class __lowercase ( A ): '''simple docstring''' _A : UNetaDModel _A : RePaintScheduler def __init__( self : Tuple , _a : Any , _a : Tuple ): super().__init__() self.register_modules(unet=_a , scheduler=_a ) @torch.no_grad() def __call__( self : Optional[int] , _a : Union[torch.Tensor, PIL.Image.Image] , _a : Union[torch.Tensor, PIL.Image.Image] , _a : int = 250 , _a : float = 0.0 , _a : int = 10 , _a : int = 10 , _a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , _a : Optional[str] = "pil" , _a : bool = True , ): UpperCamelCase__ = image UpperCamelCase__ = _preprocess_image(_a ) UpperCamelCase__ = original_image.to(device=self.device , dtype=self.unet.dtype ) UpperCamelCase__ = _preprocess_mask(_a ) UpperCamelCase__ = mask_image.to(device=self.device , dtype=self.unet.dtype ) UpperCamelCase__ = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(_a , _a ) and len(_a ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(_a )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) UpperCamelCase__ = original_image.shape UpperCamelCase__ = randn_tensor(_a , generator=_a , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(_a , _a , _a , self.device ) UpperCamelCase__ = eta UpperCamelCase__ = self.scheduler.timesteps[0] + 1 UpperCamelCase__ = generator[0] if isinstance(_a , _a ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual UpperCamelCase__ = self.unet(_a , _a ).sample # compute previous image: x_t -> x_t-1 UpperCamelCase__ = self.scheduler.step(_a , _a , _a , _a , _a , _a ).prev_sample else: # compute the reverse: x_t-1 -> x_t UpperCamelCase__ = self.scheduler.undo_step(_a , _a , _a ) UpperCamelCase__ = t UpperCamelCase__ = (image / 2 + 0.5).clamp(0 , 1 ) UpperCamelCase__ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCamelCase__ = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
35
1
import argparse import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_dummies.py UpperCamelCase = '''src/diffusers''' # Matches is_xxx_available() UpperCamelCase = re.compile(R'''is\_([a-z_]*)_available\(\)''') # Matches from xxx import bla UpperCamelCase = re.compile(R'''\s+from\s+\S*\s+import\s+([^\(\s].*)\n''') UpperCamelCase = ''' {0} = None ''' UpperCamelCase = ''' class {0}(metaclass=DummyObject): _backends = {1} def __init__(self, *args, **kwargs): requires_backends(self, {1}) @classmethod def from_config(cls, *args, **kwargs): requires_backends(cls, {1}) @classmethod def from_pretrained(cls, *args, **kwargs): requires_backends(cls, {1}) ''' UpperCamelCase = ''' def {0}(*args, **kwargs): requires_backends({0}, {1}) ''' def lowercase_ ( _lowerCamelCase : List[str]): lowercase__ : Union[str, Any] = _re_backend.findall(_lowerCamelCase) if len(_lowerCamelCase) == 0: return None return "_and_".join(_lowerCamelCase) def lowercase_ ( ): with open(os.path.join(_lowerCamelCase , "__init__.py") , "r" , encoding="utf-8" , newline="\n") as f: lowercase__ : Any = f.readlines() # Get to the point we do the actual imports for type checking lowercase__ : List[Any] = 0 lowercase__ : Tuple = {} # Go through the end of the file while line_index < len(_lowerCamelCase): # If the line contains is_backend_available, we grab all objects associated with the `else` block lowercase__ : Tuple = find_backend(lines[line_index]) if backend is not None: while not lines[line_index].startswith("else:"): line_index += 1 line_index += 1 lowercase__ : str = [] # Until we unindent, add backend objects to the list while line_index < len(_lowerCamelCase) and len(lines[line_index]) > 1: lowercase__ : Any = lines[line_index] lowercase__ : Union[str, Any] = _re_single_line_import.search(_lowerCamelCase) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", ")) elif line.startswith(" " * 8): objects.append(line[8:-2]) line_index += 1 if len(_lowerCamelCase) > 0: lowercase__ : List[str] = objects else: line_index += 1 return backend_specific_objects def lowercase_ ( _lowerCamelCase : List[str] , _lowerCamelCase : int): if name.isupper(): return DUMMY_CONSTANT.format(_lowerCamelCase) elif name.islower(): return DUMMY_FUNCTION.format(_lowerCamelCase , _lowerCamelCase) else: return DUMMY_CLASS.format(_lowerCamelCase , _lowerCamelCase) def lowercase_ ( _lowerCamelCase : List[Any]=None): if backend_specific_objects is None: lowercase__ : Union[str, Any] = read_init() # For special correspondence backend to module name as used in the function requires_modulename lowercase__ : Dict = {} for backend, objects in backend_specific_objects.items(): lowercase__ : Any = "[" + ", ".join(f'''"{b}"''' for b in backend.split("_and_")) + "]" lowercase__ : Union[str, Any] = "# This file is autogenerated by the command `make fix-copies`, do not edit.\n" dummy_file += "from ..utils import DummyObject, requires_backends\n\n" dummy_file += "\n".join([create_dummy_object(_lowerCamelCase , _lowerCamelCase) for o in objects]) lowercase__ : Tuple = dummy_file return dummy_files def lowercase_ ( _lowerCamelCase : int=False): lowercase__ : Optional[Any] = create_dummy_files() # For special correspondence backend to shortcut as used in utils/dummy_xxx_objects.py lowercase__ : Tuple = {"torch": "pt"} # Locate actual dummy modules and read their content. lowercase__ : str = os.path.join(_lowerCamelCase , "utils") lowercase__ : int = { backend: os.path.join(_lowerCamelCase , f'''dummy_{short_names.get(_lowerCamelCase , _lowerCamelCase)}_objects.py''') for backend in dummy_files.keys() } lowercase__ : Tuple = {} for backend, file_path in dummy_file_paths.items(): if os.path.isfile(_lowerCamelCase): with open(_lowerCamelCase , "r" , encoding="utf-8" , newline="\n") as f: lowercase__ : Optional[Any] = f.read() else: lowercase__ : str = "" for backend in dummy_files.keys(): if dummy_files[backend] != actual_dummies[backend]: if overwrite: print( f'''Updating diffusers.utils.dummy_{short_names.get(_lowerCamelCase , _lowerCamelCase)}_objects.py as the main ''' "__init__ has new objects.") with open(dummy_file_paths[backend] , "w" , encoding="utf-8" , newline="\n") as f: f.write(dummy_files[backend]) else: raise ValueError( "The main __init__ has objects that are not present in " f'''diffusers.utils.dummy_{short_names.get(_lowerCamelCase , _lowerCamelCase)}_objects.py. Run `make fix-copies` ''' "to fix this.") if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') UpperCamelCase = parser.parse_args() check_dummies(args.fix_and_overwrite)
87
from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar UpperCamelCase = TypeVar('''T''') class snake_case_ ( Generic[T] ): __A : deque[T] # Cache store of keys __A : set[T] # References of the keys in cache __A : int = 10 # Maximum capacity of cache def __init__( self : Union[str, Any] , lowercase_ : int ) -> None: lowercase__ : int = deque() lowercase__ : str = set() if not n: lowercase__ : str = sys.maxsize elif n < 0: raise ValueError("n should be an integer greater than 0." ) else: lowercase__ : List[Any] = n def __UpperCamelCase ( self : Dict , lowercase_ : T ) -> None: if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: lowercase__ : Dict = self.dq_store.pop() self.key_reference.remove(lowercase_ ) else: self.dq_store.remove(lowercase_ ) self.dq_store.appendleft(lowercase_ ) self.key_reference.add(lowercase_ ) def __UpperCamelCase ( self : Dict ) -> None: for k in self.dq_store: print(lowercase_ ) def __repr__( self : Optional[int] ) -> str: return F'''LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}''' if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = LRUCache(4) lru_cache.refer('''A''') lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer('''A''') lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
87
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCamelCase__ = {'''configuration_swin''': ['''SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwinConfig''', '''SwinOnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''SWIN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SwinForImageClassification''', '''SwinForMaskedImageModeling''', '''SwinModel''', '''SwinPreTrainedModel''', '''SwinBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFSwinForImageClassification''', '''TFSwinForMaskedImageModeling''', '''TFSwinModel''', '''TFSwinPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
63
import math class _UpperCAmelCase : '''simple docstring''' def __UpperCAmelCase ( self : Dict , lowercase_ : list[list[float]] , lowercase_ : list[int]) -> int: """simple docstring""" _UpperCamelCase = 0.0 _UpperCamelCase = 0.0 for i in range(len(lowercase_)): da += math.pow((sample[i] - weights[0][i]) , 2) da += math.pow((sample[i] - weights[1][i]) , 2) return 0 if da > da else 1 return 0 def __UpperCAmelCase ( self : Any , lowercase_ : list[list[int | float]] , lowercase_ : list[int] , lowercase_ : int , lowercase_ : float) -> list[list[int | float]]: """simple docstring""" for i in range(len(lowercase_)): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def lowerCAmelCase__ ( ) ->None: '''simple docstring''' _UpperCamelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) _UpperCamelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training _UpperCamelCase = SelfOrganizingMap() _UpperCamelCase = 3 _UpperCamelCase = 0.5 for _ in range(a__ ): for j in range(len(a__ ) ): # training sample _UpperCamelCase = training_samples[j] # Compute the winning vector _UpperCamelCase = self_organizing_map.get_winner(a__ , a__ ) # Update the winning vector _UpperCamelCase = self_organizing_map.update(a__ , a__ , a__ , a__ ) # classify test sample _UpperCamelCase = [0, 0, 0, 1] _UpperCamelCase = self_organizing_map.get_winner(a__ , a__ ) # results print(f'Clusters that the test sample belongs to : {winner}' ) print(f'Weights that have been trained : {weights}' ) # running the main() function if __name__ == "__main__": main()
63
1
'''simple docstring''' import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase ( unittest.TestCase ): def __init__(self , lowercase , lowercase=7 , lowercase=3 , lowercase=18 , lowercase=30 , lowercase=400 , lowercase=True , lowercase=None , lowercase=True , lowercase=None , lowercase=True , lowercase=[0.5, 0.5, 0.5] , lowercase=[0.5, 0.5, 0.5] , lowercase=False , ): A_ : List[str] = size if size is not None else {"""height""": 20, """width""": 20} A_ : Optional[Any] = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} A_ : Tuple = parent A_ : Tuple = batch_size A_ : str = num_channels A_ : Optional[int] = image_size A_ : Union[str, Any] = min_resolution A_ : int = max_resolution A_ : Union[str, Any] = do_resize A_ : Optional[Any] = size A_ : List[str] = do_center_crop A_ : List[Any] = crop_size A_ : List[Any] = do_normalize A_ : Optional[int] = image_mean A_ : Tuple = image_std A_ : Optional[int] = do_reduce_labels def _a (self ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def a ( ): '''simple docstring''' A_ : int = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : Union[str, Any] = Image.open(dataset[0]["""file"""] ) A_ : List[str] = Image.open(dataset[1]["""file"""] ) return image, map def a ( ): '''simple docstring''' A_ : List[str] = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : List[str] = Image.open(ds[0]["""file"""] ) A_ : int = Image.open(ds[1]["""file"""] ) A_ : List[str] = Image.open(ds[2]["""file"""] ) A_ : Union[str, Any] = Image.open(ds[3]["""file"""] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class _lowerCAmelCase ( __UpperCAmelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE : Union[str, Any] = BeitImageProcessor if is_vision_available() else None def _a (self ): A_ : Any = BeitImageProcessingTester(self ) @property def _a (self ): return self.image_processor_tester.prepare_image_processor_dict() def _a (self ): A_ : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase , """do_resize""" ) ) self.assertTrue(hasattr(lowercase , """size""" ) ) self.assertTrue(hasattr(lowercase , """do_center_crop""" ) ) self.assertTrue(hasattr(lowercase , """center_crop""" ) ) self.assertTrue(hasattr(lowercase , """do_normalize""" ) ) self.assertTrue(hasattr(lowercase , """image_mean""" ) ) self.assertTrue(hasattr(lowercase , """image_std""" ) ) def _a (self ): A_ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 20, """width""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) self.assertEqual(image_processor.do_reduce_labels , lowercase ) A_ : Union[str, Any] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=lowercase ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) self.assertEqual(image_processor.do_reduce_labels , lowercase ) def _a (self ): pass def _a (self ): # Initialize image_processing A_ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , Image.Image ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : List[Any] = image_processing(lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _a (self ): # Initialize image_processing A_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , numpify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , np.ndarray ) # Test not batched input A_ : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : List[Any] = image_processing(lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _a (self ): # Initialize image_processing A_ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , torchify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , torch.Tensor ) # Test not batched input A_ : Optional[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : Optional[int] = image_processing(lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _a (self ): # Initialize image_processing A_ : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , torchify=lowercase ) A_ : Optional[int] = [] for image in image_inputs: self.assertIsInstance(lowercase , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input A_ : List[Any] = image_processing(image_inputs[0] , maps[0] , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched A_ : Optional[int] = image_processing(lowercase , lowercase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test not batched input (PIL images) A_, A_ : Any = prepare_semantic_single_inputs() A_ : List[Any] = image_processing(lowercase , lowercase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched input (PIL images) A_, A_ : Union[str, Any] = prepare_semantic_batch_inputs() A_ : int = image_processing(lowercase , lowercase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 2, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) def _a (self ): # Initialize image_processing A_ : Dict = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 A_, A_ : Union[str, Any] = prepare_semantic_single_inputs() A_ : Union[str, Any] = image_processing(lowercase , lowercase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 150 ) A_ : Union[str, Any] = True A_ : Union[str, Any] = image_processing(lowercase , lowercase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 )
206
'''simple docstring''' import argparse import struct import unittest class _lowerCAmelCase : def __init__(self , lowercase ): A_ : List[str] = data # Initialize hash values A_ : Tuple = [ 0X6A09_E667, 0XBB67_AE85, 0X3C6E_F372, 0XA54F_F53A, 0X510E_527F, 0X9B05_688C, 0X1F83_D9AB, 0X5BE0_CD19, ] # Initialize round constants A_ : List[Any] = [ 0X428A_2F98, 0X7137_4491, 0XB5C0_FBCF, 0XE9B5_DBA5, 0X3956_C25B, 0X59F1_11F1, 0X923F_82A4, 0XAB1C_5ED5, 0XD807_AA98, 0X1283_5B01, 0X2431_85BE, 0X550C_7DC3, 0X72BE_5D74, 0X80DE_B1FE, 0X9BDC_06A7, 0XC19B_F174, 0XE49B_69C1, 0XEFBE_4786, 0X0FC1_9DC6, 0X240C_A1CC, 0X2DE9_2C6F, 0X4A74_84AA, 0X5CB0_A9DC, 0X76F9_88DA, 0X983E_5152, 0XA831_C66D, 0XB003_27C8, 0XBF59_7FC7, 0XC6E0_0BF3, 0XD5A7_9147, 0X06CA_6351, 0X1429_2967, 0X27B7_0A85, 0X2E1B_2138, 0X4D2C_6DFC, 0X5338_0D13, 0X650A_7354, 0X766A_0ABB, 0X81C2_C92E, 0X9272_2C85, 0XA2BF_E8A1, 0XA81A_664B, 0XC24B_8B70, 0XC76C_51A3, 0XD192_E819, 0XD699_0624, 0XF40E_3585, 0X106A_A070, 0X19A4_C116, 0X1E37_6C08, 0X2748_774C, 0X34B0_BCB5, 0X391C_0CB3, 0X4ED8_AA4A, 0X5B9C_CA4F, 0X682E_6FF3, 0X748F_82EE, 0X78A5_636F, 0X84C8_7814, 0X8CC7_0208, 0X90BE_FFFA, 0XA450_6CEB, 0XBEF9_A3F7, 0XC671_78F2, ] A_ : Tuple = self.preprocessing(self.data ) self.final_hash() @staticmethod def _a (lowercase ): A_ : Any = b"""\x80""" + (b"""\x00""" * (63 - (len(lowercase ) + 8) % 64)) A_ : Optional[Any] = struct.pack(""">Q""" , (len(lowercase ) * 8) ) return data + padding + big_endian_integer def _a (self ): # Convert into blocks of 64 bytes A_ : str = [ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers A_ : Tuple = list(struct.unpack(""">16L""" , lowercase ) ) # add 48 0-ed integers words += [0] * 48 A_, A_, A_, A_, A_, A_, A_, A_ : Dict = self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array A_ : Optional[Any] = ( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) A_ : Tuple = ( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) A_ : Any = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X1_0000_0000 # Compression A_ : Union[str, Any] = self.ror(lowercase , 6 ) ^ self.ror(lowercase , 11 ) ^ self.ror(lowercase , 25 ) A_ : List[Any] = (e & f) ^ ((~e & 0XFFFF_FFFF) & g) A_ : Dict = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X1_0000_0000 A_ : List[str] = self.ror(lowercase , 2 ) ^ self.ror(lowercase , 13 ) ^ self.ror(lowercase , 22 ) A_ : str = (a & b) ^ (a & c) ^ (b & c) A_ : int = (sa + maj) % 0X1_0000_0000 A_, A_, A_, A_, A_, A_, A_, A_ : List[Any] = ( g, f, e, ((d + tempa) % 0X1_0000_0000), c, b, a, ((tempa + tempa) % 0X1_0000_0000), ) A_ : int = [a, b, c, d, e, f, g, h] # Modify final values A_ : Dict = [ ((element + mutated_hash_values[index]) % 0X1_0000_0000) for index, element in enumerate(self.hashes ) ] A_ : Any = """""".join([hex(lowercase )[2:].zfill(8 ) for value in self.hashes] ) def _a (self , lowercase , lowercase ): return 0XFFFF_FFFF & (value << (32 - rotations)) | (value >> rotations) class _lowerCAmelCase ( unittest.TestCase ): def _a (self ): import hashlib A_ : Optional[Any] = bytes("""Test String""" , """utf-8""" ) self.assertEqual(SHAaaa(lowercase ).hash , hashlib.shaaaa(lowercase ).hexdigest() ) def a ( ): '''simple docstring''' import doctest doctest.testmod() A_ : Any = argparse.ArgumentParser() parser.add_argument( """-s""" , """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument( """-f""" , """--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) A_ : List[str] = parser.parse_args() A_ : Tuple = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: A_ : Union[str, Any] = f.read() else: A_ : Optional[Any] = bytes(lowerCamelCase__ , """utf-8""" ) print(SHAaaa(lowerCamelCase__ ).hash ) if __name__ == "__main__": main()
206
1
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class _snake_case ( lowercase__): def __init__( self : List[Any], __lowercase : Tuple, __lowercase : List[Any]=13, __lowercase : Dict=7, __lowercase : int=True, __lowercase : List[str]=True, __lowercase : str=False, __lowercase : str=True, __lowercase : Dict=99, __lowercase : Any=32, __lowercase : Optional[Any]=5, __lowercase : Union[str, Any]=4, __lowercase : int=37, __lowercase : int="gelu", __lowercase : Optional[int]=0.1, __lowercase : Any=0.1, __lowercase : List[Any]=512, __lowercase : Optional[Any]=16, __lowercase : Tuple=2, __lowercase : Dict=0.02, __lowercase : Tuple=3, __lowercase : Union[str, Any]=4, __lowercase : List[str]=None, ): lowercase__ = parent lowercase__ = batch_size lowercase__ = seq_length lowercase__ = is_training lowercase__ = use_input_mask lowercase__ = use_token_type_ids lowercase__ = use_labels lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = type_vocab_size lowercase__ = type_sequence_label_size lowercase__ = initializer_range lowercase__ = num_labels lowercase__ = num_choices lowercase__ = scope def A__ ( self : List[str] ): lowercase__ = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) lowercase__ = None if self.use_input_mask: lowercase__ = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__ = None lowercase__ = None lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size], self.type_sequence_label_size ) lowercase__ = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) lowercase__ = ids_tensor([self.batch_size], self.num_choices ) lowercase__ = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A__ ( self : int ): return DistilBertConfig( vocab_size=self.vocab_size, dim=self.hidden_size, n_layers=self.num_hidden_layers, n_heads=self.num_attention_heads, hidden_dim=self.intermediate_size, hidden_act=self.hidden_act, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, initializer_range=self.initializer_range, ) def A__ ( self : Union[str, Any], __lowercase : Optional[Any], __lowercase : Dict, __lowercase : Union[str, Any], __lowercase : Optional[Any], __lowercase : Tuple, __lowercase : List[Any] ): lowercase__ = DistilBertModel(config=__lowercase ) model.to(__lowercase ) model.eval() lowercase__ = model(__lowercase, __lowercase ) lowercase__ = model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def A__ ( self : Tuple, __lowercase : List[Any], __lowercase : Dict, __lowercase : Any, __lowercase : Union[str, Any], __lowercase : List[Any], __lowercase : Tuple ): lowercase__ = DistilBertForMaskedLM(config=__lowercase ) model.to(__lowercase ) model.eval() lowercase__ = model(__lowercase, attention_mask=__lowercase, labels=__lowercase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def A__ ( self : Optional[int], __lowercase : int, __lowercase : Optional[Any], __lowercase : Optional[int], __lowercase : List[str], __lowercase : int, __lowercase : Any ): lowercase__ = DistilBertForQuestionAnswering(config=__lowercase ) model.to(__lowercase ) model.eval() lowercase__ = model( __lowercase, attention_mask=__lowercase, start_positions=__lowercase, end_positions=__lowercase ) self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length) ) def A__ ( self : Any, __lowercase : int, __lowercase : str, __lowercase : Optional[Any], __lowercase : List[Any], __lowercase : str, __lowercase : Optional[Any] ): lowercase__ = self.num_labels lowercase__ = DistilBertForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() lowercase__ = model(__lowercase, attention_mask=__lowercase, labels=__lowercase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def A__ ( self : Union[str, Any], __lowercase : Union[str, Any], __lowercase : Tuple, __lowercase : List[str], __lowercase : str, __lowercase : Optional[Any], __lowercase : str ): lowercase__ = self.num_labels lowercase__ = DistilBertForTokenClassification(config=__lowercase ) model.to(__lowercase ) model.eval() lowercase__ = model(__lowercase, attention_mask=__lowercase, labels=__lowercase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels) ) def A__ ( self : Optional[int], __lowercase : List[Any], __lowercase : List[str], __lowercase : List[Any], __lowercase : Optional[Any], __lowercase : Optional[int], __lowercase : Tuple ): lowercase__ = self.num_choices lowercase__ = DistilBertForMultipleChoice(config=__lowercase ) model.to(__lowercase ) model.eval() lowercase__ = input_ids.unsqueeze(1 ).expand(-1, self.num_choices, -1 ).contiguous() lowercase__ = input_mask.unsqueeze(1 ).expand(-1, self.num_choices, -1 ).contiguous() lowercase__ = model( __lowercase, attention_mask=__lowercase, labels=__lowercase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices) ) def A__ ( self : Any ): lowercase__ = self.prepare_config_and_inputs() ((lowercase__) , (lowercase__) , (lowercase__) , (lowercase__) , (lowercase__) , (lowercase__)) = config_and_inputs lowercase__ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _snake_case ( lowercase__ , lowercase__ , unittest.TestCase): UpperCamelCase__ : Dict =( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) UpperCamelCase__ : Any =( { """feature-extraction""": DistilBertModel, """fill-mask""": DistilBertForMaskedLM, """question-answering""": DistilBertForQuestionAnswering, """text-classification""": DistilBertForSequenceClassification, """token-classification""": DistilBertForTokenClassification, """zero-shot""": DistilBertForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase__ : Optional[Any] =True UpperCamelCase__ : str =True UpperCamelCase__ : Any =True UpperCamelCase__ : List[str] =True def A__ ( self : Dict ): lowercase__ = DistilBertModelTester(self ) lowercase__ = ConfigTester(self, config_class=__lowercase, dim=37 ) def A__ ( self : Union[str, Any] ): self.config_tester.run_common_tests() def A__ ( self : str ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*__lowercase ) def A__ ( self : Optional[Any] ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*__lowercase ) def A__ ( self : List[str] ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*__lowercase ) def A__ ( self : Union[str, Any] ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*__lowercase ) def A__ ( self : Optional[int] ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*__lowercase ) def A__ ( self : List[str] ): lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*__lowercase ) @slow def A__ ( self : Dict ): for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = DistilBertModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) @slow @require_torch_gpu def A__ ( self : List[str] ): lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return lowercase__ = True lowercase__ = model_class(config=__lowercase ) lowercase__ = self._prepare_for_class(__lowercase, __lowercase ) lowercase__ = torch.jit.trace( __lowercase, (inputs_dict["input_ids"].to("cpu" ), inputs_dict["attention_mask"].to("cpu" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(__lowercase, os.path.join(__lowercase, "traced_model.pt" ) ) lowercase__ = torch.jit.load(os.path.join(__lowercase, "traced_model.pt" ), map_location=__lowercase ) loaded(inputs_dict["input_ids"].to(__lowercase ), inputs_dict["attention_mask"].to(__lowercase ) ) @require_torch class _snake_case ( unittest.TestCase): @slow def A__ ( self : str ): lowercase__ = DistilBertModel.from_pretrained("distilbert-base-uncased" ) lowercase__ = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) lowercase__ = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): lowercase__ = model(__lowercase, attention_mask=__lowercase )[0] lowercase__ = torch.Size((1, 11, 768) ) self.assertEqual(output.shape, __lowercase ) lowercase__ = torch.tensor( [[[-0.1639, 0.3299, 0.1648], [-0.1746, 0.3289, 0.1710], [-0.1884, 0.3357, 0.1810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4], __lowercase, atol=1e-4 ) )
224
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """transfo-xl-wt103""": """https://huggingface.co/transfo-xl-wt103/resolve/main/config.json""", } class _snake_case ( lowercase__): UpperCamelCase__ : Optional[Any] ="""transfo-xl""" UpperCamelCase__ : Dict =["""mems"""] UpperCamelCase__ : Optional[int] ={ """n_token""": """vocab_size""", """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Optional[Any], __lowercase : Optional[Any]=26_7735, __lowercase : int=[2_0000, 4_0000, 20_0000], __lowercase : Union[str, Any]=1024, __lowercase : Tuple=1024, __lowercase : Tuple=16, __lowercase : Optional[Any]=64, __lowercase : str=4096, __lowercase : Optional[int]=4, __lowercase : Union[str, Any]=False, __lowercase : Union[str, Any]=18, __lowercase : List[str]=1600, __lowercase : List[Any]=1000, __lowercase : Union[str, Any]=True, __lowercase : Tuple=True, __lowercase : Optional[Any]=0, __lowercase : List[str]=-1, __lowercase : int=True, __lowercase : Dict=0.1, __lowercase : Union[str, Any]=0.0, __lowercase : str=True, __lowercase : Optional[Any]="normal", __lowercase : str=0.01, __lowercase : Tuple=0.01, __lowercase : List[Any]=0.02, __lowercase : Any=1e-5, __lowercase : Union[str, Any]=0, **__lowercase : Union[str, Any], ): lowercase__ = vocab_size lowercase__ = [] self.cutoffs.extend(__lowercase ) if proj_share_all_but_first: lowercase__ = [False] + [True] * len(self.cutoffs ) else: lowercase__ = [False] + [False] * len(self.cutoffs ) lowercase__ = d_model lowercase__ = d_embed lowercase__ = d_head lowercase__ = d_inner lowercase__ = div_val lowercase__ = pre_lnorm lowercase__ = n_layer lowercase__ = n_head lowercase__ = mem_len lowercase__ = same_length lowercase__ = attn_type lowercase__ = clamp_len lowercase__ = sample_softmax lowercase__ = adaptive lowercase__ = dropout lowercase__ = dropatt lowercase__ = untie_r lowercase__ = init lowercase__ = init_range lowercase__ = proj_init_std lowercase__ = init_std lowercase__ = layer_norm_epsilon super().__init__(eos_token_id=__lowercase, **__lowercase ) @property def A__ ( self : Optional[Any] ): # Message copied from Transformer-XL documentation logger.info(F'''The model {self.model_type} is one of the few models that has no sequence length limit.''' ) return -1 @max_position_embeddings.setter def A__ ( self : List[str], __lowercase : Union[str, Any] ): # Message copied from Transformer-XL documentation raise NotImplementedError( F'''The model {self.model_type} is one of the few models that has no sequence length limit.''' )
224
1
import math def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): if ( not isinstance(__lowerCamelCase , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError("power_factor must be a valid float value between -1 and 1." ) return apparent_power * power_factor def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): if ( not isinstance(__lowerCamelCase , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError("power_factor must be a valid float value between -1 and 1." ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
123
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _snake_case : int = logging.get_logger(__name__) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Any = DPTConfig(embedding_type="hybrid" ) if "large" in checkpoint_url: __snake_case : Optional[int] = 1_0_2_4 __snake_case : List[Any] = 4_0_9_6 __snake_case : List[Any] = 2_4 __snake_case : Optional[Any] = 1_6 __snake_case : str = [5, 1_1, 1_7, 2_3] __snake_case : List[str] = [2_5_6, 5_1_2, 1_0_2_4, 1_0_2_4] __snake_case : Union[str, Any] = (1, 3_8_4, 3_8_4) if "nyu" or "midas" in checkpoint_url: __snake_case : Tuple = 7_6_8 __snake_case : Any = [1, 1, 1, 0.5] __snake_case : Any = [2_5_6, 5_1_2, 7_6_8, 7_6_8] __snake_case : Any = 1_5_0 __snake_case : Optional[Any] = 1_6 __snake_case : List[str] = (1, 3_8_4, 3_8_4) __snake_case : Tuple = False __snake_case : Optional[Any] = "project" if "ade" in checkpoint_url: __snake_case : Optional[int] = True __snake_case : List[str] = 7_6_8 __snake_case : int = [1, 1, 1, 0.5] __snake_case : Any = 1_5_0 __snake_case : Tuple = 1_6 __snake_case : List[str] = "huggingface/label-files" __snake_case : Union[str, Any] = "ade20k-id2label.json" __snake_case : List[str] = json.load(open(cached_download(hf_hub_url(__lowerCamelCase , __lowerCamelCase , repo_type="dataset" ) ) , "r" ) ) __snake_case : int = {int(__lowerCamelCase ): v for k, v in idalabel.items()} __snake_case : Optional[Any] = idalabel __snake_case : str = {v: k for k, v in idalabel.items()} __snake_case : Tuple = [1, 1_5_0, 4_8_0, 4_8_0] return config, expected_shape def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = ["pretrained.model.head.weight", "pretrained.model.head.bias"] for k in ignore_keys: state_dict.pop(__lowerCamelCase , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): __snake_case : Tuple = name.replace("pretrained.model" , "dpt.encoder" ) if "pretrained.model" in name: __snake_case : Tuple = name.replace("pretrained.model" , "dpt.embeddings" ) if "patch_embed" in name: __snake_case : Optional[Any] = name.replace("patch_embed" , "" ) if "pos_embed" in name: __snake_case : Optional[int] = name.replace("pos_embed" , "position_embeddings" ) if "attn.proj" in name: __snake_case : List[str] = name.replace("attn.proj" , "attention.output.dense" ) if "proj" in name and "project" not in name: __snake_case : Union[str, Any] = name.replace("proj" , "projection" ) if "blocks" in name: __snake_case : int = name.replace("blocks" , "layer" ) if "mlp.fc1" in name: __snake_case : Tuple = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: __snake_case : Any = name.replace("mlp.fc2" , "output.dense" ) if "norm1" in name and "backbone" not in name: __snake_case : Optional[Any] = name.replace("norm1" , "layernorm_before" ) if "norm2" in name and "backbone" not in name: __snake_case : Any = name.replace("norm2" , "layernorm_after" ) if "scratch.output_conv" in name: __snake_case : Dict = name.replace("scratch.output_conv" , "head" ) if "scratch" in name: __snake_case : Union[str, Any] = name.replace("scratch" , "neck" ) if "layer1_rn" in name: __snake_case : List[Any] = name.replace("layer1_rn" , "convs.0" ) if "layer2_rn" in name: __snake_case : str = name.replace("layer2_rn" , "convs.1" ) if "layer3_rn" in name: __snake_case : List[str] = name.replace("layer3_rn" , "convs.2" ) if "layer4_rn" in name: __snake_case : Optional[int] = name.replace("layer4_rn" , "convs.3" ) if "refinenet" in name: __snake_case : Optional[int] = int(name[len("neck.refinenet" ) : len("neck.refinenet" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 __snake_case : int = name.replace(F'refinenet{layer_idx}' , F'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: __snake_case : Any = name.replace("out_conv" , "projection" ) if "resConfUnit1" in name: __snake_case : List[Any] = name.replace("resConfUnit1" , "residual_layer1" ) if "resConfUnit2" in name: __snake_case : Tuple = name.replace("resConfUnit2" , "residual_layer2" ) if "conv1" in name: __snake_case : List[str] = name.replace("conv1" , "convolution1" ) if "conv2" in name: __snake_case : str = name.replace("conv2" , "convolution2" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: __snake_case : Union[str, Any] = name.replace("pretrained.act_postprocess1.0.project.0" , "neck.reassemble_stage.readout_projects.0.0" ) if "pretrained.act_postprocess2.0.project.0" in name: __snake_case : Optional[int] = name.replace("pretrained.act_postprocess2.0.project.0" , "neck.reassemble_stage.readout_projects.1.0" ) if "pretrained.act_postprocess3.0.project.0" in name: __snake_case : List[str] = name.replace("pretrained.act_postprocess3.0.project.0" , "neck.reassemble_stage.readout_projects.2.0" ) if "pretrained.act_postprocess4.0.project.0" in name: __snake_case : Dict = name.replace("pretrained.act_postprocess4.0.project.0" , "neck.reassemble_stage.readout_projects.3.0" ) # resize blocks if "pretrained.act_postprocess1.3" in name: __snake_case : Tuple = name.replace("pretrained.act_postprocess1.3" , "neck.reassemble_stage.layers.0.projection" ) if "pretrained.act_postprocess1.4" in name: __snake_case : int = name.replace("pretrained.act_postprocess1.4" , "neck.reassemble_stage.layers.0.resize" ) if "pretrained.act_postprocess2.3" in name: __snake_case : Union[str, Any] = name.replace("pretrained.act_postprocess2.3" , "neck.reassemble_stage.layers.1.projection" ) if "pretrained.act_postprocess2.4" in name: __snake_case : Optional[Any] = name.replace("pretrained.act_postprocess2.4" , "neck.reassemble_stage.layers.1.resize" ) if "pretrained.act_postprocess3.3" in name: __snake_case : Optional[int] = name.replace("pretrained.act_postprocess3.3" , "neck.reassemble_stage.layers.2.projection" ) if "pretrained.act_postprocess4.3" in name: __snake_case : Dict = name.replace("pretrained.act_postprocess4.3" , "neck.reassemble_stage.layers.3.projection" ) if "pretrained.act_postprocess4.4" in name: __snake_case : Union[str, Any] = name.replace("pretrained.act_postprocess4.4" , "neck.reassemble_stage.layers.3.resize" ) if "pretrained" in name: __snake_case : Union[str, Any] = name.replace("pretrained" , "dpt" ) if "bn" in name: __snake_case : Tuple = name.replace("bn" , "batch_norm" ) if "head" in name: __snake_case : Dict = name.replace("head" , "head.head" ) if "encoder.norm" in name: __snake_case : Optional[int] = name.replace("encoder.norm" , "layernorm" ) if "auxlayer" in name: __snake_case : Tuple = name.replace("auxlayer" , "auxiliary_head.head" ) if "backbone" in name: __snake_case : str = name.replace("backbone" , "backbone.bit.encoder" ) if ".." in name: __snake_case : Tuple = name.replace(".." , "." ) if "stem.conv" in name: __snake_case : int = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: __snake_case : Any = name.replace("blocks" , "layers" ) if "convolution" in name and "backbone" in name: __snake_case : Optional[int] = name.replace("convolution" , "conv" ) if "layer" in name and "backbone" in name: __snake_case : List[Any] = name.replace("layer" , "layers" ) if "backbone.bit.encoder.bit" in name: __snake_case : Optional[int] = name.replace("backbone.bit.encoder.bit" , "backbone.bit" ) if "embedder.conv" in name: __snake_case : int = name.replace("embedder.conv" , "embedder.convolution" ) if "backbone.bit.encoder.stem.norm" in name: __snake_case : Optional[Any] = name.replace("backbone.bit.encoder.stem.norm" , "backbone.bit.embedder.norm" ) return name def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __snake_case : int = state_dict.pop(F'dpt.encoder.layer.{i}.attn.qkv.weight' ) __snake_case : Any = state_dict.pop(F'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict __snake_case : str = in_proj_weight[: config.hidden_size, :] __snake_case : List[Any] = in_proj_bias[: config.hidden_size] __snake_case : str = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __snake_case : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __snake_case : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] __snake_case : int = in_proj_bias[-config.hidden_size :] def lowerCAmelCase_ ( ): __snake_case : List[str] = "http://images.cocodataset.org/val2017/000000039769.jpg" __snake_case : int = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case , __snake_case : Optional[int] = get_dpt_config(__lowerCamelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") __snake_case : Optional[int] = torch.load(__lowerCamelCase , map_location="cpu" ) # remove certain keys remove_ignore_keys_(__lowerCamelCase ) # rename keys for key in state_dict.copy().keys(): __snake_case : Optional[int] = state_dict.pop(__lowerCamelCase ) __snake_case : Optional[Any] = val # read in qkv matrices read_in_q_k_v(__lowerCamelCase , __lowerCamelCase ) # load HuggingFace model __snake_case : Dict = DPTForSemanticSegmentation(__lowerCamelCase ) if "ade" in checkpoint_url else DPTForDepthEstimation(__lowerCamelCase ) model.load_state_dict(__lowerCamelCase ) model.eval() # Check outputs on an image __snake_case : str = 4_8_0 if "ade" in checkpoint_url else 3_8_4 __snake_case : Any = DPTImageProcessor(size=__lowerCamelCase ) __snake_case : int = prepare_img() __snake_case : Union[str, Any] = image_processor(__lowerCamelCase , return_tensors="pt" ) # forward pass __snake_case : Dict = model(**__lowerCamelCase ).logits if "ade" in checkpoint_url else model(**__lowerCamelCase ).predicted_depth if show_prediction: __snake_case : int = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode="bicubic" , align_corners=__lowerCamelCase , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 2_5_5 ).show() if pytorch_dump_folder_path is not None: Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(F'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(__lowerCamelCase ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(__lowerCamelCase ) if push_to_hub: model.push_to_hub("ybelkada/dpt-hybrid-midas" ) image_processor.push_to_hub("ybelkada/dpt-hybrid-midas" ) if __name__ == "__main__": _snake_case : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt", type=str, help="URL of the original DPT checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=False, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", action="store_true", ) parser.add_argument( "--model_name", default="dpt-large", type=str, help="Name of the model, in case you're pushing to the hub.", ) parser.add_argument( "--show_prediction", action="store_true", ) _snake_case : str = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
123
1
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(__UpperCAmelCase ) ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): # Base Case if index == len(__UpperCAmelCase ): return True # Recursive Step for i in range(__UpperCAmelCase ): if valid_coloring(graph[index] , __UpperCAmelCase , __UpperCAmelCase ): # Color current vertex _lowercase : Dict = i # Validate coloring if util_color(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , index + 1 ): return True # Backtrack _lowercase : Dict = -1 return False def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase , __UpperCAmelCase ): _lowercase : Union[str, Any] = [-1] * len(__UpperCAmelCase ) if util_color(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , 0 ): return colored_vertices return []
336
"""simple docstring""" import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCamelCase ( snake_case , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = LEDTokenizer SCREAMING_SNAKE_CASE_ : List[str] = LEDTokenizerFast SCREAMING_SNAKE_CASE_ : List[str] = True def lowerCamelCase__ ( self ): super().setUp() _lowercase : Union[str, Any] = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] _lowercase : List[Any] = dict(zip(UpperCAmelCase_ ,range(len(UpperCAmelCase_ ) ) ) ) _lowercase : Optional[int] = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] _lowercase : Dict = {"""unk_token""": """<unk>"""} _lowercase : Optional[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) _lowercase : List[str] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + """\n""" ) with open(self.merges_file ,"""w""" ,encoding="""utf-8""" ) as fp: fp.write("""\n""".join(UpperCAmelCase_ ) ) def lowerCamelCase__ ( self ,**UpperCAmelCase_ ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**UpperCAmelCase_ ) def lowerCamelCase__ ( self ,**UpperCAmelCase_ ): kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**UpperCAmelCase_ ) def lowerCamelCase__ ( self ,UpperCAmelCase_ ): return "lower newer", "lower newer" @cached_property def lowerCamelCase__ ( self ): return LEDTokenizer.from_pretrained("""allenai/led-base-16384""" ) @cached_property def lowerCamelCase__ ( self ): return LEDTokenizerFast.from_pretrained("""allenai/led-base-16384""" ) @require_torch def lowerCamelCase__ ( self ): _lowercase : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] _lowercase : Any = [0, 2_50, 2_51, 1_78_18, 13, 3_91_86, 19_38, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _lowercase : Tuple = tokenizer(UpperCAmelCase_ ,max_length=len(UpperCAmelCase_ ) ,padding=UpperCAmelCase_ ,return_tensors="""pt""" ) self.assertIsInstance(UpperCAmelCase_ ,UpperCAmelCase_ ) self.assertEqual((2, 9) ,batch.input_ids.shape ) self.assertEqual((2, 9) ,batch.attention_mask.shape ) _lowercase : Optional[Any] = batch.input_ids.tolist()[0] self.assertListEqual(UpperCAmelCase_ ,UpperCAmelCase_ ) @require_torch def lowerCamelCase__ ( self ): _lowercase : Tuple = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _lowercase : Dict = tokenizer(UpperCAmelCase_ ,padding=UpperCAmelCase_ ,return_tensors="""pt""" ) self.assertIn("""input_ids""" ,UpperCAmelCase_ ) self.assertIn("""attention_mask""" ,UpperCAmelCase_ ) self.assertNotIn("""labels""" ,UpperCAmelCase_ ) self.assertNotIn("""decoder_attention_mask""" ,UpperCAmelCase_ ) @require_torch def lowerCamelCase__ ( self ): _lowercase : Dict = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _lowercase : Tuple = tokenizer(text_target=UpperCAmelCase_ ,max_length=32 ,padding="""max_length""" ,return_tensors="""pt""" ) self.assertEqual(32 ,targets["""input_ids"""].shape[1] ) @require_torch def lowerCamelCase__ ( self ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _lowercase : List[Any] = tokenizer( ["""I am a small frog""" * 10_24, """I am a small frog"""] ,padding=UpperCAmelCase_ ,truncation=UpperCAmelCase_ ,return_tensors="""pt""" ) self.assertIsInstance(UpperCAmelCase_ ,UpperCAmelCase_ ) self.assertEqual(batch.input_ids.shape ,(2, 51_22) ) @require_torch def lowerCamelCase__ ( self ): _lowercase : List[Any] = ["""A long paragraph for summarization."""] _lowercase : Dict = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _lowercase : Dict = tokenizer(UpperCAmelCase_ ,return_tensors="""pt""" ) _lowercase : List[str] = tokenizer(text_target=UpperCAmelCase_ ,return_tensors="""pt""" ) _lowercase : Union[str, Any] = inputs["""input_ids"""] _lowercase : List[str] = targets["""input_ids"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def lowerCamelCase__ ( self ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _lowercase : str = ["""Summary of the text.""", """Another summary."""] _lowercase : Optional[int] = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] _lowercase : Any = tokenizer(UpperCAmelCase_ ,padding=UpperCAmelCase_ ) _lowercase : str = [[0] * len(UpperCAmelCase_ ) for x in encoded_output["""input_ids"""]] _lowercase : Optional[int] = tokenizer.pad(UpperCAmelCase_ ) self.assertSequenceEqual(outputs["""global_attention_mask"""] ,UpperCAmelCase_ ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowercase : int = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase_ ,**UpperCAmelCase_ ) _lowercase : Optional[int] = self.tokenizer_class.from_pretrained(UpperCAmelCase_ ,**UpperCAmelCase_ ) _lowercase : Dict = """A, <mask> AllenNLP sentence.""" _lowercase : List[Any] = tokenizer_r.encode_plus(UpperCAmelCase_ ,add_special_tokens=UpperCAmelCase_ ,return_token_type_ids=UpperCAmelCase_ ) _lowercase : Any = tokenizer_p.encode_plus(UpperCAmelCase_ ,add_special_tokens=UpperCAmelCase_ ,return_token_type_ids=UpperCAmelCase_ ) self.assertEqual(sum(tokens_r["""token_type_ids"""] ) ,sum(tokens_p["""token_type_ids"""] ) ) self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) ,sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) ,) _lowercase : str = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) _lowercase : str = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) self.assertSequenceEqual(tokens_p["""input_ids"""] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( UpperCAmelCase_ ,["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( UpperCAmelCase_ ,["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] )
336
1
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class lowercase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : Tuple=13 , _UpperCAmelCase : Tuple=7 , _UpperCAmelCase : Tuple=True , _UpperCAmelCase : Dict=True , _UpperCAmelCase : str=True , _UpperCAmelCase : List[str]=True , _UpperCAmelCase : Union[str, Any]=99 , _UpperCAmelCase : Any=32 , _UpperCAmelCase : Union[str, Any]=5 , _UpperCAmelCase : int=4 , _UpperCAmelCase : str=37 , _UpperCAmelCase : Any="gelu" , _UpperCAmelCase : int=0.1 , _UpperCAmelCase : Union[str, Any]=0.1 , _UpperCAmelCase : Tuple=512 , _UpperCAmelCase : Dict=16 , _UpperCAmelCase : Any=2 , _UpperCAmelCase : Optional[int]=0.02 , _UpperCAmelCase : List[str]=4 , ): _A = parent _A = batch_size _A = seq_length _A = is_training _A = use_attention_mask _A = use_token_type_ids _A = use_labels _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = type_sequence_label_size _A = initializer_range _A = num_choices def lowerCAmelCase_ ( self : Any ): _A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A = None if self.use_attention_mask: _A = random_attention_mask([self.batch_size, self.seq_length] ) _A = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=_UpperCAmelCase , ) return config, input_ids, attention_mask def lowerCAmelCase_ ( self : Optional[int] ): _A = self.prepare_config_and_inputs() _A , _A , _A = config_and_inputs _A = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class lowercase_ ( __lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : Dict = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase_ ( self : Optional[int] ): _A = FlaxDistilBertModelTester(self ) @slow def lowerCAmelCase_ ( self : List[Any] ): for model_class_name in self.all_model_classes: _A = model_class_name.from_pretrained('distilbert-base-uncased' ) _A = model(np.ones((1, 1) ) ) self.assertIsNotNone(_UpperCAmelCase ) @require_flax class lowercase_ ( unittest.TestCase ): '''simple docstring''' @slow def lowerCAmelCase_ ( self : List[Any] ): _A = FlaxDistilBertModel.from_pretrained('distilbert-base-uncased' ) _A = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) _A = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _A = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase )[0] _A = (1, 11, 768) self.assertEqual(output.shape , _UpperCAmelCase ) _A = np.array([[[-0.1639, 0.3299, 0.1648], [-0.1746, 0.3289, 0.1710], [-0.1884, 0.3357, 0.1810]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , _UpperCAmelCase , atol=1E-4 ) )
315
"""simple docstring""" import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor a = logging.get_logger(__name__) class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' def __init__( self : Any , *_UpperCAmelCase : List[str] , **_UpperCAmelCase : Union[str, Any] ): warnings.warn( 'The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use ImageGPTImageProcessor instead.' , _UpperCAmelCase , ) super().__init__(*_UpperCAmelCase , **_UpperCAmelCase )
315
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase : List[Any] = { '''configuration_clipseg''': [ '''CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CLIPSegConfig''', '''CLIPSegTextConfig''', '''CLIPSegVisionConfig''', ], '''processing_clipseg''': ['''CLIPSegProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : Optional[Any] = [ '''CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CLIPSegModel''', '''CLIPSegPreTrainedModel''', '''CLIPSegTextModel''', '''CLIPSegVisionModel''', '''CLIPSegForImageSegmentation''', ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys lowerCamelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
352
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def lowerCAmelCase ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase : Dict = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) __lowercase : List[str] = tf.convert_to_tensor( [[5, 121, 11, 660, 16, 730, 25543, 110, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" __lowercase : Optional[Any] = model(__a )["""last_hidden_state"""] __lowercase : Any = tf.TensorShape((1, 10, 768) ) self.assertEqual(output.shape , __a ) # compare the actual values for a slice. __lowercase : Dict = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
306
0
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): UpperCAmelCase = True from torch.cuda.amp import autocast UpperCAmelCase = logging.getLogger(__name__) @dataclass class A_ : '''simple docstring''' _UpperCamelCase : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) _UpperCamelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) _UpperCamelCase : Optional[bool] = field( default=__lowerCamelCase , metadata={"""help""": """Whether to freeze the feature extractor layers of the model."""} ) _UpperCamelCase : Optional[bool] = field( default=__lowerCamelCase , metadata={"""help""": """Whether to log verbose messages or not."""} , ) _UpperCamelCase : Optional[float] = field( default=2.0 , metadata={"""help""": """Maximum temperature for gumbel softmax."""} ) _UpperCamelCase : Optional[float] = field( default=0.5 , metadata={"""help""": """Minimum temperature for gumbel softmax."""} ) _UpperCamelCase : Optional[float] = field( default=0.999995 , metadata={"""help""": """Decay of gumbel temperature during training."""} ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) lowercase = logging.WARNING if model_args.verbose_logging: lowercase = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): lowercase = logging.INFO logger.setLevel(__SCREAMING_SNAKE_CASE ) @dataclass class A_ : '''simple docstring''' _UpperCamelCase : str = field( default=__lowerCamelCase , metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) _UpperCamelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) _UpperCamelCase : Optional[str] = field( default="""train""" , metadata={ """help""": """The name of the training data set split to use (via the datasets library). Defaults to 'train'""" } , ) _UpperCamelCase : Optional[str] = field( default="""validation""" , metadata={ """help""": ( """The name of the validation data set split to use (via the datasets library). Defaults to 'validation'""" ) } , ) _UpperCamelCase : Optional[str] = field( default="""file""" , metadata={"""help""": """Column in the dataset that contains speech file path. Defaults to 'file'"""} , ) _UpperCamelCase : bool = field( default=__lowerCamelCase , metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) _UpperCamelCase : Optional[int] = field( default=1 , metadata={ """help""": """The percentage of the train set used as validation set in case there's no validation split""" } , ) _UpperCamelCase : Optional[int] = field( default=__lowerCamelCase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) _UpperCamelCase : Optional[float] = field( default=20.0 , metadata={"""help""": """Filter audio files that are longer than `max_duration_in_seconds` seconds"""} ) @dataclass class A_ : '''simple docstring''' _UpperCamelCase : WavaVecaForPreTraining _UpperCamelCase : WavaVecaFeatureExtractor _UpperCamelCase : Union[bool, str] = "longest" _UpperCamelCase : Optional[int] = None _UpperCamelCase : Optional[int] = None def __call__( self , snake_case ): # reformat list to dict and set to pytorch format lowercase = self.feature_extractor.pad( snake_case , max_length=self.max_length , padding=self.padding , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) lowercase = self.model._get_feat_extract_output_lengths(batch['input_values'].shape[-1] ) lowercase = batch['input_values'].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula lowercase = self.model._get_feat_extract_output_lengths(batch['attention_mask'].sum(-1 ) ).to( torch.long ) lowercase = torch.zeros( (batch_size, mask_indices_seq_length) , dtype=torch.long , device=batch['input_values'].device ) # these two operations makes sure that all values # before the output lengths indices are attended to lowercase = 1 lowercase = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices lowercase = _compute_mask_indices( (batch_size, mask_indices_seq_length) , self.model.config.mask_time_prob , self.model.config.mask_time_length , attention_mask=snake_case , min_masks=2 , ) return batch class A_ ( __lowerCamelCase ): '''simple docstring''' def __init__( self , *snake_case , snake_case=1 , snake_case=0 , snake_case=1.0 , **snake_case ): super().__init__(*snake_case , **snake_case ) lowercase = 0 lowercase = max_gumbel_temp lowercase = min_gumbel_temp lowercase = gumbel_temp_decay def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): model.train() lowercase = self._prepare_inputs(snake_case ) if self.use_amp: with autocast(): lowercase = self.compute_loss(snake_case , snake_case ) else: lowercase = self.compute_loss(snake_case , snake_case ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": lowercase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowercase = loss.sum() / (inputs['mask_time_indices']).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowercase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(snake_case ).backward() elif self.use_apex: with amp.scale_loss(snake_case , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(snake_case ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp ) ) return loss.detach() def UpperCAmelCase_ ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowercase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase , lowercase , lowercase = parser.parse_args_into_dataclasses() configure_logger(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Downloading and loading a dataset from the hub. lowercase = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" lowercase = DatasetDict() lowercase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) lowercase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" lowercase = DatasetDict() lowercase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split='validation' , cache_dir=model_args.cache_dir , ) lowercase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported lowercase = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=__SCREAMING_SNAKE_CASE ) def prepare_dataset(__SCREAMING_SNAKE_CASE ): # check that all files have the correct sampling rate lowercase , lowercase = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays lowercase = datasets.map( __SCREAMING_SNAKE_CASE , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets['train'].column_names ) # filter audio files that are too long lowercase = vectorized_datasets.filter( lambda __SCREAMING_SNAKE_CASE : len(data['speech'] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(__SCREAMING_SNAKE_CASE ): return feature_extractor(batch['speech'] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` lowercase = vectorized_datasets.map( __SCREAMING_SNAKE_CASE , batched=__SCREAMING_SNAKE_CASE , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets['train'].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 lowercase = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( 'PreTraining is only supported for ``config.do_stable_layer_norm=True`` and' ' ``config.feat_extract_norm=\'layer\'' ) lowercase = WavaVecaForPreTraining(__SCREAMING_SNAKE_CASE ) lowercase = DataCollatorForWavaVecaPretraining(model=__SCREAMING_SNAKE_CASE , feature_extractor=__SCREAMING_SNAKE_CASE ) lowercase = WavaVecaPreTrainer( model=__SCREAMING_SNAKE_CASE , data_collator=__SCREAMING_SNAKE_CASE , args=__SCREAMING_SNAKE_CASE , train_dataset=vectorized_datasets['train'] , eval_dataset=vectorized_datasets['validation'] , tokenizer=__SCREAMING_SNAKE_CASE , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
195
import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class A_ : '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case ): return None class A_ : '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case ): return None class A_ ( unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Tuple = [ # (model_name, model_kwargs) ("""bert-base-cased""", {}), ("""gpt2""", {"""use_cache""": False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def SCREAMING_SNAKE_CASE__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(snake_case , 'tf' , 12 , **snake_case ) @require_torch @slow def SCREAMING_SNAKE_CASE__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(snake_case , 'pt' , 12 , **snake_case ) @require_torch @slow def SCREAMING_SNAKE_CASE__ ( self ): from transformers import BertModel lowercase = ['[UNK]', '[SEP]', '[CLS]', '[PAD]', '[MASK]', 'some', 'other', 'words'] with NamedTemporaryFile(mode='w+t' ) as vocab_file: vocab_file.write('\n'.join(snake_case ) ) vocab_file.flush() lowercase = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: lowercase = BertModel(BertConfig(vocab_size=len(snake_case ) ) ) model.save_pretrained(snake_case ) self._test_export(snake_case , 'pt' , 12 , snake_case ) @require_tf @slow def SCREAMING_SNAKE_CASE__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: lowercase = self._test_export(snake_case , 'tf' , 12 , **snake_case ) lowercase = quantize(Path(snake_case ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(snake_case ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) @require_torch @slow def SCREAMING_SNAKE_CASE__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: lowercase = self._test_export(snake_case , 'pt' , 12 , **snake_case ) lowercase = quantize(snake_case ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(snake_case ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case=None , **snake_case ): try: # Compute path with TemporaryDirectory() as tempdir: lowercase = Path(snake_case ).joinpath('model.onnx' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(snake_case , snake_case , snake_case , snake_case , snake_case , **snake_case ) return path except Exception as e: self.fail(snake_case ) @require_torch @require_tokenizers @slow def SCREAMING_SNAKE_CASE__ ( self ): from transformers import BertModel lowercase = BertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) lowercase = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(snake_case , snake_case , 'pt' ) @require_tf @require_tokenizers @slow def SCREAMING_SNAKE_CASE__ ( self ): from transformers import TFBertModel lowercase = TFBertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) lowercase = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(snake_case , snake_case , 'tf' ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case ): lowercase = FeatureExtractionPipeline(snake_case , snake_case ) lowercase = ['input_ids', 'token_type_ids', 'attention_mask', 'output_0', 'output_1'] lowercase , lowercase , lowercase , lowercase = infer_shapes(snake_case , snake_case ) # Assert all variables are present self.assertEqual(len(snake_case ) , len(snake_case ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , snake_case ) self.assertSequenceEqual(variable_names[3:] , snake_case ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: 'batch', 1: 'sequence'} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['output_0'] , {0: 'batch', 1: 'sequence'} ) self.assertDictEqual(shapes['output_1'] , {0: 'batch'} ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ['input_ids', 'attention_mask', 'token_type_ids'] lowercase = {'input_ids': [1, 2, 3, 4], 'attention_mask': [0, 0, 0, 0], 'token_type_ids': [1, 1, 1, 1]} lowercase , lowercase = ensure_valid_input(FuncContiguousArgs() , snake_case , snake_case ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(snake_case ) , 3 ) # Should have exactly the same input names self.assertEqual(set(snake_case ) , set(snake_case ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(snake_case , (tokens['input_ids'], tokens['token_type_ids'], tokens['attention_mask']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) lowercase , lowercase = ensure_valid_input(FuncNonContiguousArgs() , snake_case , snake_case ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(snake_case ) , 1 ) self.assertEqual(len(snake_case ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['input_ids'] ) self.assertEqual(ordered_input_names[0] , 'input_ids' ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = generate_identified_filename(Path('/home/something/my_fake_model.onnx' ) , '-test' ) self.assertEqual('/home/something/my_fake_model-test.onnx' , generated.as_posix() )
195
1
'''simple docstring''' from __future__ import annotations from scipy.special import comb # type: ignore class __magic_name__ : def __init__( self : List[Any] ,_UpperCAmelCase : list[tuple[float, float]] ): _a : Optional[Any] = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. _a : List[Any] = len(_UpperCAmelCase ) - 1 def __lowercase ( self : List[str] ,_UpperCAmelCase : float ): assert 0 <= t <= 1, "Time t must be between 0 and 1." _a : list[float] = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree ,_UpperCAmelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(_UpperCAmelCase ) ,5 ) == 1 return output_values def __lowercase ( self : Optional[Any] ,_UpperCAmelCase : float ): assert 0 <= t <= 1, "Time t must be between 0 and 1." _a : Optional[Any] = self.basis_function(_UpperCAmelCase ) _a : Optional[int] = 0.0 _a : str = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def __lowercase ( self : Optional[Any] ,_UpperCAmelCase : float = 0.01 ): from matplotlib import pyplot as plt # type: ignore _a : list[float] = [] # x coordinates of points to plot _a : list[float] = [] # y coordinates of points to plot _a : List[Any] = 0.0 while t <= 1: _a : Tuple = self.bezier_curve_function(_UpperCAmelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size _a : Union[str, Any] = [i[0] for i in self.list_of_points] _a : str = [i[1] for i in self.list_of_points] plt.plot( _UpperCAmelCase ,_UpperCAmelCase ,color='blue' ,label='Curve of Degree ' + str(self.degree ) ,) plt.scatter(_UpperCAmelCase ,_UpperCAmelCase ,color='red' ,label='Control Points' ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
107
'''simple docstring''' import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from tensorflow.python.eager import context from tensorflow.python.framework import ops from transformers import GradientAccumulator, create_optimizer @require_tf class __magic_name__ ( unittest.TestCase ): def __lowercase ( self : int ,_UpperCAmelCase : int ,_UpperCAmelCase : str ,_UpperCAmelCase : Optional[Any] ): self.assertEqual(len(_UpperCAmelCase ) ,len(_UpperCAmelCase ) ) for a, b in zip(_UpperCAmelCase ,_UpperCAmelCase ): self.assertAlmostEqual(_UpperCAmelCase ,_UpperCAmelCase ,delta=_UpperCAmelCase ) def __lowercase ( self : int ): _a : int = GradientAccumulator() accumulator([tf.constant([1.0, 2.0] )] ) accumulator([tf.constant([-2.0, 1.0] )] ) accumulator([tf.constant([-1.0, 2.0] )] ) with self.assertRaises(_UpperCAmelCase ): accumulator([tf.constant([1.0, 1.0] ), tf.constant([2.0, 2.0] )] ) self.assertEqual(accumulator.step ,3 ) self.assertEqual(len(accumulator.gradients ) ,1 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() ,[-2.0, 5.0] ,tol=1E-2 ) accumulator.reset() self.assertEqual(accumulator.step ,0 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() ,[0.0, 0.0] ,tol=1E-2 ) def __lowercase ( self : Any ): _a : int = None ops.enable_eager_execution_internal() _a : Optional[int] = tf.config.list_physical_devices('CPU' ) if len(_UpperCAmelCase ) == 1: tf.config.set_logical_device_configuration( physical_devices[0] ,[tf.config.LogicalDeviceConfiguration(), tf.config.LogicalDeviceConfiguration()] ) _a : Tuple = tf.config.list_logical_devices(device_type='CPU' ) _a : List[str] = tf.distribute.MirroredStrategy(devices=devices[:2] ) with strategy.scope(): _a : Tuple = GradientAccumulator() _a : List[Any] = tf.Variable([4.0, 3.0] ) _a , _a : Dict = create_optimizer(5E-5 ,10 ,5 ) _a : Tuple = tf.Variable([0.0, 0.0] ,trainable=_UpperCAmelCase ) def accumulate_on_replica(_UpperCAmelCase : str ): accumulator([gradient] ) def apply_on_replica(): optimizer.apply_gradients(list(zip(accumulator.gradients ,[variable] ) ) ) @tf.function def accumulate(_UpperCAmelCase : Optional[int] ,_UpperCAmelCase : Optional[int] ): with strategy.scope(): _a : Union[str, Any] = strategy.experimental_local_results(_UpperCAmelCase ) local_variables[0].assign(_UpperCAmelCase ) local_variables[1].assign(_UpperCAmelCase ) strategy.run(_UpperCAmelCase ,args=(gradient_placeholder,) ) @tf.function def apply_grad(): with strategy.scope(): strategy.run(_UpperCAmelCase ) def _check_local_values(_UpperCAmelCase : Tuple ,_UpperCAmelCase : List[str] ): _a : List[Any] = strategy.experimental_local_results(accumulator._gradients[0] ) self.assertListAlmostEqual(values[0].value() ,_UpperCAmelCase ,tol=1E-2 ) self.assertListAlmostEqual(values[1].value() ,_UpperCAmelCase ,tol=1E-2 ) accumulate([1.0, 2.0] ,[-1.0, 1.0] ) accumulate([3.0, -1.0] ,[-1.0, -1.0] ) accumulate([-2.0, 2.0] ,[3.0, -2.0] ) self.assertEqual(accumulator.step ,3 ) _check_local_values([2.0, 3.0] ,[1.0, -2.0] ) apply_grad() self.assertListAlmostEqual(variable.value() ,[4.0, 3.0] ,tol=1E-2 ) accumulator.reset() self.assertEqual(accumulator.step ,0 ) _check_local_values([0.0, 0.0] ,[0.0, 0.0] )
107
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { "facebook/data2vec-text-base": "https://huggingface.co/data2vec/resolve/main/config.json", } class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = 'data2vec-text' def __init__( self , _a=30_522 , _a=768 , _a=12 , _a=12 , _a=3_072 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=2 , _a=0.02 , _a=1E-12 , _a=1 , _a=0 , _a=2 , _a="absolute" , _a=True , _a=None , **_a , ): super().__init__(pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a ) __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = hidden_act __a = intermediate_size __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = max_position_embeddings __a = type_vocab_size __a = initializer_range __a = layer_norm_eps __a = position_embedding_type __a = use_cache __a = classifier_dropout class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def __UpperCAmelCase ( self ): if self.task == "multiple-choice": __a = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: __a = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
45
from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image lowercase__ =['text', 'image', 'audio'] def __UpperCamelCase ( lowerCAmelCase__ : List[str] ): __a : Optional[int] = [] for input_type in input_types: if input_type == "text": inputs.append('''Text input''' ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir('''fixtures/tests_samples/COCO''' ) ) / '''000000039769.png''' ).resize((5_1_2, 5_1_2) ) ) elif input_type == "audio": inputs.append(torch.ones(3_0_0_0 ) ) elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): inputs.append(create_inputs(lowerCAmelCase__ ) ) else: raise ValueError(f"Invalid type requested: {input_type}" ) return inputs def __UpperCamelCase ( lowerCAmelCase__ : List ): __a : List[str] = [] for output in outputs: if isinstance(lowerCAmelCase__ , (str, AgentText) ): output_types.append('''text''' ) elif isinstance(lowerCAmelCase__ , (Image.Image, AgentImage) ): output_types.append('''image''' ) elif isinstance(lowerCAmelCase__ , (torch.Tensor, AgentAudio) ): output_types.append('''audio''' ) else: raise ValueError(f"Invalid output: {output}" ) return output_types @is_tool_test class UpperCamelCase__ : def lowerCAmelCase (self : Any ): self.assertTrue(hasattr(self.tool , '''inputs''' ) ) self.assertTrue(hasattr(self.tool , '''outputs''' ) ) __a : Any = self.tool.inputs for _input in inputs: if isinstance(_input , snake_case_ ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) __a : Optional[int] = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def lowerCAmelCase (self : List[Any] ): __a : Union[str, Any] = create_inputs(self.tool.inputs ) __a : List[Any] = self.tool(*snake_case_ ) # There is a single output if len(self.tool.outputs ) == 1: __a : Tuple = [outputs] self.assertListEqual(output_types(snake_case_ ) , self.tool.outputs ) def lowerCAmelCase (self : List[Any] ): self.assertTrue(hasattr(self.tool , '''description''' ) ) self.assertTrue(hasattr(self.tool , '''default_checkpoint''' ) ) self.assertTrue(self.tool.description.startswith('''This is a tool that''' ) ) def lowerCAmelCase (self : Any ): __a : Any = create_inputs(self.tool.inputs ) __a : Union[str, Any] = self.tool(*snake_case_ ) if not isinstance(snake_case_ , snake_case_ ): __a : Tuple = [outputs] self.assertEqual(len(snake_case_ ) , len(self.tool.outputs ) ) for output, output_type in zip(snake_case_ , self.tool.outputs ): __a : List[Any] = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(snake_case_ , snake_case_ ) ) def lowerCAmelCase (self : Optional[int] ): __a : Any = create_inputs(self.tool.inputs ) __a : Dict = [] for _input, input_type in zip(snake_case_ , self.tool.inputs ): if isinstance(snake_case_ , snake_case_ ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error __a : Optional[Any] = self.tool(*snake_case_ ) if not isinstance(snake_case_ , snake_case_ ): __a : Dict = [outputs] self.assertEqual(len(snake_case_ ) , len(self.tool.outputs ) )
216
0
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy _A = logging.getLogger(__name__) _A = "pytorch_model.bin" @dataclasses.dataclass class lowerCamelCase : UpperCAmelCase__ : str = dataclasses.field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default=A_ , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co."} , ) @dataclasses.dataclass class lowerCamelCase : UpperCAmelCase__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the training data."} ) UpperCAmelCase__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the data to predict on."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default=A_ , metadata={"help": "A csv or a json file containing the validation data."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default=A_ , metadata={"help": "The name of the task to train on."} , ) UpperCAmelCase__ : Optional[List[str]] = dataclasses.field( default=A_ , metadata={"help": "The list of labels for the task."} ) @dataclasses.dataclass class lowerCamelCase : UpperCAmelCase__ : str = dataclasses.field( metadata={"help": "The output directory where the model predictions and checkpoints will be written."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default="accuracy" , metadata={"help": "The evaluation metric used for the task."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default="no" , metadata={ "help": "The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]" } , ) UpperCAmelCase__ : Optional[int] = dataclasses.field( default=10 , metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} , ) UpperCAmelCase__ : Optional[float] = dataclasses.field( default=0.0 , metadata={ "help": "How much the specified evaluation metric must improve to satisfy early stopping conditions." } , ) UpperCAmelCase__ : Optional[bool] = dataclasses.field( default=A_ , metadata={"help": "Whether to filter the pseudo-labeled data based on the confidence score."} , ) UpperCAmelCase__ : Optional[bool] = dataclasses.field( default=A_ , metadata={"help": "Whether to filter the pseudo-labeled data based on the validation performance."} , ) UpperCAmelCase__ : Optional[bool] = dataclasses.field( default=A_ , metadata={"help": "Whether to fine-tune on labeled data after pseudo training."} , ) UpperCAmelCase__ : Optional[float] = dataclasses.field( default=0.0 , metadata={"help": "Confidence threshold for pseudo-labeled data filtering."} , ) UpperCAmelCase__ : Optional[int] = dataclasses.field( default=1_00 , metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} , ) UpperCAmelCase__ : Optional[int] = dataclasses.field( default=A_ , metadata={"help": "Random seed for initialization."} , ) def lowercase_ ( A__ , A__ , A__ , A__ , A__ , A__ ) -> Union[str, Any]: """simple docstring""" snake_case = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: snake_case = dataset.filter(lambda A__ : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 snake_case = int(eval_result * len(A__ ) ) print(A__ ) snake_case = dataset.sort("probability" , reverse=A__ ) snake_case = dataset.select(range(A__ ) ) snake_case = dataset.remove_columns(["label", "probability"] ) snake_case = dataset.rename_column("prediction" , "label" ) snake_case = dataset.map(lambda A__ : {"label": idalabel[example["label"]]} ) snake_case = dataset.shuffle(seed=args.seed ) snake_case = os.path.join(A__ , F'train_pseudo.{args.data_file_extension}' ) if args.data_file_extension == "csv": dataset.to_csv(A__ , index=A__ ) else: dataset.to_json(A__ ) def lowercase_ ( A__ , A__ , A__ , A__ , **A__ ) -> List[Any]: """simple docstring""" snake_case = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() snake_case = STModelArguments(model_name_or_path=A__ ) snake_case = STDataArguments(train_file=A__ , infer_file=A__ ) snake_case = STTrainingArguments(output_dir=A__ ) snake_case = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(A__ ).items(): setattr(A__ , A__ , A__ ) for key, value in kwargs.items(): if hasattr(A__ , A__ ): setattr(A__ , A__ , A__ ) # Sanity checks snake_case = {} snake_case = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None snake_case = args.train_file snake_case = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None snake_case = args.eval_file for key in data_files: snake_case = data_files[key].split("." )[-1] assert extension in ["csv", "json"], F'`{key}_file` should be a csv or a json file.' if args.data_file_extension is None: snake_case = extension else: assert extension == args.data_file_extension, F'`{key}_file` should be a {args.data_file_extension} file`.' assert ( args.eval_metric in datasets.list_metrics() ), F'{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.' # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info("Creating the initial data directory for self-training..." ) snake_case = F'{args.output_dir}/self-train_iter-{{}}'.format snake_case = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=A__ ) os.makedirs(A__ , exist_ok=A__ ) accelerator.wait_for_everyone() snake_case = None snake_case = None snake_case = 0 snake_case = False # Show the progress bar snake_case = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): snake_case = data_dir_format(A__ ) assert os.path.exists(A__ ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 snake_case = os.path.join(A__ , "stage-1" ) snake_case = { "accelerator": accelerator, "model_name_or_path": args.model_name_or_path, "cache_dir": args.cache_dir, "do_train": True, "train_file": data_files["train"] if iteration == 0 else data_files["train_pseudo"], "do_eval": True if args.eval_file is not None else False, "eval_file": data_files["eval"], "do_predict": True, "infer_file": data_files["infer"], "task_name": args.task_name, "label_list": args.label_list, "output_dir": current_output_dir, "eval_metric": args.eval_metric, "evaluation_strategy": args.evaluation_strategy, "early_stopping_patience": args.early_stopping_patience, "early_stopping_threshold": args.early_stopping_threshold, "seed": args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(A__ , A__ ): arguments_dict.update({key: value} ) snake_case = os.path.join(A__ , "best-checkpoint" , A__ ) if os.path.exists(A__ ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1." , A__ , A__ , ) else: logger.info("***** Running self-training: iteration: %d, stage: 1 *****" , A__ ) finetune(**A__ ) accelerator.wait_for_everyone() assert os.path.exists(A__ ) logger.info("Self-training job completed: iteration: %d, stage: 1." , A__ ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data snake_case = os.path.join(A__ , "best-checkpoint" ) snake_case = os.path.join(A__ , "stage-2" ) # Update arguments_dict snake_case = model_path snake_case = data_files["train"] snake_case = current_output_dir snake_case = os.path.join(A__ , "best-checkpoint" , A__ ) if os.path.exists(A__ ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2." , A__ , A__ , ) else: logger.info("***** Running self-training: iteration: %d, stage: 2 *****" , A__ ) finetune(**A__ ) accelerator.wait_for_everyone() assert os.path.exists(A__ ) logger.info("Self-training job completed: iteration: %d, stage: 2." , A__ ) snake_case = iteration snake_case = data_dir_format(iteration + 1 ) snake_case = AutoConfig.from_pretrained(os.path.join(A__ , "best-checkpoint" ) ) snake_case = config.idalabel snake_case = os.path.join(A__ , "eval_results_best-checkpoint.json" ) snake_case = os.path.join(A__ , "test_results_best-checkpoint.json" ) assert os.path.exists(A__ ) with open(A__ , "r" ) as f: snake_case = float(json.load(A__ )[args.eval_metric] ) snake_case = os.path.join(A__ , "infer_output_best-checkpoint.csv" ) assert os.path.exists(A__ ) # Loading the dataset from local csv or json files. snake_case = load_dataset(args.data_file_extension , data_files={"data": data_files["infer"]} )["data"] snake_case = load_dataset("csv" , data_files={"data": infer_output_file} )["data"] if accelerator.is_main_process: os.makedirs(A__ , exist_ok=A__ ) shutil.copy(A__ , os.path.join(A__ , F'eval_results_iter-{iteration}.json' ) ) if os.path.exists(A__ ): shutil.copy(A__ , os.path.join(A__ , F'test_results_iter-{iteration}.json' ) ) create_pseudo_labeled_data(A__ , A__ , A__ , A__ , A__ , A__ ) accelerator.wait_for_everyone() snake_case = os.path.join(A__ , F'train_pseudo.{args.data_file_extension}' ) if args.evaluation_strategy != IntervalStrategy.NO.value: snake_case = eval_result if best_iteration is None: snake_case = new_iteration snake_case = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: snake_case = new_iteration snake_case = new_eval_result snake_case = 0 else: if new_eval_result == best_eval_result: snake_case = new_iteration snake_case = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: snake_case = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info("Best iteration: %d" , A__ ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , A__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(A__ , F'eval_results_iter-{iteration}.json' ) , os.path.join(A__ , "eval_results_best-iteration.json" ) , ) else: # Assume that the last iteration is the best logger.info("Best iteration: %d" , args.max_selftrain_iterations - 1 ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , A__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(A__ , F'eval_results_iter-{args.max_selftrain_iterations - 1}.json' ) , os.path.join(A__ , "eval_results_best-iteration.json" ) , )
137
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy _A = logging.getLogger(__name__) _A = "pytorch_model.bin" @dataclasses.dataclass class lowerCamelCase : UpperCAmelCase__ : str = dataclasses.field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default=A_ , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co."} , ) @dataclasses.dataclass class lowerCamelCase : UpperCAmelCase__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the training data."} ) UpperCAmelCase__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the data to predict on."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default=A_ , metadata={"help": "A csv or a json file containing the validation data."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default=A_ , metadata={"help": "The name of the task to train on."} , ) UpperCAmelCase__ : Optional[List[str]] = dataclasses.field( default=A_ , metadata={"help": "The list of labels for the task."} ) @dataclasses.dataclass class lowerCamelCase : UpperCAmelCase__ : str = dataclasses.field( metadata={"help": "The output directory where the model predictions and checkpoints will be written."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default="accuracy" , metadata={"help": "The evaluation metric used for the task."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default="no" , metadata={ "help": "The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]" } , ) UpperCAmelCase__ : Optional[int] = dataclasses.field( default=10 , metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} , ) UpperCAmelCase__ : Optional[float] = dataclasses.field( default=0.0 , metadata={ "help": "How much the specified evaluation metric must improve to satisfy early stopping conditions." } , ) UpperCAmelCase__ : Optional[bool] = dataclasses.field( default=A_ , metadata={"help": "Whether to filter the pseudo-labeled data based on the confidence score."} , ) UpperCAmelCase__ : Optional[bool] = dataclasses.field( default=A_ , metadata={"help": "Whether to filter the pseudo-labeled data based on the validation performance."} , ) UpperCAmelCase__ : Optional[bool] = dataclasses.field( default=A_ , metadata={"help": "Whether to fine-tune on labeled data after pseudo training."} , ) UpperCAmelCase__ : Optional[float] = dataclasses.field( default=0.0 , metadata={"help": "Confidence threshold for pseudo-labeled data filtering."} , ) UpperCAmelCase__ : Optional[int] = dataclasses.field( default=1_00 , metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} , ) UpperCAmelCase__ : Optional[int] = dataclasses.field( default=A_ , metadata={"help": "Random seed for initialization."} , ) def lowercase_ ( A__ , A__ , A__ , A__ , A__ , A__ ) -> Union[str, Any]: """simple docstring""" snake_case = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: snake_case = dataset.filter(lambda A__ : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 snake_case = int(eval_result * len(A__ ) ) print(A__ ) snake_case = dataset.sort("probability" , reverse=A__ ) snake_case = dataset.select(range(A__ ) ) snake_case = dataset.remove_columns(["label", "probability"] ) snake_case = dataset.rename_column("prediction" , "label" ) snake_case = dataset.map(lambda A__ : {"label": idalabel[example["label"]]} ) snake_case = dataset.shuffle(seed=args.seed ) snake_case = os.path.join(A__ , F'train_pseudo.{args.data_file_extension}' ) if args.data_file_extension == "csv": dataset.to_csv(A__ , index=A__ ) else: dataset.to_json(A__ ) def lowercase_ ( A__ , A__ , A__ , A__ , **A__ ) -> List[Any]: """simple docstring""" snake_case = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() snake_case = STModelArguments(model_name_or_path=A__ ) snake_case = STDataArguments(train_file=A__ , infer_file=A__ ) snake_case = STTrainingArguments(output_dir=A__ ) snake_case = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(A__ ).items(): setattr(A__ , A__ , A__ ) for key, value in kwargs.items(): if hasattr(A__ , A__ ): setattr(A__ , A__ , A__ ) # Sanity checks snake_case = {} snake_case = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None snake_case = args.train_file snake_case = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None snake_case = args.eval_file for key in data_files: snake_case = data_files[key].split("." )[-1] assert extension in ["csv", "json"], F'`{key}_file` should be a csv or a json file.' if args.data_file_extension is None: snake_case = extension else: assert extension == args.data_file_extension, F'`{key}_file` should be a {args.data_file_extension} file`.' assert ( args.eval_metric in datasets.list_metrics() ), F'{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.' # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info("Creating the initial data directory for self-training..." ) snake_case = F'{args.output_dir}/self-train_iter-{{}}'.format snake_case = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=A__ ) os.makedirs(A__ , exist_ok=A__ ) accelerator.wait_for_everyone() snake_case = None snake_case = None snake_case = 0 snake_case = False # Show the progress bar snake_case = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): snake_case = data_dir_format(A__ ) assert os.path.exists(A__ ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 snake_case = os.path.join(A__ , "stage-1" ) snake_case = { "accelerator": accelerator, "model_name_or_path": args.model_name_or_path, "cache_dir": args.cache_dir, "do_train": True, "train_file": data_files["train"] if iteration == 0 else data_files["train_pseudo"], "do_eval": True if args.eval_file is not None else False, "eval_file": data_files["eval"], "do_predict": True, "infer_file": data_files["infer"], "task_name": args.task_name, "label_list": args.label_list, "output_dir": current_output_dir, "eval_metric": args.eval_metric, "evaluation_strategy": args.evaluation_strategy, "early_stopping_patience": args.early_stopping_patience, "early_stopping_threshold": args.early_stopping_threshold, "seed": args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(A__ , A__ ): arguments_dict.update({key: value} ) snake_case = os.path.join(A__ , "best-checkpoint" , A__ ) if os.path.exists(A__ ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1." , A__ , A__ , ) else: logger.info("***** Running self-training: iteration: %d, stage: 1 *****" , A__ ) finetune(**A__ ) accelerator.wait_for_everyone() assert os.path.exists(A__ ) logger.info("Self-training job completed: iteration: %d, stage: 1." , A__ ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data snake_case = os.path.join(A__ , "best-checkpoint" ) snake_case = os.path.join(A__ , "stage-2" ) # Update arguments_dict snake_case = model_path snake_case = data_files["train"] snake_case = current_output_dir snake_case = os.path.join(A__ , "best-checkpoint" , A__ ) if os.path.exists(A__ ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2." , A__ , A__ , ) else: logger.info("***** Running self-training: iteration: %d, stage: 2 *****" , A__ ) finetune(**A__ ) accelerator.wait_for_everyone() assert os.path.exists(A__ ) logger.info("Self-training job completed: iteration: %d, stage: 2." , A__ ) snake_case = iteration snake_case = data_dir_format(iteration + 1 ) snake_case = AutoConfig.from_pretrained(os.path.join(A__ , "best-checkpoint" ) ) snake_case = config.idalabel snake_case = os.path.join(A__ , "eval_results_best-checkpoint.json" ) snake_case = os.path.join(A__ , "test_results_best-checkpoint.json" ) assert os.path.exists(A__ ) with open(A__ , "r" ) as f: snake_case = float(json.load(A__ )[args.eval_metric] ) snake_case = os.path.join(A__ , "infer_output_best-checkpoint.csv" ) assert os.path.exists(A__ ) # Loading the dataset from local csv or json files. snake_case = load_dataset(args.data_file_extension , data_files={"data": data_files["infer"]} )["data"] snake_case = load_dataset("csv" , data_files={"data": infer_output_file} )["data"] if accelerator.is_main_process: os.makedirs(A__ , exist_ok=A__ ) shutil.copy(A__ , os.path.join(A__ , F'eval_results_iter-{iteration}.json' ) ) if os.path.exists(A__ ): shutil.copy(A__ , os.path.join(A__ , F'test_results_iter-{iteration}.json' ) ) create_pseudo_labeled_data(A__ , A__ , A__ , A__ , A__ , A__ ) accelerator.wait_for_everyone() snake_case = os.path.join(A__ , F'train_pseudo.{args.data_file_extension}' ) if args.evaluation_strategy != IntervalStrategy.NO.value: snake_case = eval_result if best_iteration is None: snake_case = new_iteration snake_case = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: snake_case = new_iteration snake_case = new_eval_result snake_case = 0 else: if new_eval_result == best_eval_result: snake_case = new_iteration snake_case = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: snake_case = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info("Best iteration: %d" , A__ ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , A__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(A__ , F'eval_results_iter-{iteration}.json' ) , os.path.join(A__ , "eval_results_best-iteration.json" ) , ) else: # Assume that the last iteration is the best logger.info("Best iteration: %d" , args.max_selftrain_iterations - 1 ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , A__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(A__ , F'eval_results_iter-{args.max_selftrain_iterations - 1}.json' ) , os.path.join(A__ , "eval_results_best-iteration.json" ) , )
137
1
"""simple docstring""" from __future__ import annotations def _SCREAMING_SNAKE_CASE ( _lowercase : list[int] , _lowercase : list[int] , _lowercase : int ) ->tuple[float, list[float]]: '''simple docstring''' a : List[str] = list(range(len(_lowerCamelCase ) ) ) a : int = [v / w for v, w in zip(_lowerCamelCase , _lowerCamelCase )] index.sort(key=lambda _lowercase : ratio[i] , reverse=_lowerCamelCase ) a : float = 0 a : list[float] = [0] * len(_lowerCamelCase ) for i in index: if weight[i] <= capacity: a : Dict = 1 max_value += value[i] capacity -= weight[i] else: a : Any = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
105
"""simple docstring""" from pathlib import PurePosixPath from typing import Optional import fsspec from fsspec import AbstractFileSystem from huggingface_hub.hf_api import DatasetInfo from ..utils.file_utils import get_authentication_headers_for_url from ..utils.hub import hf_hub_url class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : Union[str, Any] = "" _UpperCamelCase : str = "hf-legacy" # "hf://"" is reserved for hffs def __init__( self , a__ = None , a__ = None , **a__ , ): super().__init__(self , **a__ ) _lowerCAmelCase : Any = repo_info _lowerCAmelCase : Optional[Any] = token _lowerCAmelCase : Optional[int] = None def __A ( self ): if self.dir_cache is None: _lowerCAmelCase : Optional[Any] = {} for hf_file in self.repo_info.siblings: # TODO(QL): add sizes _lowerCAmelCase : Any = { """name""": hf_file.rfilename, """size""": None, """type""": """file""", } self.dir_cache.update( { str(a__ ): {"""name""": str(a__ ), """size""": None, """type""": """directory"""} for d in list(PurePosixPath(hf_file.rfilename ).parents )[:-1] } ) def __A ( self , a__ , a__ = "rb" , **a__ , ): if not isinstance(self.repo_info , a__ ): raise NotImplementedError(F"Open is only implemented for dataset repositories, but got {self.repo_info}" ) _lowerCAmelCase : Tuple = hf_hub_url(self.repo_info.id , a__ , revision=self.repo_info.sha ) return fsspec.open( a__ , mode=a__ , headers=get_authentication_headers_for_url(a__ , use_auth_token=self.token ) , client_kwargs={"""trust_env""": True} , ).open() def __A ( self , a__ , **a__ ): self._get_dirs() _lowerCAmelCase : Union[str, Any] = self._strip_protocol(a__ ) if path in self.dir_cache: return self.dir_cache[path] else: raise FileNotFoundError(a__ ) def __A ( self , a__ , a__=False , **a__ ): self._get_dirs() _lowerCAmelCase : Any = PurePosixPath(path.strip("""/""" ) ) _lowerCAmelCase : List[str] = {} for p, f in self.dir_cache.items(): _lowerCAmelCase : Any = PurePosixPath(p.strip("""/""" ) ) _lowerCAmelCase : Optional[int] = p.parent if root == path: _lowerCAmelCase : Dict = f _lowerCAmelCase : Union[str, Any] = list(paths.values() ) if detail: return out else: return sorted(f["""name"""] for f in out )
44
0
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets UpperCamelCase__ = datasets.logging.get_logger(__name__) UpperCamelCase__ = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' UpperCamelCase__ = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' UpperCamelCase__ = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="dummy_doc" ) -> Optional[int]: UpperCAmelCase__ : Any = {doc: key_lines} UpperCAmelCase__ : Dict = {doc: sys_lines} UpperCAmelCase__ : Tuple = {} UpperCAmelCase__ : List[str] = 0 UpperCAmelCase__ : Union[str, Any] = 0 UpperCAmelCase__ : Dict = 0 UpperCAmelCase__ : str = 0 UpperCAmelCase__ : str = 0 UpperCAmelCase__ : Dict = 0 UpperCAmelCase__ , UpperCAmelCase__ : int = reader.get_doc_mentions(lowerCAmelCase__ , key_doc_lines[doc] , lowerCAmelCase__ ) key_singletons_num += singletons_num if NP_only or min_span: UpperCAmelCase__ : Union[str, Any] = reader.set_annotated_parse_trees(lowerCAmelCase__ , key_doc_lines[doc] , lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase__ , UpperCAmelCase__ : str = reader.get_doc_mentions(lowerCAmelCase__ , sys_doc_lines[doc] , lowerCAmelCase__ ) sys_singletons_num += singletons_num if NP_only or min_span: UpperCAmelCase__ : str = reader.set_annotated_parse_trees(lowerCAmelCase__ , key_doc_lines[doc] , lowerCAmelCase__ , lowerCAmelCase__ ) if remove_nested: UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = reader.remove_nested_coref_mentions(lowerCAmelCase__ , lowerCAmelCase__ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters UpperCAmelCase__ , UpperCAmelCase__ : Dict = reader.remove_nested_coref_mentions(lowerCAmelCase__ , lowerCAmelCase__ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters UpperCAmelCase__ : Dict = reader.get_mention_assignments(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase__ : Dict = reader.get_mention_assignments(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase__ : Dict = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' F"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( '''Number of resulting singleton clusters in the key ''' F"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( F"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ '''files, respectively''' ) return doc_coref_infos def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: UpperCAmelCase__ : Optional[int] = get_coref_infos(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase__ : int = {} UpperCAmelCase__ : Dict = 0 UpperCAmelCase__ : Tuple = 0 for name, metric in metrics: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : str = evaluator.evaluate_documents(lowerCAmelCase__ , lowerCAmelCase__ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F"""{name}/recall""": recall, F"""{name}/precision""": precision, F"""{name}/f1""": fa} ) logger.info( name.ljust(10 ) , F"""Recall: {recall * 1_00:.2f}""" , F""" Precision: {precision * 1_00:.2f}""" , F""" F1: {fa * 1_00:.2f}""" , ) if conll_subparts_num == 3: UpperCAmelCase__ : Union[str, Any] = (conll / 3) * 1_00 logger.info(F"""CoNLL score: {conll:.2f}""" ) output_scores.update({'''conll_score''': conll} ) return output_scores def a__ ( lowerCAmelCase__ ) -> Dict: UpperCAmelCase__ : Tuple = False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: UpperCAmelCase__ : Optional[int] = line.split()[5] if not parse_col == "-": UpperCAmelCase__ : str = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase_ ( datasets.Metric ): def lowercase_ ( self : str ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def lowercase_ ( self : str , _A : Union[str, Any] , _A : Dict , _A : str=True , _A : int=False , _A : str=False , _A : Union[str, Any]=False ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = [ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: UpperCAmelCase__ : List[Any] = util.check_gold_parse_annotation(_A ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" UpperCAmelCase__ : str = evaluate( key_lines=_A , sys_lines=_A , metrics=_A , NP_only=_A , remove_nested=_A , keep_singletons=_A , min_span=_A , ) return score
299
'''simple docstring''' from __future__ import annotations import math def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: if depth < 0: raise ValueError('''Depth cannot be less than 0''' ) if not scores: raise ValueError('''Scores cannot be empty''' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) , minimax(depth + 1 , node_index * 2 + 1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) , minimax(depth + 1 , node_index * 2 + 1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) , ) ) def a__ ( ) -> None: UpperCAmelCase__ : Union[str, Any] = [90, 23, 6, 33, 21, 65, 1_23, 3_44_23] UpperCAmelCase__ : Optional[Any] = math.log(len(lowerCAmelCase__ ) , 2 ) print(F"""Optimal value : {minimax(0 , 0 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ )}""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
299
1
from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( """The RoBERTa Model transformer with early exiting (DeeRoBERTa). """ , UpperCAmelCase_ , ) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[int] = RobertaConfig a__ : List[str] = """roberta""" def __init__( self , __lowercase) -> str: super().__init__(__lowercase) __UpperCamelCase :Dict = RobertaEmbeddings(__lowercase) self.init_weights() @add_start_docstrings( """RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top, also takes care of multi-layer training. """ , UpperCAmelCase_ , ) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : List[Any] = RobertaConfig a__ : str = """roberta""" def __init__( self , __lowercase) -> Any: super().__init__(__lowercase) __UpperCamelCase :List[str] = config.num_labels __UpperCamelCase :List[Any] = config.num_hidden_layers __UpperCamelCase :Optional[Any] = DeeRobertaModel(__lowercase) __UpperCamelCase :Any = nn.Dropout(config.hidden_dropout_prob) __UpperCamelCase :str = nn.Linear(config.hidden_size , self.config.num_labels) @add_start_docstrings_to_model_forward(__lowercase) def UpperCamelCase__ ( self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=-1 , __lowercase=False , ) -> List[Any]: __UpperCamelCase :List[str] = self.num_layers try: __UpperCamelCase :Tuple = self.roberta( __lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , position_ids=__lowercase , head_mask=__lowercase , inputs_embeds=__lowercase , ) __UpperCamelCase :int = outputs[1] __UpperCamelCase :Optional[int] = self.dropout(__lowercase) __UpperCamelCase :Any = self.classifier(__lowercase) __UpperCamelCase :Tuple = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: __UpperCamelCase :Optional[Any] = e.message __UpperCamelCase :List[str] = e.exit_layer __UpperCamelCase :Any = outputs[0] if not self.training: __UpperCamelCase :Optional[Any] = entropy(__lowercase) __UpperCamelCase :Dict = [] __UpperCamelCase :int = [] if labels is not None: if self.num_labels == 1: # We are doing regression __UpperCamelCase :Dict = MSELoss() __UpperCamelCase :str = loss_fct(logits.view(-1) , labels.view(-1)) else: __UpperCamelCase :Optional[Any] = CrossEntropyLoss() __UpperCamelCase :Any = loss_fct(logits.view(-1 , self.num_labels) , labels.view(-1)) # work with highway exits __UpperCamelCase :List[Any] = [] for highway_exit in outputs[-1]: __UpperCamelCase :Any = highway_exit[0] if not self.training: highway_logits_all.append(__lowercase) highway_entropy.append(highway_exit[2]) if self.num_labels == 1: # We are doing regression __UpperCamelCase :Any = MSELoss() __UpperCamelCase :Optional[Any] = loss_fct(highway_logits.view(-1) , labels.view(-1)) else: __UpperCamelCase :Union[str, Any] = CrossEntropyLoss() __UpperCamelCase :Union[str, Any] = loss_fct(highway_logits.view(-1 , self.num_labels) , labels.view(-1)) highway_losses.append(__lowercase) if train_highway: __UpperCamelCase :Any = (sum(highway_losses[:-1]),) + outputs # exclude the final highway, of course else: __UpperCamelCase :Union[str, Any] = (loss,) + outputs if not self.training: __UpperCamelCase :Tuple = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: __UpperCamelCase :List[str] = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
43
import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def __lowercase ( lowerCamelCase : str , lowerCamelCase : str , **lowerCamelCase : List[Any] ): UpperCamelCase_ : Union[str, Any] = AutoConfig.from_pretrained(lowerCamelCase , **lowerCamelCase ) UpperCamelCase_ : str = AutoModelForSeqaSeqLM.from_config(lowerCamelCase ) model.save_pretrained(lowerCamelCase ) AutoTokenizer.from_pretrained(lowerCamelCase ).save_pretrained(lowerCamelCase ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
175
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: UpperCamelCase = None UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} UpperCamelCase = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", }, """tokenizer_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json""", }, } UpperCamelCase = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } UpperCamelCase = """▁""" class snake_case_ ( __UpperCamelCase ): __A : List[Any] = VOCAB_FILES_NAMES __A : str = PRETRAINED_VOCAB_FILES_MAP __A : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A : List[Any] = AlbertTokenizer def __init__( self : Optional[Any] , lowercase_ : Optional[Any]=None , lowercase_ : List[Any]=None , lowercase_ : Tuple=True , lowercase_ : List[Any]=True , lowercase_ : Union[str, Any]=False , lowercase_ : Dict="[CLS]" , lowercase_ : List[Any]="[SEP]" , lowercase_ : str="<unk>" , lowercase_ : Any="[SEP]" , lowercase_ : List[str]="<pad>" , lowercase_ : Union[str, Any]="[CLS]" , lowercase_ : int="[MASK]" , **lowercase_ : Tuple , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. lowercase__ : Optional[int] = ( AddedToken(lowercase_ , lstrip=lowercase_ , rstrip=lowercase_ , normalized=lowercase_ ) if isinstance(lowercase_ , lowercase_ ) else mask_token ) super().__init__( lowercase_ , tokenizer_file=lowercase_ , do_lower_case=lowercase_ , remove_space=lowercase_ , keep_accents=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , unk_token=lowercase_ , sep_token=lowercase_ , pad_token=lowercase_ , cls_token=lowercase_ , mask_token=lowercase_ , **lowercase_ , ) lowercase__ : Dict = do_lower_case lowercase__ : Any = remove_space lowercase__ : str = keep_accents lowercase__ : Any = vocab_file lowercase__ : str = False if not self.vocab_file else True def __UpperCamelCase ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None ) -> List[int]: lowercase__ : Tuple = [self.sep_token_id] lowercase__ : List[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __UpperCamelCase ( self : Tuple , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None ) -> List[int]: lowercase__ : Optional[int] = [self.sep_token_id] lowercase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCamelCase ( self : Optional[int] , lowercase_ : str , lowercase_ : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowercase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase__ : Tuple = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase_ ): copyfile(self.vocab_file , lowercase_ ) return (out_vocab_file,)
351
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { '''microsoft/unispeech-large-1500h-cv''': ( '''https://huggingface.co/microsoft/unispeech-large-1500h-cv/resolve/main/config.json''' ), # See all UniSpeech models at https://huggingface.co/models?filter=unispeech } class snake_case_ ( __A ): __A : List[str] = "unispeech" def __init__( self : List[Any] , lowercase_ : Optional[int]=32 , lowercase_ : Optional[int]=7_68 , lowercase_ : List[str]=12 , lowercase_ : Union[str, Any]=12 , lowercase_ : Union[str, Any]=30_72 , lowercase_ : List[Any]="gelu" , lowercase_ : int=0.1 , lowercase_ : Union[str, Any]=0.1 , lowercase_ : str=0.1 , lowercase_ : Union[str, Any]=0.0 , lowercase_ : List[str]=0.0 , lowercase_ : List[Any]=0.1 , lowercase_ : Any=0.1 , lowercase_ : Optional[Any]=0.02 , lowercase_ : int=1E-5 , lowercase_ : int="group" , lowercase_ : Tuple="gelu" , lowercase_ : Dict=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , lowercase_ : Union[str, Any]=(5, 2, 2, 2, 2, 2, 2) , lowercase_ : List[str]=(10, 3, 3, 3, 3, 2, 2) , lowercase_ : int=False , lowercase_ : List[Any]=1_28 , lowercase_ : Optional[Any]=16 , lowercase_ : Union[str, Any]=False , lowercase_ : Tuple=True , lowercase_ : Union[str, Any]=0.05 , lowercase_ : Optional[Any]=10 , lowercase_ : Any=2 , lowercase_ : int=0.0 , lowercase_ : Union[str, Any]=10 , lowercase_ : Optional[Any]=0 , lowercase_ : List[str]=3_20 , lowercase_ : Dict=2 , lowercase_ : Optional[int]=0.1 , lowercase_ : Tuple=1_00 , lowercase_ : Dict=2_56 , lowercase_ : Optional[Any]=2_56 , lowercase_ : Union[str, Any]=0.1 , lowercase_ : List[Any]="mean" , lowercase_ : Union[str, Any]=False , lowercase_ : Tuple=False , lowercase_ : Dict=2_56 , lowercase_ : Union[str, Any]=80 , lowercase_ : int=0 , lowercase_ : Union[str, Any]=1 , lowercase_ : Dict=2 , lowercase_ : Optional[int]=0.5 , **lowercase_ : Union[str, Any] , ) -> Any: super().__init__(**lowercase_ , pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ ) lowercase__ : List[str] = hidden_size lowercase__ : Any = feat_extract_norm lowercase__ : Optional[Any] = feat_extract_activation lowercase__ : Dict = list(lowercase_ ) lowercase__ : Union[str, Any] = list(lowercase_ ) lowercase__ : List[str] = list(lowercase_ ) lowercase__ : List[str] = conv_bias lowercase__ : Any = num_conv_pos_embeddings lowercase__ : Dict = num_conv_pos_embedding_groups lowercase__ : int = len(self.conv_dim ) lowercase__ : str = num_hidden_layers lowercase__ : Any = intermediate_size lowercase__ : Optional[int] = hidden_act lowercase__ : int = num_attention_heads lowercase__ : Union[str, Any] = hidden_dropout lowercase__ : Any = attention_dropout lowercase__ : Union[str, Any] = activation_dropout lowercase__ : Any = feat_proj_dropout lowercase__ : str = final_dropout lowercase__ : int = layerdrop lowercase__ : Optional[int] = layer_norm_eps lowercase__ : List[Any] = initializer_range lowercase__ : Any = num_ctc_classes lowercase__ : int = vocab_size lowercase__ : str = do_stable_layer_norm lowercase__ : Any = use_weighted_layer_sum lowercase__ : Dict = classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==" " `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =" F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowercase__ : List[Any] = apply_spec_augment lowercase__ : Dict = mask_time_prob lowercase__ : Tuple = mask_time_length lowercase__ : str = mask_time_min_masks lowercase__ : List[Any] = mask_feature_prob lowercase__ : int = mask_feature_length lowercase__ : Optional[int] = mask_feature_min_masks # parameters for pretraining with codevector quantized representations lowercase__ : Optional[int] = num_codevectors_per_group lowercase__ : List[str] = num_codevector_groups lowercase__ : Dict = contrastive_logits_temperature lowercase__ : Tuple = feat_quantizer_dropout lowercase__ : Any = num_negatives lowercase__ : Dict = codevector_dim lowercase__ : Tuple = proj_codevector_dim lowercase__ : List[str] = diversity_loss_weight # ctc loss lowercase__ : Tuple = ctc_loss_reduction lowercase__ : Dict = ctc_zero_infinity # pretraining loss lowercase__ : Optional[Any] = replace_prob @property def __UpperCamelCase ( self : Dict ) -> Tuple: return functools.reduce(operator.mul , self.conv_stride , 1 )
333
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase__ = { '''configuration_squeezebert''': [ '''SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SqueezeBertConfig''', '''SqueezeBertOnnxConfig''', ], '''tokenization_squeezebert''': ['''SqueezeBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = ['''SqueezeBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ '''SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SqueezeBertForMaskedLM''', '''SqueezeBertForMultipleChoice''', '''SqueezeBertForQuestionAnswering''', '''SqueezeBertForSequenceClassification''', '''SqueezeBertForTokenClassification''', '''SqueezeBertModel''', '''SqueezeBertModule''', '''SqueezeBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
5
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCAmelCase__ = { '''configuration_efficientnet''': [ '''EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''EfficientNetConfig''', '''EfficientNetOnnxConfig''', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = ['''EfficientNetImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ '''EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''EfficientNetForImageClassification''', '''EfficientNetModel''', '''EfficientNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
5
1
"""simple docstring""" import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class __magic_name__ ( lowerCamelCase_ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase = MvpTokenizer __UpperCamelCase = MvpTokenizerFast __UpperCamelCase = True __UpperCamelCase = filter_roberta_detectors def _lowerCAmelCase ( self ): """simple docstring""" super().setUp() lowerCamelCase = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] lowerCamelCase = dict(zip(__snake_case , range(len(__snake_case ) ) ) ) lowerCamelCase = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] lowerCamelCase = {"""unk_token""": """<unk>"""} lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__snake_case ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__snake_case ) ) def _lowerCAmelCase ( self , **_a ): """simple docstring""" kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **__snake_case ) def _lowerCAmelCase ( self , **_a ): """simple docstring""" kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **__snake_case ) def _lowerCAmelCase ( self , _a ): """simple docstring""" return "lower newer", "lower newer" @cached_property def _lowerCAmelCase ( self ): """simple docstring""" return MvpTokenizer.from_pretrained("""RUCAIBox/mvp""" ) @cached_property def _lowerCAmelCase ( self ): """simple docstring""" return MvpTokenizerFast.from_pretrained("""RUCAIBox/mvp""" ) @require_torch def _lowerCAmelCase ( self ): """simple docstring""" lowerCamelCase = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] lowerCamelCase = [0, 250, 251, 17_818, 13, 39_186, 1_938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCamelCase = tokenizer(__snake_case , max_length=len(__snake_case ) , padding=__snake_case , return_tensors="""pt""" ) self.assertIsInstance(__snake_case , __snake_case ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) lowerCamelCase = batch.input_ids.tolist()[0] self.assertListEqual(__snake_case , __snake_case ) # Test that special tokens are reset @require_torch def _lowerCAmelCase ( self ): """simple docstring""" lowerCamelCase = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCamelCase = tokenizer(__snake_case , padding=__snake_case , return_tensors="""pt""" ) # check if input_ids are returned and no labels self.assertIn("""input_ids""" , __snake_case ) self.assertIn("""attention_mask""" , __snake_case ) self.assertNotIn("""labels""" , __snake_case ) self.assertNotIn("""decoder_attention_mask""" , __snake_case ) @require_torch def _lowerCAmelCase ( self ): """simple docstring""" lowerCamelCase = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCamelCase = tokenizer(text_target=__snake_case , max_length=32 , padding="""max_length""" , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) @require_torch def _lowerCAmelCase ( self ): """simple docstring""" for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCamelCase = tokenizer( ["""I am a small frog""" * 1_024, """I am a small frog"""] , padding=__snake_case , truncation=__snake_case , return_tensors="""pt""" ) self.assertIsInstance(__snake_case , __snake_case ) self.assertEqual(batch.input_ids.shape , (2, 1_024) ) @require_torch def _lowerCAmelCase ( self ): """simple docstring""" lowerCamelCase = ["""A long paragraph for summarization."""] lowerCamelCase = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCamelCase = tokenizer(__snake_case , text_target=__snake_case , return_tensors="""pt""" ) lowerCamelCase = inputs["""input_ids"""] lowerCamelCase = inputs["""labels"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) def _lowerCAmelCase ( self ): """simple docstring""" pass def _lowerCAmelCase ( self ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})' ): lowerCamelCase = self.rust_tokenizer_class.from_pretrained(__snake_case , **__snake_case ) lowerCamelCase = self.tokenizer_class.from_pretrained(__snake_case , **__snake_case ) lowerCamelCase = """A, <mask> AllenNLP sentence.""" lowerCamelCase = tokenizer_r.encode_plus(__snake_case , add_special_tokens=__snake_case , return_token_type_ids=__snake_case ) lowerCamelCase = tokenizer_p.encode_plus(__snake_case , add_special_tokens=__snake_case , return_token_type_ids=__snake_case ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) lowerCamelCase = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) lowerCamelCase = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual( __snake_case , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( __snake_case , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] )
371
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase : List[str] = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = {"""tokenizer_file""": """tokenizer.json"""} lowerCAmelCase : Tuple = { """tokenizer_file""": { """bigscience/tokenizer""": """https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json""", """bigscience/bloom-560m""": """https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json""", """bigscience/bloom-1b1""": """https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json""", """bigscience/bloom-1b7""": """https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json""", """bigscience/bloom-3b""": """https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json""", """bigscience/bloom-7b1""": """https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json""", """bigscience/bloom""": """https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json""", }, } class __magic_name__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = ["input_ids", "attention_mask"] __UpperCamelCase = None def __init__( self , _a=None , _a=None , _a=None , _a="<unk>" , _a="<s>" , _a="</s>" , _a="<pad>" , _a=False , _a=False , **_a , ): """simple docstring""" super().__init__( _a , _a , tokenizer_file=_a , unk_token=_a , bos_token=_a , eos_token=_a , pad_token=_a , add_prefix_space=_a , clean_up_tokenization_spaces=_a , **_a , ) lowerCamelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , _a ) != add_prefix_space: lowerCamelCase = getattr(_a , pre_tok_state.pop("""type""" ) ) lowerCamelCase = add_prefix_space lowerCamelCase = pre_tok_class(**_a ) lowerCamelCase = add_prefix_space def _lowerCAmelCase ( self , *_a , **_a ): """simple docstring""" lowerCamelCase = kwargs.get("""is_split_into_words""" , _a ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._batch_encode_plus(*_a , **_a ) def _lowerCAmelCase ( self , *_a , **_a ): """simple docstring""" lowerCamelCase = kwargs.get("""is_split_into_words""" , _a ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._encode_plus(*_a , **_a ) def _lowerCAmelCase ( self , _a , _a = None ): """simple docstring""" lowerCamelCase = self._tokenizer.model.save(_a , name=_a ) return tuple(_a ) def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(_a , add_special_tokens=_a ) + [self.eos_token_id] ) if len(_a ) > self.model_max_length: lowerCamelCase = input_ids[-self.model_max_length :] return input_ids
168
0
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): """simple docstring""" _snake_case = [[1, 2, 4], [1, 2, 3, 4]] _snake_case = DisjunctiveConstraint(lowerCAmelCase_ ) self.assertTrue(isinstance(dc.token_ids , lowerCAmelCase_ ) ) with self.assertRaises(lowerCAmelCase_ ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(lowerCAmelCase_ ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(lowerCAmelCase_ ): DisjunctiveConstraint(lowerCAmelCase_ ) # fails here def lowerCamelCase ( self ): """simple docstring""" _snake_case = [[1, 2, 3], [1, 2, 4]] _snake_case = DisjunctiveConstraint(lowerCAmelCase_ ) _snake_case , _snake_case , _snake_case = dc.update(1 ) _snake_case = stepped is True and completed is False and reset is False self.assertTrue(lowerCAmelCase_ ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) _snake_case , _snake_case , _snake_case = dc.update(2 ) _snake_case = stepped is True and completed is False and reset is False self.assertTrue(lowerCAmelCase_ ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) _snake_case , _snake_case , _snake_case = dc.update(3 ) _snake_case = stepped is True and completed is True and reset is False self.assertTrue(lowerCAmelCase_ ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] _snake_case = DisjunctiveConstraint(lowerCAmelCase_ ) _snake_case , _snake_case , _snake_case = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) _snake_case , _snake_case , _snake_case = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) _snake_case , _snake_case , _snake_case = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) _snake_case , _snake_case , _snake_case = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() _snake_case , _snake_case , _snake_case = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) _snake_case , _snake_case , _snake_case = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) _snake_case , _snake_case , _snake_case = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
42
import os # Precomputes a list of the 100 first triangular numbers a : Optional[Any] = [int(0.5 * n * (n + 1)) for n in range(1, 101)] def lowerCAmelCase_ (): """simple docstring""" UpperCAmelCase_: Any = os.path.dirname(os.path.realpath(lowerCAmelCase__ ) ) UpperCAmelCase_: List[Any] = os.path.join(lowerCAmelCase__ , """words.txt""" ) UpperCAmelCase_: int = """""" with open(lowerCAmelCase__ ) as f: UpperCAmelCase_: str = f.readline() UpperCAmelCase_: Optional[int] = [word.strip("""\"""" ) for word in words.strip("""\r\n""" ).split(""",""" )] UpperCAmelCase_: Union[str, Any] = [ word for word in [sum(ord(lowerCAmelCase__ ) - 6_4 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(lowerCAmelCase__ ) if __name__ == "__main__": print(solution())
147
0
"""simple docstring""" class __lowerCamelCase : '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None ) -> Optional[Any]: _a = data _a = previous _a = next_node def __str__( self ) -> str: return F'{self.data}' def _UpperCAmelCase ( self ) -> int: return self.data def _UpperCAmelCase ( self ) -> str: return self.next def _UpperCAmelCase ( self ) -> Dict: return self.previous class __lowerCamelCase : '''simple docstring''' def __init__( self , __UpperCAmelCase ) -> int: _a = head def __iter__( self ) -> Tuple: return self def _UpperCAmelCase ( self ) -> Dict: if not self.current: raise StopIteration else: _a = self.current.get_data() _a = self.current.get_next() return value class __lowerCamelCase : '''simple docstring''' def __init__( self ) -> Any: _a = None # First node in list _a = None # Last node in list def __str__( self ) -> Optional[Any]: _a = self.head _a = [] while current is not None: nodes.append(current.get_data() ) _a = current.get_next() return " ".join(str(__UpperCAmelCase ) for node in nodes ) def __contains__( self , __UpperCAmelCase ) -> Dict: _a = self.head while current: if current.get_data() == value: return True _a = current.get_next() return False def __iter__( self ) -> int: return LinkedListIterator(self.head ) def _UpperCAmelCase ( self ) -> List[Any]: if self.head: return self.head.get_data() return None def _UpperCAmelCase ( self ) -> Dict: if self.tail: return self.tail.get_data() return None def _UpperCAmelCase ( self , __UpperCAmelCase ) -> None: if self.head is None: _a = node _a = node else: self.insert_before_node(self.head , __UpperCAmelCase ) def _UpperCAmelCase ( self , __UpperCAmelCase ) -> None: if self.head is None: self.set_head(__UpperCAmelCase ) else: self.insert_after_node(self.tail , __UpperCAmelCase ) def _UpperCAmelCase ( self , __UpperCAmelCase ) -> None: _a = Node(__UpperCAmelCase ) if self.head is None: self.set_head(__UpperCAmelCase ) else: self.set_tail(__UpperCAmelCase ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ) -> None: _a = node _a = node.previous if node.get_previous() is None: _a = node_to_insert else: _a = node_to_insert _a = node_to_insert def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ) -> None: _a = node _a = node.next if node.get_next() is None: _a = node_to_insert else: _a = node_to_insert _a = node_to_insert def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ) -> None: _a = 1 _a = Node(__UpperCAmelCase ) _a = self.head while node: if current_position == position: self.insert_before_node(__UpperCAmelCase , __UpperCAmelCase ) return current_position += 1 _a = node.next self.insert_after_node(self.tail , __UpperCAmelCase ) def _UpperCAmelCase ( self , __UpperCAmelCase ) -> Node: _a = self.head while node: if node.get_data() == item: return node _a = node.get_next() raise Exception('''Node not found''' ) def _UpperCAmelCase ( self , __UpperCAmelCase ) -> int: if (node := self.get_node(__UpperCAmelCase )) is not None: if node == self.head: _a = self.head.get_next() if node == self.tail: _a = self.tail.get_previous() self.remove_node_pointers(__UpperCAmelCase ) @staticmethod def _UpperCAmelCase ( __UpperCAmelCase ) -> None: if node.get_next(): _a = node.previous if node.get_previous(): _a = node.next _a = None _a = None def _UpperCAmelCase ( self ) -> Union[str, Any]: return self.head is None def A_ ( ): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
153
"""simple docstring""" from collections.abc import Callable import numpy as np def A_ ( _lowerCAmelCase : Callable, _lowerCAmelCase : float, _lowerCAmelCase : float, _lowerCAmelCase : float, _lowerCAmelCase : float ): """simple docstring""" _a = int(np.ceil((x_end - xa) / step_size ) ) _a = np.zeros((n + 1,) ) _a = ya _a = xa for k in range(_lowerCAmelCase ): _a = y[k] + step_size * ode_func(_lowerCAmelCase, y[k] ) _a = y[k] + ( (step_size / 2) * (ode_func(_lowerCAmelCase, y[k] ) + ode_func(x + step_size, _lowerCAmelCase )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
153
1
'''simple docstring''' from collections import defaultdict class snake_case__ : def __init__( self : str , _A : Dict , _A : int ) -> List[Any]: UpperCAmelCase_ : Optional[Any] = total # total no of tasks (N) # DP table will have a dimension of (2^M)*N # initially all values are set to -1 UpperCAmelCase_ : Union[str, Any] = [ [-1 for i in range(total + 1 )] for j in range(2 ** len(_A ) ) ] UpperCAmelCase_ : Optional[Any] = defaultdict(_A ) # stores the list of persons for each task # final_mask is used to check if all persons are included by setting all bits # to 1 UpperCAmelCase_ : Optional[int] = (1 << len(_A )) - 1 def A ( self : Any , _A : int , _A : List[str] ) -> Tuple: # if mask == self.finalmask all persons are distributed tasks, return 1 if mask == self.final_mask: return 1 # if not everyone gets the task and no more tasks are available, return 0 if task_no > self.total_tasks: return 0 # if case already considered if self.dp[mask][task_no] != -1: return self.dp[mask][task_no] # Number of ways when we don't this task in the arrangement UpperCAmelCase_ : Tuple = self.count_ways_until(_A , task_no + 1 ) # now assign the tasks one by one to all possible persons and recursively # assign for the remaining tasks. if task_no in self.task: for p in self.task[task_no]: # if p is already given a task if mask & (1 << p): continue # assign this task to p and change the mask value. And recursively # assign tasks with the new mask value. total_ways_util += self.count_ways_until(mask | (1 << p) , task_no + 1 ) # save the value. UpperCAmelCase_ : List[Any] = total_ways_util return self.dp[mask][task_no] def A ( self : Dict , _A : Tuple ) -> List[str]: # Store the list of persons for each task for i in range(len(_A ) ): for j in task_performed[i]: self.task[j].append(_A ) # call the function to fill the DP table, final answer is stored in dp[0][1] return self.count_ways_until(0 , 1 ) if __name__ == "__main__": _UpperCamelCase : Dict = 5 # total no of tasks (the value of N) # the list of tasks that can be done by M persons. _UpperCamelCase : Union[str, Any] = [[1, 3, 4], [1, 2, 5], [3, 4]] print( AssignmentUsingBitmask(task_performed, total_tasks).count_no_of_ways( task_performed ) )
304
'''simple docstring''' _UpperCamelCase : Tuple = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' _UpperCamelCase : Any = [{'type': 'code', 'content': INSTALL_CONTENT}] _UpperCamelCase : Dict = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
304
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def UpperCAmelCase_ ( __UpperCAmelCase : Dict ) -> Tuple: SCREAMING_SNAKE_CASE_ = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = StableDiffusionLatentUpscalePipeline lowercase_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "height", "width", "cross_attention_kwargs", "negative_prompt_embeds", "prompt_embeds", } lowercase_ = PipelineTesterMixin.required_optional_params - {"num_images_per_prompt"} lowercase_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase_ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowercase_ = frozenset([] ) lowercase_ = True @property def lowerCAmelCase_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = 4 SCREAMING_SNAKE_CASE_ = (16, 16) SCREAMING_SNAKE_CASE_ = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(_lowerCAmelCase ) return image def lowerCAmelCase_ ( self : Tuple ): torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ = UNetaDConditionModel( act_fn='gelu' , attention_head_dim=8 , norm_num_groups=_lowerCAmelCase , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=160 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( 'KDownBlock2D', 'KCrossAttnDownBlock2D', 'KCrossAttnDownBlock2D', 'KCrossAttnDownBlock2D', ) , in_channels=8 , mid_block_type=_lowerCAmelCase , only_cross_attention=_lowerCAmelCase , out_channels=5 , resnet_time_scale_shift='scale_shift' , time_embedding_type='fourier' , timestep_post_act='gelu' , up_block_types=('KCrossAttnUpBlock2D', 'KCrossAttnUpBlock2D', 'KCrossAttnUpBlock2D', 'KUpBlock2D') , ) SCREAMING_SNAKE_CASE_ = AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', ] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) SCREAMING_SNAKE_CASE_ = EulerDiscreteScheduler(prediction_type='sample' ) SCREAMING_SNAKE_CASE_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='quick_gelu' , projection_dim=512 , ) SCREAMING_SNAKE_CASE_ = CLIPTextModel(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) SCREAMING_SNAKE_CASE_ = { 'unet': model.eval(), 'vae': vae.eval(), 'scheduler': scheduler, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def lowerCAmelCase_ ( self : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[Any]=0 ): if str(_lowerCAmelCase ).startswith('mps' ): SCREAMING_SNAKE_CASE_ = torch.manual_seed(_lowerCAmelCase ) else: SCREAMING_SNAKE_CASE_ = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = { 'prompt': 'A painting of a squirrel eating a burger', 'image': self.dummy_image.cpu(), 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def lowerCAmelCase_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE_ = 'cpu' SCREAMING_SNAKE_CASE_ = self.get_dummy_components() SCREAMING_SNAKE_CASE_ = self.pipeline_class(**_lowerCAmelCase ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = self.get_dummy_inputs(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = pipe(**_lowerCAmelCase ).images SCREAMING_SNAKE_CASE_ = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 256, 256, 3) ) SCREAMING_SNAKE_CASE_ = np.array( [0.4722_2412, 0.4192_1633, 0.4471_7434, 0.4687_4192, 0.4258_8258, 0.4615_0726, 0.467_7534, 0.4558_3832, 0.4857_9055] ) SCREAMING_SNAKE_CASE_ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_lowerCAmelCase , 1E-3 ) def lowerCAmelCase_ ( self : List[Any] ): super().test_attention_slicing_forward_pass(expected_max_diff=7E-3 ) def lowerCAmelCase_ ( self : Dict ): super().test_cpu_offload_forward_pass(expected_max_diff=3E-3 ) def lowerCAmelCase_ ( self : Optional[int] ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def lowerCAmelCase_ ( self : int ): super().test_inference_batch_single_identical(expected_max_diff=7E-3 ) def lowerCAmelCase_ ( self : Dict ): super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3E-3 ) def lowerCAmelCase_ ( self : Any ): super().test_save_load_local(expected_max_difference=3E-3 ) def lowerCAmelCase_ ( self : Union[str, Any] ): super().test_save_load_optional_components(expected_max_difference=3E-3 ) def lowerCAmelCase_ ( self : Dict ): SCREAMING_SNAKE_CASE_ = [ 'DDIMScheduler', 'DDPMScheduler', 'PNDMScheduler', 'HeunDiscreteScheduler', 'EulerAncestralDiscreteScheduler', 'KDPM2DiscreteScheduler', 'KDPM2AncestralDiscreteScheduler', 'DPMSolverSDEScheduler', ] SCREAMING_SNAKE_CASE_ = self.get_dummy_components() SCREAMING_SNAKE_CASE_ = self.pipeline_class(**_lowerCAmelCase ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=_lowerCAmelCase ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = self.get_dummy_inputs(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = 2 SCREAMING_SNAKE_CASE_ = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue SCREAMING_SNAKE_CASE_ = getattr(_lowerCAmelCase , scheduler_enum.name ) SCREAMING_SNAKE_CASE_ = scheduler_cls.from_config(pipe.scheduler.config ) SCREAMING_SNAKE_CASE_ = pipe(**_lowerCAmelCase )[0] outputs.append(_lowerCAmelCase ) assert check_same_shape(_lowerCAmelCase ) @require_torch_gpu @slow class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def lowerCAmelCase_ ( self : List[Any] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase_ ( self : Any ): SCREAMING_SNAKE_CASE_ = torch.manual_seed(33 ) SCREAMING_SNAKE_CASE_ = StableDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' , torch_dtype=torch.floataa ) pipe.to('cuda' ) SCREAMING_SNAKE_CASE_ = StableDiffusionLatentUpscalePipeline.from_pretrained( 'stabilityai/sd-x2-latent-upscaler' , torch_dtype=torch.floataa ) upscaler.to('cuda' ) SCREAMING_SNAKE_CASE_ = 'a photo of an astronaut high resolution, unreal engine, ultra realistic' SCREAMING_SNAKE_CASE_ = pipe(_lowerCAmelCase , generator=_lowerCAmelCase , output_type='latent' ).images SCREAMING_SNAKE_CASE_ = upscaler( prompt=_lowerCAmelCase , image=_lowerCAmelCase , num_inference_steps=20 , guidance_scale=0 , generator=_lowerCAmelCase , output_type='np' , ).images[0] SCREAMING_SNAKE_CASE_ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy' ) assert np.abs((expected_image - image).mean() ) < 5E-2 def lowerCAmelCase_ ( self : Tuple ): SCREAMING_SNAKE_CASE_ = torch.manual_seed(33 ) SCREAMING_SNAKE_CASE_ = StableDiffusionLatentUpscalePipeline.from_pretrained( 'stabilityai/sd-x2-latent-upscaler' , torch_dtype=torch.floataa ) upscaler.to('cuda' ) SCREAMING_SNAKE_CASE_ = 'the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas' SCREAMING_SNAKE_CASE_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png' ) SCREAMING_SNAKE_CASE_ = upscaler( prompt=_lowerCAmelCase , image=_lowerCAmelCase , num_inference_steps=20 , guidance_scale=0 , generator=_lowerCAmelCase , output_type='np' , ).images[0] SCREAMING_SNAKE_CASE_ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy' ) assert np.abs((expected_image - image).max() ) < 5E-2
210
import importlib import json import os import sys import tempfile import unittest from pathlib import Path import transformers import transformers.models.auto from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.bert.configuration_bert import BertConfig from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCamelCase__ : str = get_tests_dir('fixtures/dummy-config.json') class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def lowerCAmelCase_ ( self : Dict ): SCREAMING_SNAKE_CASE_ = 0 def lowerCAmelCase_ ( self : Optional[int] ): self.assertIsNotNone(transformers.models.auto.__spec__ ) self.assertIsNotNone(importlib.util.find_spec('transformers.models.auto' ) ) def lowerCAmelCase_ ( self : Any ): SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained('bert-base-uncased' ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained(_lowerCAmelCase ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained(_lowerCAmelCase ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase_ ( self : Dict ): SCREAMING_SNAKE_CASE_ = AutoConfig.for_model('roberta' ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase_ ( self : Dict ): with tempfile.TemporaryDirectory() as tmp_dir: # This model name contains bert and roberta, but roberta ends up being picked. SCREAMING_SNAKE_CASE_ = os.path.join(_lowerCAmelCase , 'fake-roberta' ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) with open(os.path.join(_lowerCAmelCase , 'config.json' ) , 'w' ) as f: f.write(json.dumps({} ) ) SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained(_lowerCAmelCase ) self.assertEqual(type(_lowerCAmelCase ) , _lowerCAmelCase ) def lowerCAmelCase_ ( self : Optional[Any] ): try: AutoConfig.register('custom' , _lowerCAmelCase ) # Wrong model type will raise an error with self.assertRaises(_lowerCAmelCase ): AutoConfig.register('model' , _lowerCAmelCase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_lowerCAmelCase ): AutoConfig.register('bert' , _lowerCAmelCase ) # Now that the config is registered, it can be used as any other config with the auto-API SCREAMING_SNAKE_CASE_ = CustomConfig() with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained(_lowerCAmelCase ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] def lowerCAmelCase_ ( self : Optional[int] ): with self.assertRaisesRegex( _lowerCAmelCase , 'bert-base is not a local folder and is not a valid model identifier' ): SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained('bert-base' ) def lowerCAmelCase_ ( self : int ): with self.assertRaisesRegex( _lowerCAmelCase , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained(_lowerCAmelCase , revision='aaaaaa' ) def lowerCAmelCase_ ( self : Tuple ): with self.assertRaisesRegex( _lowerCAmelCase , 'hf-internal-testing/no-config-test-repo does not appear to have a file named config.json.' , ): SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained('hf-internal-testing/no-config-test-repo' ) def lowerCAmelCase_ ( self : Union[str, Any] ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_lowerCAmelCase ): SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' ) # If remote code is disabled, we can't load this config. with self.assertRaises(_lowerCAmelCase ): SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' , trust_remote_code=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' , trust_remote_code=_lowerCAmelCase ) self.assertEqual(config.__class__.__name__ , 'NewModelConfig' ) # Test config can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained(_lowerCAmelCase , trust_remote_code=_lowerCAmelCase ) self.assertEqual(reloaded_config.__class__.__name__ , 'NewModelConfig' ) def lowerCAmelCase_ ( self : Any ): class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = "new-model" try: AutoConfig.register('new-model' , _lowerCAmelCase ) # If remote code is not set, the default is to use local SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' ) self.assertEqual(config.__class__.__name__ , 'NewModelConfigLocal' ) # If remote code is disabled, we load the local one. SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' , trust_remote_code=_lowerCAmelCase ) self.assertEqual(config.__class__.__name__ , 'NewModelConfigLocal' ) # If remote is enabled, we load from the Hub SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' , trust_remote_code=_lowerCAmelCase ) self.assertEqual(config.__class__.__name__ , 'NewModelConfig' ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"]
210
1
'''simple docstring''' from collections.abc import Generator def __magic_name__ ( ) -> Generator[int, None, None]: '''simple docstring''' snake_case_ = 0, 1 while True: snake_case_ = b, a + b yield b def __magic_name__ ( __UpperCAmelCase = 1000 ) -> int: '''simple docstring''' snake_case_ = 1 snake_case_ = fibonacci_generator() while len(str(next(_lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
56
'''simple docstring''' import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 _UpperCamelCase = data_utils.TransfoXLTokenizer _UpperCamelCase = data_utils.TransfoXLCorpus _UpperCamelCase = data_utils _UpperCamelCase = data_utils def a_ ( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) -> Optional[Any]: if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(_lowerCAmelCase ,'rb' ) as fp: __lowerCamelCase : Optional[Any] = pickle.load(_lowerCAmelCase ,encoding='latin1' ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) __lowerCamelCase : Tuple = pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['pretrained_vocab_file'] print(F'Save vocabulary to {pytorch_vocab_dump_path}' ) __lowerCamelCase : str = corpus.vocab.__dict__ torch.save(_lowerCAmelCase ,_lowerCAmelCase ) __lowerCamelCase : Optional[Any] = corpus.__dict__ corpus_dict_no_vocab.pop('vocab' ,_lowerCAmelCase ) __lowerCamelCase : Optional[Any] = pytorch_dump_folder_path + '/' + CORPUS_NAME print(F'Save dataset to {pytorch_dataset_dump_path}' ) torch.save(_lowerCAmelCase ,_lowerCAmelCase ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model __lowerCamelCase : int = os.path.abspath(_lowerCAmelCase ) __lowerCamelCase : Any = os.path.abspath(_lowerCAmelCase ) print(F'Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.' ) # Initialise PyTorch model if transfo_xl_config_file == "": __lowerCamelCase : Optional[int] = TransfoXLConfig() else: __lowerCamelCase : Optional[int] = TransfoXLConfig.from_json_file(_lowerCAmelCase ) print(F'Building PyTorch model from configuration: {config}' ) __lowerCamelCase : List[str] = TransfoXLLMHeadModel(_lowerCAmelCase ) __lowerCamelCase : Dict = load_tf_weights_in_transfo_xl(_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) # Save pytorch-model __lowerCamelCase : List[str] = os.path.join(_lowerCAmelCase ,_lowerCAmelCase ) __lowerCamelCase : int = os.path.join(_lowerCAmelCase ,_lowerCAmelCase ) print(F'Save PyTorch model to {os.path.abspath(_lowerCAmelCase )}' ) torch.save(model.state_dict() ,_lowerCAmelCase ) print(F'Save configuration file to {os.path.abspath(_lowerCAmelCase )}' ) with open(_lowerCAmelCase ,'w' ,encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) _UpperCamelCase = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
208
0
import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( __UpperCamelCase , unittest.TestCase ): A : Optional[int] = None A : str = BloomTokenizerFast A : Optional[Any] = BloomTokenizerFast A : str = True A : Optional[int] = False A : List[str] = "tokenizer_file" A : int = {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"} def snake_case__ ( self : str ): super().setUp() __snake_case : List[Any] = BloomTokenizerFast.from_pretrained("""bigscience/tokenizer""" ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self : int , **_lowerCAmelCase : List[Any] ): kwargs.update(self.special_tokens_map ) return BloomTokenizerFast.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def snake_case__ ( self : Optional[Any] ): __snake_case : List[str] = self.get_rust_tokenizer() __snake_case : int = ["""The quick brown fox</s>""", """jumps over the lazy dog</s>"""] __snake_case : Optional[int] = [[21_75, 2_37_14, 7_31_73, 14_42_52, 2], [77, 13_26_19, 34_78, 3_68, 10_95_86, 3_54_33, 2]] __snake_case : Optional[int] = tokenizer.batch_encode_plus(_lowerCAmelCase )["""input_ids"""] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) __snake_case : Optional[Any] = tokenizer.batch_decode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def snake_case__ ( self : Dict , _lowerCAmelCase : Dict=6 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __snake_case : List[str] = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input __snake_case : int = """This is a simple input""" __snake_case : List[str] = ["""This is a simple input 1""", """This is a simple input 2"""] __snake_case : Optional[int] = ("""This is a simple input""", """This is a pair""") __snake_case : Optional[Any] = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests try: tokenizer_r.encode(_lowerCAmelCase , max_length=_lowerCAmelCase ) tokenizer_r.encode_plus(_lowerCAmelCase , max_length=_lowerCAmelCase ) tokenizer_r.batch_encode_plus(_lowerCAmelCase , max_length=_lowerCAmelCase ) tokenizer_r.encode(_lowerCAmelCase , max_length=_lowerCAmelCase ) tokenizer_r.batch_encode_plus(_lowerCAmelCase , max_length=_lowerCAmelCase ) except ValueError: self.fail("""Bloom Tokenizer should be able to deal with padding""" ) __snake_case : List[Any] = None # Hotfixing padding = None self.assertRaises(_lowerCAmelCase , tokenizer_r.encode , _lowerCAmelCase , max_length=_lowerCAmelCase , padding="""max_length""" ) # Simple input self.assertRaises(_lowerCAmelCase , tokenizer_r.encode_plus , _lowerCAmelCase , max_length=_lowerCAmelCase , padding="""max_length""" ) # Simple input self.assertRaises( _lowerCAmelCase , tokenizer_r.batch_encode_plus , _lowerCAmelCase , max_length=_lowerCAmelCase , padding="""max_length""" , ) # Pair input self.assertRaises(_lowerCAmelCase , tokenizer_r.encode , _lowerCAmelCase , max_length=_lowerCAmelCase , padding="""max_length""" ) # Pair input self.assertRaises(_lowerCAmelCase , tokenizer_r.encode_plus , _lowerCAmelCase , max_length=_lowerCAmelCase , padding="""max_length""" ) # Pair input self.assertRaises( _lowerCAmelCase , tokenizer_r.batch_encode_plus , _lowerCAmelCase , max_length=_lowerCAmelCase , padding="""max_length""" , ) def snake_case__ ( self : List[str] ): __snake_case : str = self.get_rust_tokenizer() __snake_case : str = load_dataset("""xnli""" , """all_languages""" , split="""test""" , streaming=_lowerCAmelCase ) __snake_case : List[str] = next(iter(_lowerCAmelCase ) )["""premise"""] # pick up one data __snake_case : Tuple = list(sample_data.values() ) __snake_case : Any = list(map(tokenizer.encode , _lowerCAmelCase ) ) __snake_case : Optional[int] = [tokenizer.decode(_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) for x in output_tokens] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def snake_case__ ( self : Dict ): # The test has to be overriden because BLOOM uses ALiBi positional embeddings that does not have # any sequence length constraints. This test of the parent class will fail since it relies on the # maximum sequence length of the positoonal embeddings. self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map ) , 1 ) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values() )[0] ) , 1 )
353
import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType lowercase_ = None lowercase_ = "<" if sys.byteorder == "little" else ">" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image lowercase_ = [ np.dtype("|b1"), np.dtype("|u1"), np.dtype("<u2"), np.dtype(">u2"), np.dtype("<i2"), np.dtype(">i2"), np.dtype("<u4"), np.dtype(">u4"), np.dtype("<i4"), np.dtype(">i4"), np.dtype("<f4"), np.dtype(">f4"), np.dtype("<f8"), np.dtype(">f8"), ] @dataclass class SCREAMING_SNAKE_CASE__ : A : bool = True A : Optional[str] = None # Automatically constructed A : ClassVar[str] = "PIL.Image.Image" A : ClassVar[Any] = pa.struct({"bytes": pa.binary(), "path": pa.string()} ) A : str = field(default="Image" , init=__UpperCamelCase , repr=__UpperCamelCase ) def __call__( self : Any ): return self.pa_type def snake_case__ ( self : List[Any] , _lowerCAmelCase : Union[str, bytes, dict, np.ndarray, "PIL.Image.Image"] ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ): __snake_case : str = np.array(_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ): return {"path": value, "bytes": None} elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): return {"path": None, "bytes": value} elif isinstance(_lowerCAmelCase , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(_lowerCAmelCase ) elif isinstance(_lowerCAmelCase , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(_lowerCAmelCase ) elif value.get("""path""" ) is not None and os.path.isfile(value["""path"""] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get("""path""" )} elif value.get("""bytes""" ) is not None or value.get("""path""" ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get("""bytes""" ), "path": value.get("""path""" )} else: raise ValueError( f'''An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def snake_case__ ( self : List[str] , _lowerCAmelCase : dict , _lowerCAmelCase : Dict=None ): if not self.decode: raise RuntimeError("""Decoding is disabled for this feature. Please use Image(decode=True) instead.""" ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support decoding images, please install 'Pillow'.""" ) if token_per_repo_id is None: __snake_case : Tuple = {} __snake_case , __snake_case : str = value["""path"""], value["""bytes"""] if bytes_ is None: if path is None: raise ValueError(f'''An image should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) else: if is_local_path(_lowerCAmelCase ): __snake_case : str = PIL.Image.open(_lowerCAmelCase ) else: __snake_case : List[str] = path.split("""::""" )[-1] try: __snake_case : Dict = string_to_dict(_lowerCAmelCase , config.HUB_DATASETS_URL )["""repo_id"""] __snake_case : int = token_per_repo_id.get(_lowerCAmelCase ) except ValueError: __snake_case : List[Any] = None with xopen(_lowerCAmelCase , """rb""" , use_auth_token=_lowerCAmelCase ) as f: __snake_case : Union[str, Any] = BytesIO(f.read() ) __snake_case : Dict = PIL.Image.open(bytes_ ) else: __snake_case : Optional[Any] = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def snake_case__ ( self : Union[str, Any] ): from .features import Value return ( self if self.decode else { "bytes": Value("""binary""" ), "path": Value("""string""" ), } ) def snake_case__ ( self : Optional[int] , _lowerCAmelCase : Union[pa.StringArray, pa.StructArray, pa.ListArray] ): if pa.types.is_string(storage.type ): __snake_case : Optional[Any] = pa.array([None] * len(_lowerCAmelCase ) , type=pa.binary() ) __snake_case : Any = pa.StructArray.from_arrays([bytes_array, storage] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): __snake_case : Optional[Any] = pa.array([None] * len(_lowerCAmelCase ) , type=pa.string() ) __snake_case : List[str] = pa.StructArray.from_arrays([storage, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("""bytes""" ) >= 0: __snake_case : List[str] = storage.field("""bytes""" ) else: __snake_case : List[Any] = pa.array([None] * len(_lowerCAmelCase ) , type=pa.binary() ) if storage.type.get_field_index("""path""" ) >= 0: __snake_case : Optional[int] = storage.field("""path""" ) else: __snake_case : int = pa.array([None] * len(_lowerCAmelCase ) , type=pa.string() ) __snake_case : Tuple = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): __snake_case : Optional[Any] = pa.array( [encode_np_array(np.array(_lowerCAmelCase ) )["""bytes"""] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) __snake_case : Optional[int] = pa.array([None] * len(_lowerCAmelCase ) , type=pa.string() ) __snake_case : List[str] = pa.StructArray.from_arrays( [bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(_lowerCAmelCase , self.pa_type ) def snake_case__ ( self : Union[str, Any] , _lowerCAmelCase : pa.StructArray ): @no_op_if_value_is_null def path_to_bytes(_lowerCAmelCase : Tuple ): with xopen(_lowerCAmelCase , """rb""" ) as f: __snake_case : Optional[int] = f.read() return bytes_ __snake_case : Tuple = pa.array( [ (path_to_bytes(x["""path"""] ) if x["""bytes"""] is None else x["""bytes"""]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) __snake_case : Optional[Any] = pa.array( [os.path.basename(_lowerCAmelCase ) if path is not None else None for path in storage.field("""path""" ).to_pylist()] , type=pa.string() , ) __snake_case : Any = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(_lowerCAmelCase , self.pa_type ) def __lowerCAmelCase ( ): '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() __snake_case : Optional[Any] = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def __lowerCAmelCase ( __SCREAMING_SNAKE_CASE : "PIL.Image.Image" ): '''simple docstring''' __snake_case : List[Any] = BytesIO() if image.format in list_image_compression_formats(): __snake_case : Union[str, Any] = image.format else: __snake_case : List[Any] = """PNG""" if image.mode in ["""1""", """L""", """LA""", """RGB""", """RGBA"""] else """TIFF""" image.save(__SCREAMING_SNAKE_CASE , format=__SCREAMING_SNAKE_CASE ) return buffer.getvalue() def __lowerCAmelCase ( __SCREAMING_SNAKE_CASE : "PIL.Image.Image" ): '''simple docstring''' if hasattr(__SCREAMING_SNAKE_CASE , """filename""" ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(__SCREAMING_SNAKE_CASE )} def __lowerCAmelCase ( __SCREAMING_SNAKE_CASE : np.ndarray ): '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) __snake_case : List[Any] = array.dtype __snake_case : List[Any] = dtype.byteorder if dtype.byteorder != """=""" else _NATIVE_BYTEORDER __snake_case : Dict = dtype.kind __snake_case : Union[str, Any] = dtype.itemsize __snake_case : Tuple = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: __snake_case : int = np.dtype("""|u1""" ) if dtype_kind not in ["u", "i"]: raise TypeError( F'''Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.''' ) if dtype is not dest_dtype: warnings.warn(F'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: __snake_case : List[str] = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: __snake_case : int = dtype_byteorder + dtype_kind + str(__SCREAMING_SNAKE_CASE ) __snake_case : Any = np.dtype(__SCREAMING_SNAKE_CASE ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(F'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( F'''Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}''' ) __snake_case : Optional[int] = PIL.Image.fromarray(array.astype(__SCREAMING_SNAKE_CASE ) ) return {"path": None, "bytes": image_to_bytes(__SCREAMING_SNAKE_CASE )} def __lowerCAmelCase ( __SCREAMING_SNAKE_CASE : Union[List[str], List[dict], List[np.ndarray], List["PIL.Image.Image"]] ): '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) if objs: __snake_case , __snake_case : Any = first_non_null_value(__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(__SCREAMING_SNAKE_CASE , np.ndarray ): __snake_case : int = no_op_if_value_is_null(__SCREAMING_SNAKE_CASE ) return [obj_to_image_dict_func(__SCREAMING_SNAKE_CASE ) for obj in objs] elif isinstance(__SCREAMING_SNAKE_CASE , PIL.Image.Image ): __snake_case : List[str] = no_op_if_value_is_null(__SCREAMING_SNAKE_CASE ) return [obj_to_image_dict_func(__SCREAMING_SNAKE_CASE ) for obj in objs] else: return objs else: return objs
20
0
"""simple docstring""" def _A ( lowercase , lowercase ): """simple docstring""" return number | (1 << position) def _A ( lowercase , lowercase ): """simple docstring""" return number & ~(1 << position) def _A ( lowercase , lowercase ): """simple docstring""" return number ^ (1 << position) def _A ( lowercase , lowercase ): """simple docstring""" return ((number >> position) & 1) == 1 def _A ( lowercase , lowercase ): """simple docstring""" return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
81
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase__ = { "configuration_groupvit": [ "GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "GroupViTConfig", "GroupViTOnnxConfig", "GroupViTTextConfig", "GroupViTVisionConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ "GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "GroupViTModel", "GroupViTPreTrainedModel", "GroupViTTextModel", "GroupViTVisionModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ "TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFGroupViTModel", "TFGroupViTPreTrainedModel", "TFGroupViTTextModel", "TFGroupViTVisionModel", ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
0
0
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def snake_case_ ( lowerCAmelCase_ )-> Optional[Any]: '''simple docstring''' return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase ( _lowerCamelCase ): """simple docstring""" @staticmethod def _snake_case ( a_ ) -> Union[str, Any]: _UpperCAmelCase : Dict = parser.add_parser("""download""" ) download_parser.add_argument( """--cache-dir""" ,type=a_ ,default=a_ ,help="""Path to location to store the models""" ) download_parser.add_argument( """--force""" ,action="""store_true""" ,help="""Force the model to be download even if already in cache-dir""" ) download_parser.add_argument( """--trust-remote-code""" ,action="""store_true""" ,help="""Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you've reviewed the code as it will execute on your local machine""" ,) download_parser.add_argument("""model""" ,type=a_ ,help="""Name of the model to download""" ) download_parser.set_defaults(func=a_ ) def __init__( self ,a_ ,a_ ,a_ ,a_ ) -> List[Any]: _UpperCAmelCase : Optional[Any] = model _UpperCAmelCase : Tuple = cache _UpperCAmelCase : Dict = force _UpperCAmelCase : Dict = trust_remote_code def _snake_case ( self ) -> Union[str, Any]: from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
349
'''simple docstring''' import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class lowercase ( unittest.TestCase ): """simple docstring""" def _snake_case ( self ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _snake_case ( self ) -> Optional[int]: _UpperCAmelCase : List[str] = 1 _UpperCAmelCase : List[str] = 3 _UpperCAmelCase : Union[str, Any] = (32, 32) _UpperCAmelCase : str = floats_tensor((batch_size, num_channels) + sizes ,rng=random.Random(0 ) ).to(a_ ) return image @property def _snake_case ( self ) -> List[Any]: torch.manual_seed(0 ) _UpperCAmelCase : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") ,up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") ,cross_attention_dim=32 ,) return model @property def _snake_case ( self ) -> Optional[int]: torch.manual_seed(0 ) _UpperCAmelCase : str = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] ,up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] ,latent_channels=4 ,) return model @property def _snake_case ( self ) -> Dict: torch.manual_seed(0 ) _UpperCAmelCase : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_000 ,) return CLIPTextModel(a_ ) @property def _snake_case ( self ) -> Union[str, Any]: def extract(*a_ ,**a_ ): class lowercase : """simple docstring""" def __init__( self ) -> Any: _UpperCAmelCase : str = torch.ones([0] ) def _snake_case ( self ,a_ ) -> Any: self.pixel_values.to(a_ ) return self return Out() return extract def _snake_case ( self ) -> List[str]: _UpperCAmelCase : List[str] = """cpu""" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase : Union[str, Any] = self.dummy_cond_unet _UpperCAmelCase : int = DDIMScheduler( beta_start=0.0_0085 ,beta_end=0.012 ,beta_schedule="""scaled_linear""" ,clip_sample=a_ ,set_alpha_to_one=a_ ,) _UpperCAmelCase : Optional[int] = self.dummy_vae _UpperCAmelCase : Optional[int] = self.dummy_text_encoder _UpperCAmelCase : str = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk _UpperCAmelCase : int = StableDiffusionPipeline( unet=a_ ,scheduler=a_ ,vae=a_ ,text_encoder=a_ ,tokenizer=a_ ,safety_checker=a_ ,feature_extractor=self.dummy_extractor ,) _UpperCAmelCase : Optional[Any] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) _UpperCAmelCase : Union[str, Any] = """A painting of a squirrel eating a burger""" _UpperCAmelCase : Optional[int] = torch.Generator(device=a_ ).manual_seed(0 ) _UpperCAmelCase : str = sd_pipe([prompt] ,generator=a_ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type="""np""" ) _UpperCAmelCase : int = output.images _UpperCAmelCase : Union[str, Any] = torch.Generator(device=a_ ).manual_seed(0 ) _UpperCAmelCase : str = sd_pipe( [prompt] ,generator=a_ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type="""np""" ,return_dict=a_ ,)[0] _UpperCAmelCase : str = image[0, -3:, -3:, -1] _UpperCAmelCase : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCAmelCase : Optional[int] = np.array([0.5756, 0.6118, 0.5005, 0.5041, 0.5471, 0.4726, 0.4976, 0.4865, 0.4864] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def _snake_case ( self ) -> Any: _UpperCAmelCase : Any = """cpu""" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase : Tuple = self.dummy_cond_unet _UpperCAmelCase : Optional[int] = PNDMScheduler(skip_prk_steps=a_ ) _UpperCAmelCase : int = self.dummy_vae _UpperCAmelCase : int = self.dummy_text_encoder _UpperCAmelCase : Union[str, Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk _UpperCAmelCase : str = StableDiffusionPipeline( unet=a_ ,scheduler=a_ ,vae=a_ ,text_encoder=a_ ,tokenizer=a_ ,safety_checker=a_ ,feature_extractor=self.dummy_extractor ,) _UpperCAmelCase : str = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) _UpperCAmelCase : int = """A painting of a squirrel eating a burger""" _UpperCAmelCase : Any = torch.Generator(device=a_ ).manual_seed(0 ) _UpperCAmelCase : List[Any] = sd_pipe([prompt] ,generator=a_ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type="""np""" ) _UpperCAmelCase : Dict = output.images _UpperCAmelCase : List[Any] = torch.Generator(device=a_ ).manual_seed(0 ) _UpperCAmelCase : Any = sd_pipe( [prompt] ,generator=a_ ,guidance_scale=6.0 ,num_inference_steps=2 ,output_type="""np""" ,return_dict=a_ ,)[0] _UpperCAmelCase : Optional[int] = image[0, -3:, -3:, -1] _UpperCAmelCase : int = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCAmelCase : Union[str, Any] = np.array([0.5125, 0.5716, 0.4828, 0.5060, 0.5650, 0.4768, 0.5185, 0.4895, 0.4993] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def _snake_case ( self ) -> Optional[int]: _UpperCAmelCase : Optional[int] = StableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-lms-pipe""" ,safety_checker=a_ ) assert isinstance(a_ ,a_ ) assert isinstance(pipe.scheduler ,a_ ) assert pipe.safety_checker is None _UpperCAmelCase : Dict = pipe("""example prompt""" ,num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(a_ ) _UpperCAmelCase : Any = StableDiffusionPipeline.from_pretrained(a_ ) # sanity check that the pipeline still works assert pipe.safety_checker is None _UpperCAmelCase : Union[str, Any] = pipe("""example prompt""" ,num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != """cuda""" ,"""This test requires a GPU""" ) def _snake_case ( self ) -> str: _UpperCAmelCase : Optional[int] = self.dummy_cond_unet _UpperCAmelCase : str = PNDMScheduler(skip_prk_steps=a_ ) _UpperCAmelCase : List[str] = self.dummy_vae _UpperCAmelCase : int = self.dummy_text_encoder _UpperCAmelCase : str = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # put models in fp16 _UpperCAmelCase : str = unet.half() _UpperCAmelCase : List[str] = vae.half() _UpperCAmelCase : Dict = bert.half() # make sure here that pndm scheduler skips prk _UpperCAmelCase : Dict = StableDiffusionPipeline( unet=a_ ,scheduler=a_ ,vae=a_ ,text_encoder=a_ ,tokenizer=a_ ,safety_checker=a_ ,feature_extractor=self.dummy_extractor ,) _UpperCAmelCase : List[str] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) _UpperCAmelCase : str = """A painting of a squirrel eating a burger""" _UpperCAmelCase : int = sd_pipe([prompt] ,num_inference_steps=2 ,output_type="""np""" ).images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class lowercase ( unittest.TestCase ): """simple docstring""" def _snake_case ( self ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case ( self ) -> str: _UpperCAmelCase : List[str] = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ,safety_checker=a_ ) _UpperCAmelCase : Dict = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) _UpperCAmelCase : int = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) _UpperCAmelCase : List[Any] = ( """portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle""" """ coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with""" """ anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and""" """ children from bahnhof zoo, detailed """ ) _UpperCAmelCase : Any = 4_003_660_346 _UpperCAmelCase : List[Any] = 7 # without safety guidance (sld_guidance_scale = 0) _UpperCAmelCase : int = torch.manual_seed(a_ ) _UpperCAmelCase : str = sd_pipe( [prompt] ,generator=a_ ,guidance_scale=a_ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=0 ,) _UpperCAmelCase : str = output.images _UpperCAmelCase : Optional[int] = image[0, -3:, -3:, -1] _UpperCAmelCase : List[str] = [0.2278, 0.2231, 0.2249, 0.2333, 0.2303, 0.1885, 0.2273, 0.2144, 0.2176] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 # without safety guidance (strong configuration) _UpperCAmelCase : List[str] = torch.manual_seed(a_ ) _UpperCAmelCase : Optional[Any] = sd_pipe( [prompt] ,generator=a_ ,guidance_scale=a_ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=2_000 ,sld_warmup_steps=7 ,sld_threshold=0.025 ,sld_momentum_scale=0.5 ,sld_mom_beta=0.7 ,) _UpperCAmelCase : List[str] = output.images _UpperCAmelCase : List[str] = image[0, -3:, -3:, -1] _UpperCAmelCase : List[str] = [0.2383, 0.2276, 0.236, 0.2192, 0.2186, 0.2053, 0.1971, 0.1901, 0.1719] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _snake_case ( self ) -> int: _UpperCAmelCase : Any = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ,safety_checker=a_ ) _UpperCAmelCase : Union[str, Any] = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) _UpperCAmelCase : Union[str, Any] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) _UpperCAmelCase : Any = """padme amidala taking a bath artwork, safe for work, no nudity""" _UpperCAmelCase : Optional[Any] = 2_734_971_755 _UpperCAmelCase : Optional[int] = 7 _UpperCAmelCase : int = torch.manual_seed(a_ ) _UpperCAmelCase : int = sd_pipe( [prompt] ,generator=a_ ,guidance_scale=a_ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=0 ,) _UpperCAmelCase : Optional[int] = output.images _UpperCAmelCase : List[Any] = image[0, -3:, -3:, -1] _UpperCAmelCase : Optional[int] = [0.3502, 0.3622, 0.3396, 0.3642, 0.3478, 0.3318, 0.35, 0.3348, 0.3297] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 _UpperCAmelCase : Optional[int] = torch.manual_seed(a_ ) _UpperCAmelCase : int = sd_pipe( [prompt] ,generator=a_ ,guidance_scale=a_ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=2_000 ,sld_warmup_steps=7 ,sld_threshold=0.025 ,sld_momentum_scale=0.5 ,sld_mom_beta=0.7 ,) _UpperCAmelCase : Union[str, Any] = output.images _UpperCAmelCase : Any = image[0, -3:, -3:, -1] _UpperCAmelCase : List[Any] = [0.5531, 0.5206, 0.4895, 0.5156, 0.5182, 0.4751, 0.4802, 0.4803, 0.4443] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _snake_case ( self ) -> Any: _UpperCAmelCase : Any = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ) _UpperCAmelCase : List[str] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) _UpperCAmelCase : Optional[int] = ( """the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c.""" """ leyendecker""" ) _UpperCAmelCase : Dict = 1_044_355_234 _UpperCAmelCase : int = 12 _UpperCAmelCase : Optional[Any] = torch.manual_seed(a_ ) _UpperCAmelCase : List[str] = sd_pipe( [prompt] ,generator=a_ ,guidance_scale=a_ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=0 ,) _UpperCAmelCase : List[str] = output.images _UpperCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1] _UpperCAmelCase : Dict = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-7 _UpperCAmelCase : Tuple = torch.manual_seed(a_ ) _UpperCAmelCase : Dict = sd_pipe( [prompt] ,generator=a_ ,guidance_scale=a_ ,num_inference_steps=50 ,output_type="""np""" ,width=512 ,height=512 ,sld_guidance_scale=2_000 ,sld_warmup_steps=7 ,sld_threshold=0.025 ,sld_momentum_scale=0.5 ,sld_mom_beta=0.7 ,) _UpperCAmelCase : Optional[Any] = output.images _UpperCAmelCase : Dict = image[0, -3:, -3:, -1] _UpperCAmelCase : int = np.array([0.5818, 0.6285, 0.6835, 0.6019, 0.625, 0.6754, 0.6096, 0.6334, 0.6561] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
349
1
"""simple docstring""" import os import sys import unittest _a = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path _a = os.path.join(git_repo_path, 'src', 'transformers') _a = '\n{0} = None\n' _a = '\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n' _a = '\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n' class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : Optional[int] ): __lowercase = find_backend(" _import_structure[\"models.albert\"].append(\"AlbertTokenizerFast\")" ) self.assertIsNone(UpperCAmelCase__ ) __lowercase = find_backend(" if not is_tokenizers_available():" ) self.assertEqual(UpperCAmelCase__, "tokenizers" ) __lowercase = find_backend(" if not is_tensorflow_text_available():" ) self.assertEqual(UpperCAmelCase__, "tensorflow_text" ) __lowercase = find_backend(" if not (is_sentencepiece_available() and is_tokenizers_available()):" ) self.assertEqual(UpperCAmelCase__, "sentencepiece_and_tokenizers" ) __lowercase = find_backend( " if not (is_sentencepiece_available() and is_tensorflow_text_available()):" ) self.assertEqual(UpperCAmelCase__, "sentencepiece_and_tensorflow_text" ) __lowercase = find_backend( " if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):" ) self.assertEqual(UpperCAmelCase__, "sentencepiece_and_tokenizers_and_vision" ) def _lowercase ( self : Tuple ): __lowercase = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn("torch", UpperCAmelCase__ ) self.assertIn("tensorflow_text", UpperCAmelCase__ ) self.assertIn("sentencepiece_and_tokenizers", UpperCAmelCase__ ) # Likewise, we can't assert on the exact content of a key self.assertIn("BertModel", objects["torch"] ) self.assertIn("TFBertModel", objects["tf"] ) self.assertIn("FlaxBertModel", objects["flax"] ) self.assertIn("BertModel", objects["torch"] ) self.assertIn("TFBertTokenizer", objects["tensorflow_text"] ) self.assertIn("convert_slow_tokenizer", objects["sentencepiece_and_tokenizers"] ) def _lowercase ( self : Optional[Any] ): __lowercase = create_dummy_object("CONSTANT", "'torch'" ) self.assertEqual(UpperCAmelCase__, "\nCONSTANT = None\n" ) __lowercase = create_dummy_object("function", "'torch'" ) self.assertEqual( UpperCAmelCase__, "\ndef function(*args, **kwargs):\n requires_backends(function, 'torch')\n" ) __lowercase = "\nclass FakeClass(metaclass=DummyObject):\n _backends = 'torch'\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, 'torch')\n" __lowercase = create_dummy_object("FakeClass", "'torch'" ) self.assertEqual(UpperCAmelCase__, UpperCAmelCase__ ) def _lowercase ( self : List[Any] ): __lowercase = "# This file is autogenerated by the command `make fix-copies`, do not edit.\nfrom ..utils import DummyObject, requires_backends\n\n\nCONSTANT = None\n\n\ndef function(*args, **kwargs):\n requires_backends(function, [\"torch\"])\n\n\nclass FakeClass(metaclass=DummyObject):\n _backends = [\"torch\"]\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, [\"torch\"])\n" __lowercase = create_dummy_files({"torch": ["CONSTANT", "function", "FakeClass"]} ) self.assertEqual(dummy_files["torch"], UpperCAmelCase__ )
17
"""simple docstring""" import inspect import unittest import numpy as np from transformers import BeitConfig from transformers.testing_utils import require_flax, require_vision, slow from transformers.utils import cached_property, is_flax_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor if is_flax_available(): import jax from transformers import FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : str, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : int=1_0_0, UpperCAmelCase__ : Any=1_3, UpperCAmelCase__ : List[Any]=3_0, UpperCAmelCase__ : Dict=2, UpperCAmelCase__ : Any=3, UpperCAmelCase__ : Optional[Any]=True, UpperCAmelCase__ : List[str]=True, UpperCAmelCase__ : Optional[Any]=3_2, UpperCAmelCase__ : Any=5, UpperCAmelCase__ : Any=4, UpperCAmelCase__ : Any=3_7, UpperCAmelCase__ : Optional[int]="gelu", UpperCAmelCase__ : Dict=0.1, UpperCAmelCase__ : Optional[int]=0.1, UpperCAmelCase__ : Dict=1_0, UpperCAmelCase__ : Tuple=0.02, UpperCAmelCase__ : List[Any]=3, ): __lowercase = parent __lowercase = vocab_size __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __lowercase = (image_size // patch_size) ** 2 __lowercase = num_patches + 1 def _lowercase ( self : int ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size], self.type_sequence_label_size ) __lowercase = BeitConfig( vocab_size=self.vocab_size, image_size=self.image_size, patch_size=self.patch_size, num_channels=self.num_channels, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, is_decoder=UpperCAmelCase__, initializer_range=self.initializer_range, ) return config, pixel_values, labels def _lowercase ( self : Any, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[str], UpperCAmelCase__ : List[str] ): __lowercase = FlaxBeitModel(config=UpperCAmelCase__ ) __lowercase = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : int, UpperCAmelCase__ : List[Any] ): __lowercase = FlaxBeitForMaskedImageModeling(config=UpperCAmelCase__ ) __lowercase = model(UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length - 1, self.vocab_size) ) def _lowercase ( self : Dict, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Any, UpperCAmelCase__ : Union[str, Any] ): __lowercase = self.type_sequence_label_size __lowercase = FlaxBeitForImageClassification(config=UpperCAmelCase__ ) __lowercase = model(UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowercase = 1 __lowercase = FlaxBeitForImageClassification(UpperCAmelCase__ ) __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(UpperCAmelCase__ ) def _lowercase ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) ,( __lowercase ) ,( __lowercase ) , ) = config_and_inputs __lowercase = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class _lowerCAmelCase ( lowercase ,unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (FlaxBeitModel, FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling) if is_flax_available() else () ) def _lowercase ( self : List[Any] ): __lowercase = FlaxBeitModelTester(self ) __lowercase = ConfigTester(self, config_class=UpperCAmelCase__, has_text_modality=UpperCAmelCase__, hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ): self.config_tester.run_common_tests() def _lowercase ( self : Optional[int] ): __lowercase ,__lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(UpperCAmelCase__ ) __lowercase = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["pixel_values"] self.assertListEqual(arg_names[:1], UpperCAmelCase__ ) def _lowercase ( self : Tuple ): __lowercase ,__lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): __lowercase = self._prepare_for_class(UpperCAmelCase__, UpperCAmelCase__ ) __lowercase = model_class(UpperCAmelCase__ ) @jax.jit def model_jitted(UpperCAmelCase__ : str, **UpperCAmelCase__ : Dict ): return model(pixel_values=UpperCAmelCase__, **UpperCAmelCase__ ) with self.subTest("JIT Enabled" ): __lowercase = model_jitted(**UpperCAmelCase__ ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): __lowercase = model_jitted(**UpperCAmelCase__ ).to_tuple() self.assertEqual(len(UpperCAmelCase__ ), len(UpperCAmelCase__ ) ) for jitted_output, output in zip(UpperCAmelCase__, UpperCAmelCase__ ): self.assertEqual(jitted_output.shape, output.shape ) def _lowercase ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def _lowercase ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase__ ) def _lowercase ( self : Tuple ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase__ ) @slow def _lowercase ( self : Union[str, Any] ): for model_class_name in self.all_model_classes: __lowercase = model_class_name.from_pretrained("microsoft/beit-base-patch16-224" ) __lowercase = model(np.ones((1, 3, 2_2_4, 2_2_4) ) ) self.assertIsNotNone(UpperCAmelCase__ ) def _A ( ) -> str: '''simple docstring''' __lowercase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") return image @require_vision @require_flax class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def _lowercase ( self : Optional[int] ): return BeitImageProcessor.from_pretrained("microsoft/beit-base-patch16-224" ) if is_vision_available() else None @slow def _lowercase ( self : Any ): __lowercase = FlaxBeitForMaskedImageModeling.from_pretrained("microsoft/beit-base-patch16-224-pt22k" ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=UpperCAmelCase__, return_tensors="np" ).pixel_values # prepare bool_masked_pos __lowercase = np.ones((1, 1_9_6), dtype=UpperCAmelCase__ ) # forward pass __lowercase = model(pixel_values=UpperCAmelCase__, bool_masked_pos=UpperCAmelCase__ ) __lowercase = outputs.logits # verify the logits __lowercase = (1, 1_9_6, 8_1_9_2) self.assertEqual(logits.shape, UpperCAmelCase__ ) __lowercase = np.array( [[-3.2_437, 0.5_072, -13.9_174], [-3.2_456, 0.4_948, -13.9_401], [-3.2_033, 0.5_121, -13.8_550]] ) self.assertTrue(np.allclose(logits[bool_masked_pos][:3, :3], UpperCAmelCase__, atol=1E-2 ) ) @slow def _lowercase ( self : Any ): __lowercase = FlaxBeitForImageClassification.from_pretrained("microsoft/beit-base-patch16-224" ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=UpperCAmelCase__, return_tensors="np" ) # forward pass __lowercase = model(**UpperCAmelCase__ ) __lowercase = outputs.logits # verify the logits __lowercase = (1, 1_0_0_0) self.assertEqual(logits.shape, UpperCAmelCase__ ) __lowercase = np.array([-1.2_385, -1.0_987, -1.0_108] ) self.assertTrue(np.allclose(logits[0, :3], UpperCAmelCase__, atol=1E-4 ) ) __lowercase = 2_8_1 self.assertEqual(logits.argmax(-1 ).item(), UpperCAmelCase__ ) @slow def _lowercase ( self : List[str] ): __lowercase = FlaxBeitForImageClassification.from_pretrained("microsoft/beit-large-patch16-224-pt22k-ft22k" ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=UpperCAmelCase__, return_tensors="np" ) # forward pass __lowercase = model(**UpperCAmelCase__ ) __lowercase = outputs.logits # verify the logits __lowercase = (1, 2_1_8_4_1) self.assertEqual(logits.shape, UpperCAmelCase__ ) __lowercase = np.array([1.6_881, -0.2_787, 0.5_901] ) self.assertTrue(np.allclose(logits[0, :3], UpperCAmelCase__, atol=1E-4 ) ) __lowercase = 2_3_9_6 self.assertEqual(logits.argmax(-1 ).item(), UpperCAmelCase__ )
17
1
from __future__ import annotations __lowerCamelCase : Tuple = [True] * 100_0001 __lowerCamelCase : Tuple = 2 while i * i <= 100_0000: if seive[i]: for j in range(i * i, 100_0001, i): __lowerCamelCase : List[str] = False i += 1 def SCREAMING_SNAKE_CASE ( snake_case_ : int ): return seive[n] def SCREAMING_SNAKE_CASE ( snake_case_ : int ): return any(digit in "02468" for digit in str(snake_case_ ) ) def SCREAMING_SNAKE_CASE ( snake_case_ : int = 1000000 ): snake_case__ : int = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(snake_case_ ) and not contains_an_even_digit(snake_case_ ): snake_case__ : List[str] = str(snake_case_ ) snake_case__ : int = [int(str_num[j:] + str_num[:j] ) for j in range(len(snake_case_ ) )] if all(is_prime(snake_case_ ) for i in list_nums ): result.append(snake_case_ ) return result def SCREAMING_SNAKE_CASE ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f"{len(find_circular_primes()) = }")
286
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class SCREAMING_SNAKE_CASE__ ( UpperCamelCase_ ): """simple docstring""" a_ = ["image_processor", "tokenizer"] a_ = "ViltImageProcessor" a_ = ("BertTokenizer", "BertTokenizerFast") def __init__( self : Optional[int] , __A : Optional[int]=None , __A : Optional[Any]=None , **__A : int ): snake_case__ : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , __A , ) snake_case__ : Tuple = kwargs.pop("feature_extractor" ) snake_case__ : Any = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(__A , __A ) snake_case__ : Tuple = self.image_processor def __call__( self : List[Any] , __A : int , __A : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , __A : bool = True , __A : Union[bool, str, PaddingStrategy] = False , __A : Union[bool, str, TruncationStrategy] = None , __A : Optional[int] = None , __A : int = 0 , __A : Optional[int] = None , __A : Optional[bool] = None , __A : Optional[bool] = None , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = False , __A : bool = True , __A : Optional[Union[str, TensorType]] = None , **__A : List[Any] , ): snake_case__ : Optional[int] = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_token_type_ids=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) # add pixel_values + pixel_mask snake_case__ : Optional[Any] = self.image_processor(__A , return_tensors=__A ) encoding.update(__A ) return encoding def _lowercase ( self : Optional[Any] , *__A : List[str] , **__A : Optional[int] ): return self.tokenizer.batch_decode(*__A , **__A ) def _lowercase ( self : Dict , *__A : str , **__A : str ): return self.tokenizer.decode(*__A , **__A ) @property def _lowercase ( self : str ): snake_case__ : Optional[Any] = self.tokenizer.model_input_names snake_case__ : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _lowercase ( self : List[Any] ): warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , __A , ) return self.image_processor_class @property def _lowercase ( self : str ): warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , __A , ) return self.image_processor
286
1
'''simple docstring''' import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask a_ : Optional[Any] = logging.getLogger(__name__) class snake_case ( lowercase ): """simple docstring""" def __init__( self , UpperCamelCase=-1 ): """simple docstring""" # in NER datasets, the last column is usually reserved for NER label lowerCamelCase_ = label_idx def snake_case ( self , UpperCamelCase , UpperCamelCase ): """simple docstring""" if isinstance(UpperCamelCase , UpperCamelCase ): lowerCamelCase_ = mode.value lowerCamelCase_ = os.path.join(UpperCamelCase , f'''{mode}.txt''' ) lowerCamelCase_ = 1 lowerCamelCase_ = [] with open(UpperCamelCase , encoding="utf-8" ) as f: lowerCamelCase_ = [] lowerCamelCase_ = [] for line in f: if line.startswith("-DOCSTART-" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=f'''{mode}-{guid_index}''' , words=UpperCamelCase , labels=UpperCamelCase ) ) guid_index += 1 lowerCamelCase_ = [] lowerCamelCase_ = [] else: lowerCamelCase_ = line.split(" " ) words.append(splits[0] ) if len(UpperCamelCase ) > 1: labels.append(splits[self.label_idx].replace("\n" , "" ) ) else: # Examples could have no label for mode = "test" labels.append("O" ) if words: examples.append(InputExample(guid=f'''{mode}-{guid_index}''' , words=UpperCamelCase , labels=UpperCamelCase ) ) return examples def snake_case ( self , UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCamelCase_ = 0 for line in test_input_reader: if line.startswith("-DOCSTART-" ) or line == "" or line == "\n": writer.write(UpperCamelCase ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: lowerCamelCase_ = line.split()[0] + " " + preds_list[example_id].pop(0 ) + "\n" writer.write(UpperCamelCase ) else: logger.warning("Maximum sequence length exceeded: No prediction for '%s'." , line.split()[0] ) def snake_case ( self , UpperCamelCase ): """simple docstring""" if path: with open(UpperCamelCase , "r" ) as f: lowerCamelCase_ = f.read().splitlines() if "O" not in labels: lowerCamelCase_ = ["O"] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class snake_case ( lowercase ): """simple docstring""" def __init__( self ): """simple docstring""" # in CONLL2003 dataset chunk column is second-to-last super().__init__(label_idx=-2 ) def snake_case ( self , UpperCamelCase ): """simple docstring""" if path: with open(UpperCamelCase , "r" ) as f: lowerCamelCase_ = f.read().splitlines() if "O" not in labels: lowerCamelCase_ = ["O"] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class snake_case ( lowercase ): """simple docstring""" def snake_case ( self , UpperCamelCase , UpperCamelCase ): """simple docstring""" if isinstance(UpperCamelCase , UpperCamelCase ): lowerCamelCase_ = mode.value lowerCamelCase_ = os.path.join(UpperCamelCase , f'''{mode}.txt''' ) lowerCamelCase_ = 1 lowerCamelCase_ = [] with open(UpperCamelCase , encoding="utf-8" ) as f: for sentence in parse_incr(UpperCamelCase ): lowerCamelCase_ = [] lowerCamelCase_ = [] for token in sentence: words.append(token["form"] ) labels.append(token["upos"] ) assert len(UpperCamelCase ) == len(UpperCamelCase ) if words: examples.append(InputExample(guid=f'''{mode}-{guid_index}''' , words=UpperCamelCase , labels=UpperCamelCase ) ) guid_index += 1 return examples def snake_case ( self , UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCamelCase_ = 0 for sentence in parse_incr(UpperCamelCase ): lowerCamelCase_ = preds_list[example_id] lowerCamelCase_ = "" for token in sentence: out += f'''{token["form"]} ({token["upos"]}|{s_p.pop(0 )}) ''' out += "\n" writer.write(UpperCamelCase ) example_id += 1 def snake_case ( self , UpperCamelCase ): """simple docstring""" if path: with open(UpperCamelCase , "r" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
55
'''simple docstring''' import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class snake_case ( lowercase , lowercase , lowercase , unittest.TestCase ): """simple docstring""" _lowerCamelCase = StableUnCLIPPipeline _lowerCamelCase = TEXT_TO_IMAGE_PARAMS _lowerCamelCase = TEXT_TO_IMAGE_BATCH_PARAMS _lowerCamelCase = TEXT_TO_IMAGE_IMAGE_PARAMS _lowerCamelCase = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false _lowerCamelCase = False def snake_case ( self ): """simple docstring""" lowerCamelCase_ = 32 lowerCamelCase_ = embedder_hidden_size # prior components torch.manual_seed(0 ) lowerCamelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) lowerCamelCase_ = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=UpperCamelCase , projection_dim=UpperCamelCase , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) lowerCamelCase_ = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=UpperCamelCase , num_layers=1 , ) torch.manual_seed(0 ) lowerCamelCase_ = DDPMScheduler( variance_type="fixed_small_log" , prediction_type="sample" , num_train_timesteps=1000 , clip_sample=UpperCamelCase , clip_sample_range=5.0 , beta_schedule="squaredcos_cap_v2" , ) # regular denoising components torch.manual_seed(0 ) lowerCamelCase_ = StableUnCLIPImageNormalizer(embedding_dim=UpperCamelCase ) lowerCamelCase_ = DDPMScheduler(beta_schedule="squaredcos_cap_v2" ) torch.manual_seed(0 ) lowerCamelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) lowerCamelCase_ = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=UpperCamelCase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) lowerCamelCase_ = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="projection" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=UpperCamelCase , layers_per_block=1 , upcast_attention=UpperCamelCase , use_linear_projection=UpperCamelCase , ) torch.manual_seed(0 ) lowerCamelCase_ = DDIMScheduler( beta_schedule="scaled_linear" , beta_start=0.00_085 , beta_end=0.012 , prediction_type="v_prediction" , set_alpha_to_one=UpperCamelCase , steps_offset=1 , ) torch.manual_seed(0 ) lowerCamelCase_ = AutoencoderKL() lowerCamelCase_ = { # prior components "prior_tokenizer": prior_tokenizer, "prior_text_encoder": prior_text_encoder, "prior": prior, "prior_scheduler": prior_scheduler, # image noising components "image_normalizer": image_normalizer, "image_noising_scheduler": image_noising_scheduler, # regular denoising components "tokenizer": tokenizer, "text_encoder": text_encoder, "unet": unet, "scheduler": scheduler, "vae": vae, } return components def snake_case ( self , UpperCamelCase , UpperCamelCase=0 ): """simple docstring""" if str(UpperCamelCase ).startswith("mps" ): lowerCamelCase_ = torch.manual_seed(UpperCamelCase ) else: lowerCamelCase_ = torch.Generator(device=UpperCamelCase ).manual_seed(UpperCamelCase ) lowerCamelCase_ = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "prior_num_inference_steps": 2, "output_type": "numpy", } return inputs def snake_case ( self ): """simple docstring""" lowerCamelCase_ = torch_device == "cpu" self._test_attention_slicing_forward_pass(test_max_difference=UpperCamelCase ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = torch_device in ["cpu", "mps"] self._test_inference_batch_single_identical(test_max_difference=UpperCamelCase ) @slow @require_torch_gpu class snake_case ( unittest.TestCase ): """simple docstring""" def snake_case ( self ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case ( self ): """simple docstring""" lowerCamelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy" ) lowerCamelCase_ = StableUnCLIPPipeline.from_pretrained("fusing/stable-unclip-2-1-l" , torch_dtype=torch.floataa ) pipe.to(UpperCamelCase ) pipe.set_progress_bar_config(disable=UpperCamelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCamelCase_ = torch.Generator(device="cpu" ).manual_seed(0 ) lowerCamelCase_ = pipe("anime turle" , generator=UpperCamelCase , output_type="np" ) lowerCamelCase_ = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(UpperCamelCase , UpperCamelCase ) def snake_case ( self ): """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCamelCase_ = StableUnCLIPPipeline.from_pretrained("fusing/stable-unclip-2-1-l" , torch_dtype=torch.floataa ) lowerCamelCase_ = pipe.to(UpperCamelCase ) pipe.set_progress_bar_config(disable=UpperCamelCase ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCamelCase_ = pipe( "anime turtle" , prior_num_inference_steps=2 , num_inference_steps=2 , output_type="np" , ) lowerCamelCase_ = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
55
1
"""simple docstring""" _lowercase : int = 6_55_21 def lowercase__ ( snake_case_ :str ): __UpperCAmelCase = 1 __UpperCAmelCase = 0 for plain_chr in plain_text: __UpperCAmelCase = (a + ord(snake_case_ )) % MOD_ADLER __UpperCAmelCase = (b + a) % MOD_ADLER return (b << 16) | a
86
"""simple docstring""" from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer _lowercase : int = logging.get_logger(__name__) _lowercase : Tuple = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} _lowercase : Tuple = { 'vocab_file': { 'allegro/herbert-base-cased': 'https://huggingface.co/allegro/herbert-base-cased/resolve/main/vocab.json' }, 'merges_file': { 'allegro/herbert-base-cased': 'https://huggingface.co/allegro/herbert-base-cased/resolve/main/merges.txt' }, } _lowercase : str = {'allegro/herbert-base-cased': 5_14} _lowercase : Tuple = {} class _UpperCAmelCase ( _lowerCAmelCase ): a__ : Optional[Any] = VOCAB_FILES_NAMES a__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP a__ : Any = PRETRAINED_INIT_CONFIGURATION a__ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ : Optional[Any] = HerbertTokenizer def __init__( self : List[Any] , _lowercase : Optional[int]=None , _lowercase : int=None , _lowercase : Tuple=None , _lowercase : str="<s>" , _lowercase : List[str]="<unk>" , _lowercase : int="<pad>" , _lowercase : str="<mask>" , _lowercase : List[Any]="</s>" , **_lowercase : List[Any] , ): super().__init__( _lowercase , _lowercase , tokenizer_file=_lowercase , cls_token=_lowercase , unk_token=_lowercase , pad_token=_lowercase , mask_token=_lowercase , sep_token=_lowercase , **_lowercase , ) def a ( self : Optional[int] , _lowercase : List[int] , _lowercase : Optional[List[int]] = None ): __UpperCAmelCase = [self.cls_token_id] __UpperCAmelCase = [self.sep_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def a ( self : Any , _lowercase : List[int] , _lowercase : Optional[List[int]] = None , _lowercase : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowercase , token_ids_a=_lowercase , already_has_special_tokens=_lowercase ) if token_ids_a is None: return [1] + ([0] * len(_lowercase )) + [1] return [1] + ([0] * len(_lowercase )) + [1] + ([0] * len(_lowercase )) + [1] def a ( self : str , _lowercase : List[int] , _lowercase : Optional[List[int]] = None ): __UpperCAmelCase = [self.sep_token_id] __UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def a ( self : str , _lowercase : str , _lowercase : Optional[str] = None ): __UpperCAmelCase = self._tokenizer.model.save(_lowercase , name=_lowercase ) return tuple(_lowercase )
86
1
__lowerCamelCase : List[str] = {str(digit): digit**5 for digit in range(10)} def _snake_case ( lowerCAmelCase : int ): """simple docstring""" return sum(DIGITS_FIFTH_POWER[digit] for digit in str(lowerCAmelCase ) ) def _snake_case ( ): """simple docstring""" return sum( number for number in range(1_0_0_0 , 1_0_0_0_0_0_0 ) if number == digits_fifth_powers_sum(lowerCAmelCase ) ) if __name__ == "__main__": print(solution())
18
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __UpperCamelCase ( lowerCAmelCase__ , unittest.TestCase ): """simple docstring""" lowerCAmelCase_ = KandinskyVaaPriorPipeline lowerCAmelCase_ = ['''prompt'''] lowerCAmelCase_ = ['''prompt''', '''negative_prompt'''] lowerCAmelCase_ = [ '''num_images_per_prompt''', '''generator''', '''num_inference_steps''', '''latents''', '''negative_prompt''', '''guidance_scale''', '''output_type''', '''return_dict''', ] lowerCAmelCase_ = False @property def UpperCAmelCase__ ( self : int ): """simple docstring""" return 32 @property def UpperCAmelCase__ ( self : List[str] ): """simple docstring""" return 32 @property def UpperCAmelCase__ ( self : Dict ): """simple docstring""" return self.time_input_dim @property def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" return self.time_input_dim * 4 @property def UpperCAmelCase__ ( self : List[Any] ): """simple docstring""" return 100 @property def UpperCAmelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def UpperCAmelCase__ ( self : List[Any] ): """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : List[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_A ) @property def UpperCAmelCase__ ( self : str ): """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : Dict = { '''num_attention_heads''': 2, '''attention_head_dim''': 12, '''embedding_dim''': self.text_embedder_hidden_size, '''num_layers''': 1, } __SCREAMING_SNAKE_CASE : Optional[Any] = PriorTransformer(**_A ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 __SCREAMING_SNAKE_CASE : List[Any] = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : int = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) __SCREAMING_SNAKE_CASE : str = CLIPVisionModelWithProjection(_A ) return model @property def UpperCAmelCase__ ( self : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = CLIPImageProcessor( crop_size=224 , do_center_crop=_A , do_normalize=_A , do_resize=_A , image_mean=[0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73] , image_std=[0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11] , resample=3 , size=224 , ) return image_processor def UpperCAmelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = self.dummy_prior __SCREAMING_SNAKE_CASE : str = self.dummy_image_encoder __SCREAMING_SNAKE_CASE : Optional[int] = self.dummy_text_encoder __SCREAMING_SNAKE_CASE : List[Any] = self.dummy_tokenizer __SCREAMING_SNAKE_CASE : Union[str, Any] = self.dummy_image_processor __SCREAMING_SNAKE_CASE : str = UnCLIPScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=1000 , clip_sample=_A , clip_sample_range=10.0 , ) __SCREAMING_SNAKE_CASE : int = { '''prior''': prior, '''image_encoder''': image_encoder, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''scheduler''': scheduler, '''image_processor''': image_processor, } return components def UpperCAmelCase__ ( self : Union[str, Any] , _A : int , _A : Dict=0 ): """simple docstring""" if str(_A ).startswith('''mps''' ): __SCREAMING_SNAKE_CASE : Tuple = torch.manual_seed(_A ) else: __SCREAMING_SNAKE_CASE : str = torch.Generator(device=_A ).manual_seed(_A ) __SCREAMING_SNAKE_CASE : List[str] = { '''prompt''': '''horse''', '''generator''': generator, '''guidance_scale''': 4.0, '''num_inference_steps''': 2, '''output_type''': '''np''', } return inputs def UpperCAmelCase__ ( self : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = '''cpu''' __SCREAMING_SNAKE_CASE : Any = self.get_dummy_components() __SCREAMING_SNAKE_CASE : Any = self.pipeline_class(**_A ) __SCREAMING_SNAKE_CASE : List[Any] = pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) __SCREAMING_SNAKE_CASE : int = pipe(**self.get_dummy_inputs(_A ) ) __SCREAMING_SNAKE_CASE : Tuple = output.image_embeds __SCREAMING_SNAKE_CASE : Optional[Any] = pipe( **self.get_dummy_inputs(_A ) , return_dict=_A , )[0] __SCREAMING_SNAKE_CASE : Tuple = image[0, -10:] __SCREAMING_SNAKE_CASE : List[Any] = image_from_tuple[0, -10:] assert image.shape == (1, 32) __SCREAMING_SNAKE_CASE : List[str] = np.array( [-0.05_32, 1.71_20, 0.36_56, -1.08_52, -0.89_46, -1.17_56, 0.43_48, 0.24_82, 0.51_46, -0.11_56] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def UpperCAmelCase__ ( self : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = torch_device == '''cpu''' __SCREAMING_SNAKE_CASE : Any = True __SCREAMING_SNAKE_CASE : int = False self._test_inference_batch_single_identical( test_max_difference=_A , relax_max_difference=_A , test_mean_pixel_difference=_A , ) @skip_mps def UpperCAmelCase__ ( self : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = torch_device == '''cpu''' __SCREAMING_SNAKE_CASE : List[Any] = False self._test_attention_slicing_forward_pass( test_max_difference=_A , test_mean_pixel_difference=_A , )
303
0
"""simple docstring""" import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class UpperCAmelCase_ ( unittest.TestCase ): def __init__( self : List[Any] , A : Optional[int] , A : bool = True , A : Dict[str, int] = None , A : int = 3_2 , A : bool = True , A : Union[int, float] = 1 / 2_5_5 , A : bool = True , A : bool = True , A : Optional[Union[float, List[float]]] = [0.48_145_466, 0.4_578_275, 0.40_821_073] , A : Optional[Union[float, List[float]]] = [0.26_862_954, 0.26_130_258, 0.27_577_711] , A : bool = True , A : Optional[Any]=7 , A : str=3_0 , A : List[str]=4_0_0 , A : int=3 , ): _UpperCAmelCase : int = parent _UpperCAmelCase : Dict = do_resize _UpperCAmelCase : List[str] = size if size is not None else {"shortest_edge": 2_8_8} _UpperCAmelCase : int = size_divisor _UpperCAmelCase : List[Any] = do_rescale _UpperCAmelCase : Any = rescale_factor _UpperCAmelCase : Tuple = do_normalize _UpperCAmelCase : str = do_center_crop _UpperCAmelCase : List[str] = image_mean _UpperCAmelCase : List[Any] = image_std _UpperCAmelCase : List[str] = do_pad _UpperCAmelCase : List[Any] = batch_size _UpperCAmelCase : Union[str, Any] = num_channels _UpperCAmelCase : Optional[int] = min_resolution _UpperCAmelCase : Union[str, Any] = max_resolution def snake_case_ ( self : int ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def snake_case_ ( self : Tuple , A : Tuple , A : List[Any]=False ): if not batched: _UpperCAmelCase : List[str] = self.size["shortest_edge"] _UpperCAmelCase : Optional[int] = image_inputs[0] if isinstance(A , Image.Image ): _UpperCAmelCase , _UpperCAmelCase : List[Any] = image.size else: _UpperCAmelCase , _UpperCAmelCase : Tuple = image.shape[1], image.shape[2] _UpperCAmelCase : int = size / min(A , A ) if h < w: _UpperCAmelCase , _UpperCAmelCase : str = size, scale * w else: _UpperCAmelCase , _UpperCAmelCase : int = scale * h, size _UpperCAmelCase : List[Any] = int((1_3_3_3 / 8_0_0) * size ) if max(A , A ) > max_size: _UpperCAmelCase : Optional[Any] = max_size / max(A , A ) _UpperCAmelCase : Optional[Any] = newh * scale _UpperCAmelCase : Tuple = neww * scale _UpperCAmelCase , _UpperCAmelCase : Dict = int(newh + 0.5 ), int(neww + 0.5 ) _UpperCAmelCase , _UpperCAmelCase : Any = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: _UpperCAmelCase : List[Any] = [] for image in image_inputs: _UpperCAmelCase , _UpperCAmelCase : Any = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _UpperCAmelCase : Any = max(A , key=lambda A : item[0] )[0] _UpperCAmelCase : Tuple = max(A , key=lambda A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCAmelCase_ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE : Union[str, Any] = BridgeTowerImageProcessor if is_vision_available() else None def snake_case_ ( self : Optional[Any] ): _UpperCAmelCase : Tuple = BridgeTowerImageProcessingTester(self ) @property def snake_case_ ( self : Tuple ): return self.image_processor_tester.prepare_image_processor_dict() def snake_case_ ( self : int ): _UpperCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A , "image_mean" ) ) self.assertTrue(hasattr(A , "image_std" ) ) self.assertTrue(hasattr(A , "do_normalize" ) ) self.assertTrue(hasattr(A , "do_resize" ) ) self.assertTrue(hasattr(A , "size" ) ) self.assertTrue(hasattr(A , "size_divisor" ) ) def snake_case_ ( self : List[Any] ): pass def snake_case_ ( self : List[Any] ): # Initialize image processor _UpperCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _UpperCAmelCase : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A , Image.Image ) # Test not batched input _UpperCAmelCase : Tuple = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values _UpperCAmelCase , _UpperCAmelCase : str = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _UpperCAmelCase : Optional[Any] = image_processing(A , return_tensors="pt" ).pixel_values _UpperCAmelCase , _UpperCAmelCase : int = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def snake_case_ ( self : Dict ): # Initialize image processor _UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _UpperCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , numpify=A ) for image in image_inputs: self.assertIsInstance(A , np.ndarray ) # Test not batched input _UpperCAmelCase : Dict = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values _UpperCAmelCase , _UpperCAmelCase : List[Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _UpperCAmelCase : List[Any] = image_processing(A , return_tensors="pt" ).pixel_values _UpperCAmelCase , _UpperCAmelCase : str = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def snake_case_ ( self : Dict ): # Initialize image processor _UpperCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _UpperCAmelCase : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test not batched input _UpperCAmelCase : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _UpperCAmelCase : int = image_processing(A , return_tensors="pt" ).pixel_values _UpperCAmelCase , _UpperCAmelCase : Dict = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
202
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase : Union[str, Any] = { "configuration_roformer": ["ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "RoFormerConfig", "RoFormerOnnxConfig"], "tokenization_roformer": ["RoFormerTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Any = ["RoFormerTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Any = [ "ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "RoFormerForCausalLM", "RoFormerForMaskedLM", "RoFormerForMultipleChoice", "RoFormerForQuestionAnswering", "RoFormerForSequenceClassification", "RoFormerForTokenClassification", "RoFormerLayer", "RoFormerModel", "RoFormerPreTrainedModel", "load_tf_weights_in_roformer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Optional[int] = [ "TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRoFormerForCausalLM", "TFRoFormerForMaskedLM", "TFRoFormerForMultipleChoice", "TFRoFormerForQuestionAnswering", "TFRoFormerForSequenceClassification", "TFRoFormerForTokenClassification", "TFRoFormerLayer", "TFRoFormerModel", "TFRoFormerPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : List[str] = [ "FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "FlaxRoFormerForMaskedLM", "FlaxRoFormerForMultipleChoice", "FlaxRoFormerForQuestionAnswering", "FlaxRoFormerForSequenceClassification", "FlaxRoFormerForTokenClassification", "FlaxRoFormerModel", "FlaxRoFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys _lowerCAmelCase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
202
1
import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin snake_case_ : Optional[Any] = get_tests_dir("fixtures/spiece.model") @require_sentencepiece @require_tokenizers class __snake_case ( a , unittest.TestCase ): UpperCAmelCase__ : int = DebertaVaTokenizer UpperCAmelCase__ : int = DebertaVaTokenizerFast UpperCAmelCase__ : Tuple = True UpperCAmelCase__ : Tuple = True def lowerCamelCase ( self : Optional[int]): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase_ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE__ , unk_token='''<unk>''') tokenizer.save_pretrained(self.tmpdirname) def lowerCamelCase ( self : Dict , _snake_case : Dict): """simple docstring""" UpperCAmelCase_ = 'this is a test' UpperCAmelCase_ = 'this is a test' return input_text, output_text def lowerCamelCase ( self : Any): """simple docstring""" UpperCAmelCase_ = '<pad>' UpperCAmelCase_ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(SCREAMING_SNAKE_CASE__) , SCREAMING_SNAKE_CASE__) self.assertEqual(self.get_tokenizer()._convert_id_to_token(SCREAMING_SNAKE_CASE__) , SCREAMING_SNAKE_CASE__) def lowerCamelCase ( self : List[Any]): """simple docstring""" UpperCAmelCase_ = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '''<pad>''') self.assertEqual(vocab_keys[1] , '''<unk>''') self.assertEqual(vocab_keys[-1] , '''[PAD]''') self.assertEqual(len(SCREAMING_SNAKE_CASE__) , 30001) def lowerCamelCase ( self : List[Any]): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 30000) def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = ' \tHeLLo!how \n Are yoU? ' UpperCAmelCase_ = ['▁hello', '!', 'how', '▁are', '▁you', '?'] # fmt: on UpperCAmelCase_ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) @unittest.skip('''There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.''') def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" pass @unittest.skip('''There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.''') def lowerCamelCase ( self : int): """simple docstring""" pass def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = 'I was born in 92000, and this is falsé.' UpperCAmelCase_ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on UpperCAmelCase_ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" UpperCAmelCase_ = 'I was born in 92000, and this is falsé.' UpperCAmelCase_ = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on UpperCAmelCase_ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def lowerCamelCase ( self : List[Any]): """simple docstring""" UpperCAmelCase_ = 'I was born in 92000, and this is falsé.' UpperCAmelCase_ = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on UpperCAmelCase_ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def lowerCamelCase ( self : Tuple): """simple docstring""" UpperCAmelCase_ = 'I was born in 92000, and this is falsé.' UpperCAmelCase_ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on UpperCAmelCase_ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ = ' \tHeLLo!how \n Are yoU? ' UpperCAmelCase_ = ['▁', '<unk>', 'e', '<unk>', 'o', '!', 'how', '▁', '<unk>', 're', '▁yo', '<unk>', '?'] # fmt: on UpperCAmelCase_ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE__ , do_lower_case=SCREAMING_SNAKE_CASE__ , split_by_punct=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def lowerCamelCase ( self : int): """simple docstring""" UpperCAmelCase_ = self.get_tokenizer() UpperCAmelCase_ = self.get_rust_tokenizer() UpperCAmelCase_ = 'I was born in 92000, and this is falsé.' UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) UpperCAmelCase_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__)) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = self.get_rust_tokenizer() UpperCAmelCase_ = tokenizer.encode(SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.encode(SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = 'This is a test' UpperCAmelCase_ = [13, 1, 4398, 25, 21, 1289] UpperCAmelCase_ = ['▁', 'T', 'his', '▁is', '▁a', '▁test'] UpperCAmelCase_ = ['▁', '<unk>', 'his', '▁is', '▁a', '▁test'] UpperCAmelCase_ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE__ , keep_accents=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE__ , keep_accents=SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.tokenize(SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.tokenize(SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) # fmt: off UpperCAmelCase_ = 'I was born in 92000, and this is falsé.' UpperCAmelCase_ = [13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] UpperCAmelCase_ = ['▁', 'I', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.', ] UpperCAmelCase_ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on UpperCAmelCase_ = tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.tokenize(SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.tokenize(SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = rust_tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def lowerCamelCase ( self : Optional[Any]): """simple docstring""" UpperCAmelCase_ = DebertaVaTokenizer(SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.encode('''sequence builders''') UpperCAmelCase_ = tokenizer.encode('''multi-sequence build''') UpperCAmelCase_ = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE__) UpperCAmelCase_ = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) self.assertEqual([tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] , SCREAMING_SNAKE_CASE__) self.assertEqual( [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [tokenizer.sep_token_id] , SCREAMING_SNAKE_CASE__ , ) @slow def lowerCamelCase ( self : Any): """simple docstring""" UpperCAmelCase_ = {'input_ids': [[1, 39867, 36, 19390, 486, 27, 35052, 81436, 18, 60685, 1225, 7, 35052, 81436, 18, 9367, 16899, 18, 15937, 53, 594, 773, 18, 16287, 30465, 36, 15937, 6, 41139, 38, 36979, 60763, 191, 6, 34132, 99, 6, 50538, 390, 43230, 6, 34132, 2779, 20850, 14, 699, 1072, 1194, 36, 382, 10901, 53, 7, 699, 1072, 2084, 36, 20422, 630, 53, 19, 105, 3049, 1896, 1053, 16899, 1506, 11, 37978, 4243, 7, 1237, 31869, 200, 16566, 654, 6, 35052, 81436, 7, 55630, 13593, 4, 2], [1, 26, 15011, 13, 667, 8, 1053, 18, 23611, 1237, 72356, 12820, 34, 104134, 1209, 35, 13313, 6627, 21, 202, 347, 7, 164, 2399, 11, 46, 4485, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 5, 1232, 2864, 15785, 14951, 105, 5, 8581, 1250, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=SCREAMING_SNAKE_CASE__ , model_name='''microsoft/deberta-v2-xlarge''' , revision='''ad6e42c1532ddf3a15c39246b63f5559d558b670''' , )
51
from math import isclose, sqrt def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> tuple[float, float, float]: __lowerCamelCase : Tuple = point_y / 4 / point_x __lowerCamelCase : Tuple = 2 * normal_gradient / (1 + normal_gradient * normal_gradient) __lowerCamelCase : List[Any] = (1 - normal_gradient * normal_gradient) / ( 1 + normal_gradient * normal_gradient ) __lowerCamelCase : int = (sa - ca * incoming_gradient) / (ca + sa * incoming_gradient) # to find the next point, solve the simultaeneous equations: # y^2 + 4x^2 = 100 # y - b = m * (x - a) # ==> A x^2 + B x + C = 0 __lowerCamelCase : Any = outgoing_gradient**2 + 4 __lowerCamelCase : Optional[int] = 2 * outgoing_gradient * (point_y - outgoing_gradient * point_x) __lowerCamelCase : str = (point_y - outgoing_gradient * point_x) ** 2 - 1_0_0 __lowerCamelCase : str = ( -linear_term - sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) __lowerCamelCase : Optional[Any] = ( -linear_term + sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) # two solutions, one of which is our input point __lowerCamelCase : Optional[Any] = x_minus if isclose(lowerCamelCase__ , lowerCamelCase__ ) else x_plus __lowerCamelCase : Tuple = point_y + outgoing_gradient * (next_x - point_x) return next_x, next_y, outgoing_gradient def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ = 1.4 , lowerCamelCase__ = -9.6 ) -> int: __lowerCamelCase : int = 0 __lowerCamelCase : float = first_x_coord __lowerCamelCase : float = first_y_coord __lowerCamelCase : float = (10.1 - point_y) / (0.0 - point_x) while not (-0.01 <= point_x <= 0.01 and point_y > 0): __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : Any = next_point(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) num_reflections += 1 return num_reflections if __name__ == "__main__": print(F"""{solution() = }""")
73
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase_ = { '''configuration_clap''': [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapAudioConfig''', '''ClapConfig''', '''ClapTextConfig''', ], '''processing_clap''': ['''ClapProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapModel''', '''ClapPreTrainedModel''', '''ClapTextModel''', '''ClapTextModelWithProjection''', '''ClapAudioModel''', '''ClapAudioModelWithProjection''', ] UpperCamelCase_ = ['''ClapFeatureExtractor'''] if TYPE_CHECKING: from .configuration_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioConfig, ClapConfig, ClapTextConfig, ) from .processing_clap import ClapProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clap import ClapFeatureExtractor from .modeling_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioModel, ClapAudioModelWithProjection, ClapModel, ClapPreTrainedModel, ClapTextModel, ClapTextModelWithProjection, ) else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
353
def lowerCamelCase_ ( _a : int ): '''simple docstring''' if a < 0: raise ValueError("""Input value must be a positive integer""" ) elif isinstance(_a , _a ): raise TypeError("""Input value must be a 'int' type""" ) return bin(_a ).count("""1""" ) if __name__ == "__main__": import doctest doctest.testmod()
59
0
import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging __UpperCAmelCase : str = logging.get_logger(__name__) class __snake_case ( __lowerCamelCase ): '''simple docstring''' lowerCAmelCase__ = """linear""" lowerCAmelCase__ = """cosine""" lowerCAmelCase__ = """cosine_with_restarts""" lowerCAmelCase__ = """polynomial""" lowerCAmelCase__ = """constant""" lowerCAmelCase__ = """constant_with_warmup""" lowerCAmelCase__ = """piecewise_constant""" def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = -1) -> Any: return LambdaLR(SCREAMING_SNAKE_CASE__ , lambda SCREAMING_SNAKE_CASE__: 1 , last_epoch=SCREAMING_SNAKE_CASE__) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = -1) -> Optional[int]: def lr_lambda(SCREAMING_SNAKE_CASE__): if current_step < num_warmup_steps: return float(SCREAMING_SNAKE_CASE__) / float(max(1.0 , SCREAMING_SNAKE_CASE__)) return 1.0 return LambdaLR(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , last_epoch=SCREAMING_SNAKE_CASE__) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = -1) -> Tuple: __snake_case: str = {} __snake_case: Union[str, Any] = step_rules.split(""",""") for rule_str in rule_list[:-1]: __snake_case , __snake_case: Optional[int] = rule_str.split(""":""") __snake_case: Dict = int(SCREAMING_SNAKE_CASE__) __snake_case: int = float(SCREAMING_SNAKE_CASE__) __snake_case: Optional[Any] = value __snake_case: List[str] = float(rule_list[-1]) def create_rules_function(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): def rule_func(SCREAMING_SNAKE_CASE__) -> float: __snake_case: Union[str, Any] = sorted(rules_dict.keys()) for i, sorted_step in enumerate(SCREAMING_SNAKE_CASE__): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func __snake_case: List[Any] = create_rules_function(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) return LambdaLR(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , last_epoch=SCREAMING_SNAKE_CASE__) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=-1) -> Dict: def lr_lambda(SCREAMING_SNAKE_CASE__): if current_step < num_warmup_steps: return float(SCREAMING_SNAKE_CASE__) / float(max(1 , SCREAMING_SNAKE_CASE__)) return max( 0.0 , float(num_training_steps - current_step) / float(max(1 , num_training_steps - num_warmup_steps))) return LambdaLR(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 0.5 , SCREAMING_SNAKE_CASE__ = -1) -> Tuple: def lr_lambda(SCREAMING_SNAKE_CASE__): if current_step < num_warmup_steps: return float(SCREAMING_SNAKE_CASE__) / float(max(1 , SCREAMING_SNAKE_CASE__)) __snake_case: Any = float(current_step - num_warmup_steps) / float(max(1 , num_training_steps - num_warmup_steps)) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(SCREAMING_SNAKE_CASE__) * 2.0 * progress))) return LambdaLR(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 1 , SCREAMING_SNAKE_CASE__ = -1) -> List[Any]: def lr_lambda(SCREAMING_SNAKE_CASE__): if current_step < num_warmup_steps: return float(SCREAMING_SNAKE_CASE__) / float(max(1 , SCREAMING_SNAKE_CASE__)) __snake_case: Any = float(current_step - num_warmup_steps) / float(max(1 , num_training_steps - num_warmup_steps)) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(SCREAMING_SNAKE_CASE__) * progress) % 1.0)))) return LambdaLR(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=1e-7 , SCREAMING_SNAKE_CASE__=1.0 , SCREAMING_SNAKE_CASE__=-1) -> Dict: __snake_case: Any = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(F'''lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})''') def lr_lambda(SCREAMING_SNAKE_CASE__): if current_step < num_warmup_steps: return float(SCREAMING_SNAKE_CASE__) / float(max(1 , SCREAMING_SNAKE_CASE__)) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: __snake_case: List[str] = lr_init - lr_end __snake_case: Any = num_training_steps - num_warmup_steps __snake_case: Any = 1 - (current_step - num_warmup_steps) / decay_steps __snake_case: Any = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__) __UpperCAmelCase : Optional[int] = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def A__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = 1 , SCREAMING_SNAKE_CASE__ = 1.0 , SCREAMING_SNAKE_CASE__ = -1 , ) -> Any: __snake_case: List[str] = SchedulerType(SCREAMING_SNAKE_CASE__) __snake_case: Any = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(SCREAMING_SNAKE_CASE__ , last_epoch=SCREAMING_SNAKE_CASE__) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(SCREAMING_SNAKE_CASE__ , step_rules=SCREAMING_SNAKE_CASE__ , last_epoch=SCREAMING_SNAKE_CASE__) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F'''{name} requires `num_warmup_steps`, please provide that argument.''') if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(SCREAMING_SNAKE_CASE__ , num_warmup_steps=SCREAMING_SNAKE_CASE__ , last_epoch=SCREAMING_SNAKE_CASE__) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F'''{name} requires `num_training_steps`, please provide that argument.''') if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( SCREAMING_SNAKE_CASE__ , num_warmup_steps=SCREAMING_SNAKE_CASE__ , num_training_steps=SCREAMING_SNAKE_CASE__ , num_cycles=SCREAMING_SNAKE_CASE__ , last_epoch=SCREAMING_SNAKE_CASE__ , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( SCREAMING_SNAKE_CASE__ , num_warmup_steps=SCREAMING_SNAKE_CASE__ , num_training_steps=SCREAMING_SNAKE_CASE__ , power=SCREAMING_SNAKE_CASE__ , last_epoch=SCREAMING_SNAKE_CASE__ , ) return schedule_func( SCREAMING_SNAKE_CASE__ , num_warmup_steps=SCREAMING_SNAKE_CASE__ , num_training_steps=SCREAMING_SNAKE_CASE__ , last_epoch=SCREAMING_SNAKE_CASE__)
111
from __future__ import annotations import typing from collections import Counter def A__ ( SCREAMING_SNAKE_CASE__) -> typing.Counter[int]: __snake_case: typing.Counter[int] = Counter() for base in range(1 , max_perimeter + 1): for perpendicular in range(SCREAMING_SNAKE_CASE__ , max_perimeter + 1): __snake_case: Dict = (base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(SCREAMING_SNAKE_CASE__): __snake_case: Any = int(base + perpendicular + hypotenuse) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def A__ ( SCREAMING_SNAKE_CASE__ = 1000) -> int: __snake_case: List[str] = pythagorean_triple(SCREAMING_SNAKE_CASE__) return triplets.most_common(1)[0][0] if __name__ == "__main__": print(f'Perimeter {solution()} has maximum solutions')
111
1
from __future__ import annotations class __snake_case : def __init__( self : Any , _UpperCAmelCase : int ) -> None: '''simple docstring''' _lowerCAmelCase : Tuple = data _lowerCAmelCase : Node | None = None _lowerCAmelCase : Node | None = None def _UpperCAmelCase (UpperCamelCase_ : Node | None ): # In Order traversal of the tree '''simple docstring''' if tree: display(tree.left ) print(tree.data ) display(tree.right ) def _UpperCAmelCase (UpperCamelCase_ : Node | None ): '''simple docstring''' return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def _UpperCAmelCase (UpperCamelCase_ : Node ): '''simple docstring''' if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def _UpperCAmelCase (): # Main function for testing. '''simple docstring''' _lowerCAmelCase : Union[str, Any] = Node(1 ) _lowerCAmelCase : Optional[int] = Node(2 ) _lowerCAmelCase : List[str] = Node(3 ) _lowerCAmelCase : Optional[int] = Node(4 ) _lowerCAmelCase : Union[str, Any] = Node(5 ) _lowerCAmelCase : List[str] = Node(6 ) _lowerCAmelCase : int = Node(7 ) _lowerCAmelCase : Dict = Node(8 ) _lowerCAmelCase : List[Any] = Node(9 ) print(is_full_binary_tree(UpperCamelCase_ ) ) print(depth_of_tree(UpperCamelCase_ ) ) print("""Tree is: """ ) display(UpperCamelCase_ ) if __name__ == "__main__": main()
356
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowerCamelCase : Optional[Any] = { "configuration_conditional_detr": [ "CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConditionalDetrConfig", "ConditionalDetrOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Optional[int] = ["ConditionalDetrFeatureExtractor"] _lowerCamelCase : Optional[Any] = ["ConditionalDetrImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Dict = [ "CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST", "ConditionalDetrForObjectDetection", "ConditionalDetrForSegmentation", "ConditionalDetrModel", "ConditionalDetrPreTrainedModel", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys _lowerCamelCase : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
159
0
'''simple docstring''' import importlib.metadata import operator import re import sys from typing import Optional from packaging import version a : Dict = { '<': operator.lt, '<=': operator.le, '==': operator.eq, '!=': operator.ne, '>=': operator.ge, '>': operator.gt, } def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) -> int: '''simple docstring''' if got_ver is None or want_ver is None: raise ValueError( F"Unable to compare versions for {requirement}: need={want_ver} found={got_ver}. This is unusual. Consider" F" reinstalling {pkg}." ) if not ops[op](version.parse(__UpperCAmelCase ), version.parse(__UpperCAmelCase ) ): raise ImportError( F"{requirement} is required for a normal functioning of this module, but found {pkg}=={got_ver}.{hint}" ) def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase = None ) -> None: '''simple docstring''' snake_case_ = F"\n{hint}" if hint is not None else '''''' # non-versioned check if re.match(r'''^[\w_\-\d]+$''', __UpperCAmelCase ): snake_case_ ,snake_case_ ,snake_case_ = requirement, None, None else: snake_case_ = re.findall(r'''^([^!=<>\s]+)([\s!=<>]{1,2}.+)''', __UpperCAmelCase ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23, but''' F" got {requirement}" ) snake_case_ ,snake_case_ = match[0] snake_case_ = want_full.split(''',''' ) # there could be multiple requirements snake_case_ = {} for w in want_range: snake_case_ = re.findall(r'''^([\s!=<>]{1,2})(.+)''', __UpperCAmelCase ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23,''' F" but got {requirement}" ) snake_case_ ,snake_case_ = match[0] snake_case_ = want_ver if op not in ops: raise ValueError(F"{requirement}: need one of {list(ops.keys() )}, but got {op}" ) # special case if pkg == "python": snake_case_ = '''.'''.join([str(__UpperCAmelCase ) for x in sys.version_info[:3]] ) for op, want_ver in wanted.items(): _compare_versions(__UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) return # check if any version is installed try: snake_case_ = importlib.metadata.version(__UpperCAmelCase ) except importlib.metadata.PackageNotFoundError: raise importlib.metadata.PackageNotFoundError( F"The '{requirement}' distribution was not found and is required by this application. {hint}" ) # check that the right version is installed if version number or a range was provided if want_ver is not None: for op, want_ver in wanted.items(): _compare_versions(__UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> Dict: '''simple docstring''' snake_case_ = '''Try: pip install transformers -U or pip install -e \'.[dev]\' if you\'re working with git main''' return require_version(__UpperCAmelCase, __UpperCAmelCase )
56
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() a : Any = logging.get_logger(__name__) def __magic_name__ ( __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' snake_case_ = DPTConfig() if "large" in checkpoint_url: snake_case_ = 1024 snake_case_ = 4096 snake_case_ = 24 snake_case_ = 16 snake_case_ = [5, 11, 17, 23] snake_case_ = [256, 512, 1024, 1024] snake_case_ = (1, 384, 384) if "ade" in checkpoint_url: snake_case_ = True snake_case_ = 150 snake_case_ = '''huggingface/label-files''' snake_case_ = '''ade20k-id2label.json''' snake_case_ = json.load(open(cached_download(hf_hub_url(__UpperCAmelCase, __UpperCAmelCase, repo_type='''dataset''' ) ), '''r''' ) ) snake_case_ = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} snake_case_ = [1, 150, 480, 480] return config, expected_shape def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' snake_case_ = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(__UpperCAmelCase, __UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): snake_case_ = name.replace('''pretrained.model''', '''dpt.encoder''' ) if "pretrained.model" in name: snake_case_ = name.replace('''pretrained.model''', '''dpt.embeddings''' ) if "patch_embed" in name: snake_case_ = name.replace('''patch_embed''', '''patch_embeddings''' ) if "pos_embed" in name: snake_case_ = name.replace('''pos_embed''', '''position_embeddings''' ) if "attn.proj" in name: snake_case_ = name.replace('''attn.proj''', '''attention.output.dense''' ) if "proj" in name and "project" not in name: snake_case_ = name.replace('''proj''', '''projection''' ) if "blocks" in name: snake_case_ = name.replace('''blocks''', '''layer''' ) if "mlp.fc1" in name: snake_case_ = name.replace('''mlp.fc1''', '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case_ = name.replace('''mlp.fc2''', '''output.dense''' ) if "norm1" in name: snake_case_ = name.replace('''norm1''', '''layernorm_before''' ) if "norm2" in name: snake_case_ = name.replace('''norm2''', '''layernorm_after''' ) if "scratch.output_conv" in name: snake_case_ = name.replace('''scratch.output_conv''', '''head''' ) if "scratch" in name: snake_case_ = name.replace('''scratch''', '''neck''' ) if "layer1_rn" in name: snake_case_ = name.replace('''layer1_rn''', '''convs.0''' ) if "layer2_rn" in name: snake_case_ = name.replace('''layer2_rn''', '''convs.1''' ) if "layer3_rn" in name: snake_case_ = name.replace('''layer3_rn''', '''convs.2''' ) if "layer4_rn" in name: snake_case_ = name.replace('''layer4_rn''', '''convs.3''' ) if "refinenet" in name: snake_case_ = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 snake_case_ = name.replace(F"refinenet{layer_idx}", F"fusion_stage.layers.{abs(layer_idx-4 )}" ) if "out_conv" in name: snake_case_ = name.replace('''out_conv''', '''projection''' ) if "resConfUnit1" in name: snake_case_ = name.replace('''resConfUnit1''', '''residual_layer1''' ) if "resConfUnit2" in name: snake_case_ = name.replace('''resConfUnit2''', '''residual_layer2''' ) if "conv1" in name: snake_case_ = name.replace('''conv1''', '''convolution1''' ) if "conv2" in name: snake_case_ = name.replace('''conv2''', '''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: snake_case_ = name.replace('''pretrained.act_postprocess1.0.project.0''', '''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: snake_case_ = name.replace('''pretrained.act_postprocess2.0.project.0''', '''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: snake_case_ = name.replace('''pretrained.act_postprocess3.0.project.0''', '''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: snake_case_ = name.replace('''pretrained.act_postprocess4.0.project.0''', '''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: snake_case_ = name.replace('''pretrained.act_postprocess1.3''', '''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: snake_case_ = name.replace('''pretrained.act_postprocess1.4''', '''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: snake_case_ = name.replace('''pretrained.act_postprocess2.3''', '''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: snake_case_ = name.replace('''pretrained.act_postprocess2.4''', '''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: snake_case_ = name.replace('''pretrained.act_postprocess3.3''', '''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: snake_case_ = name.replace('''pretrained.act_postprocess4.3''', '''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: snake_case_ = name.replace('''pretrained.act_postprocess4.4''', '''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: snake_case_ = name.replace('''pretrained''', '''dpt''' ) if "bn" in name: snake_case_ = name.replace('''bn''', '''batch_norm''' ) if "head" in name: snake_case_ = name.replace('''head''', '''head.head''' ) if "encoder.norm" in name: snake_case_ = name.replace('''encoder.norm''', '''layernorm''' ) if "auxlayer" in name: snake_case_ = name.replace('''auxlayer''', '''auxiliary_head.head''' ) return name def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> Dict: '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case_ = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.weight" ) snake_case_ = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict snake_case_ = in_proj_weight[: config.hidden_size, :] snake_case_ = in_proj_bias[: config.hidden_size] snake_case_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case_ = in_proj_weight[ -config.hidden_size :, : ] snake_case_ = in_proj_bias[-config.hidden_size :] def __magic_name__ ( ) -> Any: '''simple docstring''' snake_case_ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case_ = Image.open(requests.get(__UpperCAmelCase, stream=__UpperCAmelCase ).raw ) return im @torch.no_grad() def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' snake_case_ ,snake_case_ = get_dpt_config(__UpperCAmelCase ) # load original state_dict from URL snake_case_ = torch.hub.load_state_dict_from_url(__UpperCAmelCase, map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(__UpperCAmelCase ) # rename keys for key in state_dict.copy().keys(): snake_case_ = state_dict.pop(__UpperCAmelCase ) snake_case_ = val # read in qkv matrices read_in_q_k_v(__UpperCAmelCase, __UpperCAmelCase ) # load HuggingFace model snake_case_ = DPTForSemanticSegmentation(__UpperCAmelCase ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(__UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) model.eval() # Check outputs on an image snake_case_ = 480 if '''ade''' in checkpoint_url else 384 snake_case_ = DPTImageProcessor(size=__UpperCAmelCase ) snake_case_ = prepare_img() snake_case_ = image_processor(__UpperCAmelCase, return_tensors='''pt''' ) # forward pass snake_case_ = model(**__UpperCAmelCase ).logits if '''ade''' in checkpoint_url else model(**__UpperCAmelCase ).predicted_depth # Assert logits snake_case_ = torch.tensor([[6.3_1_9_9, 6.3_6_2_9, 6.4_1_4_8], [6.3_8_5_0, 6.3_6_1_5, 6.4_1_6_6], [6.3_5_1_9, 6.3_1_7_6, 6.3_5_7_5]] ) if "ade" in checkpoint_url: snake_case_ = torch.tensor([[4.0_4_8_0, 4.2_4_2_0, 4.4_3_6_0], [4.3_1_2_4, 4.5_6_9_3, 4.8_2_6_1], [4.5_7_6_8, 4.8_9_6_5, 5.2_1_6_3]] ) assert outputs.shape == torch.Size(__UpperCAmelCase ) assert ( torch.allclose(outputs[0, 0, :3, :3], __UpperCAmelCase, atol=1e-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3], __UpperCAmelCase ) ) Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__UpperCAmelCase ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__UpperCAmelCase ) if push_to_hub: print('''Pushing model to hub...''' ) model.push_to_hub( repo_path_or_name=Path(__UpperCAmelCase, __UpperCAmelCase ), organization='''nielsr''', commit_message='''Add model''', use_temp_dir=__UpperCAmelCase, ) image_processor.push_to_hub( repo_path_or_name=Path(__UpperCAmelCase, __UpperCAmelCase ), organization='''nielsr''', commit_message='''Add image processor''', use_temp_dir=__UpperCAmelCase, ) if __name__ == "__main__": a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt', type=str, help='URL of the original DPT checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', ) parser.add_argument( '--model_name', default='dpt-large', type=str, help='Name of the model, in case you\'re pushing to the hub.', ) a : List[Any] = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
56
1
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert import BertTokenizer UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} UpperCamelCase = { "vocab_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } UpperCamelCase = { "vocab_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } UpperCamelCase = { "vocab_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json" ), }, } UpperCamelCase = { "facebook/dpr-ctx_encoder-single-nq-base": 512, "facebook/dpr-ctx_encoder-multiset-base": 512, } UpperCamelCase = { "facebook/dpr-question_encoder-single-nq-base": 512, "facebook/dpr-question_encoder-multiset-base": 512, } UpperCamelCase = { "facebook/dpr-reader-single-nq-base": 512, "facebook/dpr-reader-multiset-base": 512, } UpperCamelCase = { "facebook/dpr-ctx_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-ctx_encoder-multiset-base": {"do_lower_case": True}, } UpperCamelCase = { "facebook/dpr-question_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-question_encoder-multiset-base": {"do_lower_case": True}, } UpperCamelCase = { "facebook/dpr-reader-single-nq-base": {"do_lower_case": True}, "facebook/dpr-reader-multiset-base": {"do_lower_case": True}, } class snake_case_ ( _UpperCAmelCase ): __A : Any = VOCAB_FILES_NAMES __A : Optional[Any] = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP __A : str = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A : int = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION class snake_case_ ( _UpperCAmelCase ): __A : int = VOCAB_FILES_NAMES __A : Dict = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP __A : Union[str, Any] = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A : Optional[int] = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION UpperCamelCase = collections.namedtuple( '''DPRSpanPrediction''', ['''span_score''', '''relevance_score''', '''doc_id''', '''start_index''', '''end_index''', '''text'''] ) UpperCamelCase = collections.namedtuple('''DPRReaderOutput''', ['''start_logits''', '''end_logits''', '''relevance_logits''']) UpperCamelCase = r"\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n ```\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n ```\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `'tf'`: Return TensorFlow `tf.constant` objects.\n - `'pt'`: Return PyTorch `torch.Tensor` objects.\n - `'np'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer's default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Returns:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n " @add_start_docstrings(_UpperCAmelCase ) class snake_case_ : def __call__( self : int , lowercase_ : int , lowercase_ : Optional[str] = None , lowercase_ : Optional[str] = None , lowercase_ : Union[bool, str] = False , lowercase_ : Union[bool, str] = False , lowercase_ : Optional[int] = None , lowercase_ : Optional[Union[str, TensorType]] = None , lowercase_ : Optional[bool] = None , **lowercase_ : int , ) -> Any: if titles is None and texts is None: return super().__call__( lowercase_ , padding=lowercase_ , truncation=lowercase_ , max_length=lowercase_ , return_tensors=lowercase_ , return_attention_mask=lowercase_ , **lowercase_ , ) elif titles is None or texts is None: lowercase__ : Tuple = titles if texts is None else texts return super().__call__( lowercase_ , lowercase_ , padding=lowercase_ , truncation=lowercase_ , max_length=lowercase_ , return_tensors=lowercase_ , return_attention_mask=lowercase_ , **lowercase_ , ) lowercase__ : List[Any] = titles if not isinstance(lowercase_ , lowercase_ ) else [titles] lowercase__ : Dict = texts if not isinstance(lowercase_ , lowercase_ ) else [texts] lowercase__ : List[Any] = len(lowercase_ ) lowercase__ : Optional[Any] = questions if not isinstance(lowercase_ , lowercase_ ) else [questions] * n_passages if len(lowercase_ ) != len(lowercase_ ): raise ValueError( F'''There should be as many titles than texts but got {len(lowercase_ )} titles and {len(lowercase_ )} texts.''' ) lowercase__ : Any = super().__call__(lowercase_ , lowercase_ , padding=lowercase_ , truncation=lowercase_ )["""input_ids"""] lowercase__ : str = super().__call__(lowercase_ , add_special_tokens=lowercase_ , padding=lowercase_ , truncation=lowercase_ )["""input_ids"""] lowercase__ : List[Any] = { """input_ids""": [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(lowercase_ , lowercase_ ) ] } if return_attention_mask is not False: lowercase__ : int = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) lowercase__ : int = attention_mask return self.pad(lowercase_ , padding=lowercase_ , max_length=lowercase_ , return_tensors=lowercase_ ) def __UpperCamelCase ( self : Union[str, Any] , lowercase_ : BatchEncoding , lowercase_ : DPRReaderOutput , lowercase_ : int = 16 , lowercase_ : int = 64 , lowercase_ : int = 4 , ) -> Optional[Any]: lowercase__ : Optional[int] = reader_input["""input_ids"""] lowercase__ : Optional[int] = reader_output[:3] lowercase__ : Union[str, Any] = len(lowercase_ ) lowercase__ : List[Any] = sorted(range(lowercase_ ) , reverse=lowercase_ , key=relevance_logits.__getitem__ ) lowercase__ : List[DPRReaderOutput] = [] for doc_id in sorted_docs: lowercase__ : Optional[Any] = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence lowercase__ : Any = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: lowercase__ : Any = sequence_ids.index(self.pad_token_id ) else: lowercase__ : Dict = len(lowercase_ ) lowercase__ : Any = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=lowercase_ , top_spans=lowercase_ , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=lowercase_ , start_index=lowercase_ , end_index=lowercase_ , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(lowercase_ ) >= num_spans: break return nbest_spans_predictions[:num_spans] def __UpperCamelCase ( self : Dict , lowercase_ : List[int] , lowercase_ : List[int] , lowercase_ : int , lowercase_ : int , ) -> Dict: lowercase__ : Tuple = [] for start_index, start_score in enumerate(lowercase_ ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) lowercase__ : str = sorted(lowercase_ , key=lambda lowercase_ : x[1] , reverse=lowercase_ ) lowercase__ : Dict = [] for (start_index, end_index), score in scores: if start_index > end_index: raise ValueError(F'''Wrong span indices: [{start_index}:{end_index}]''' ) lowercase__ : Optional[int] = end_index - start_index + 1 if length > max_answer_length: raise ValueError(F'''Span is too long: {length} > {max_answer_length}''' ) if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(lowercase_ ) == top_spans: break return chosen_span_intervals @add_end_docstrings(_UpperCAmelCase ) class snake_case_ ( _UpperCAmelCase ,_UpperCAmelCase ): __A : Any = VOCAB_FILES_NAMES __A : Optional[int] = READER_PRETRAINED_VOCAB_FILES_MAP __A : Any = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A : List[Any] = READER_PRETRAINED_INIT_CONFIGURATION __A : Any = ["input_ids", "attention_mask"]
371
import argparse import datetime def lowercase_ ( _lowerCamelCase : str): lowercase__ : Optional[Any] = { "0": "Sunday", "1": "Monday", "2": "Tuesday", "3": "Wednesday", "4": "Thursday", "5": "Friday", "6": "Saturday", } lowercase__ : Any = {0: 1, 1: 2, 2: 3, 3: 4, 4: 5, 5: 6, 6: 0} # Validate if not 0 < len(_lowerCamelCase) < 11: raise ValueError("Must be 10 characters long") # Get month lowercase__ : int = int(date_input[0] + date_input[1]) # Validate if not 0 < m < 13: raise ValueError("Month must be between 1 - 12") lowercase__ : str = date_input[2] # Validate if sep_a not in ["-", "/"]: raise ValueError("Date separator must be '-' or '/'") # Get day lowercase__ : int = int(date_input[3] + date_input[4]) # Validate if not 0 < d < 32: raise ValueError("Date must be between 1 - 31") # Get second separator lowercase__ : str = date_input[5] # Validate if sep_a not in ["-", "/"]: raise ValueError("Date separator must be '-' or '/'") # Get year lowercase__ : int = int(date_input[6] + date_input[7] + date_input[8] + date_input[9]) # Arbitrary year range if not 45 < y < 8500: raise ValueError( "Year out of range. There has to be some sort of limit...right?") # Get datetime obj for validation lowercase__ : Union[str, Any] = datetime.date(int(_lowerCamelCase) , int(_lowerCamelCase) , int(_lowerCamelCase)) # Start math if m <= 2: lowercase__ : Optional[Any] = y - 1 lowercase__ : int = m + 12 # maths var lowercase__ : int = int(str(_lowerCamelCase)[:2]) lowercase__ : int = int(str(_lowerCamelCase)[2:]) lowercase__ : int = int(2.6 * m - 5.39) lowercase__ : int = int(c / 4) lowercase__ : int = int(k / 4) lowercase__ : int = int(d + k) lowercase__ : int = int(t + u + v + x) lowercase__ : int = int(z - (2 * c)) lowercase__ : int = round(w % 7) # End math # Validate math if f != convert_datetime_days[dt_ck.weekday()]: raise AssertionError("The date was evaluated incorrectly. Contact developer.") # Response lowercase__ : str = f'''Your date {date_input}, is a {days[str(_lowerCamelCase)]}!''' return response if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = argparse.ArgumentParser( description=( '''Find out what day of the week nearly any date is or was. Enter ''' '''date as a string in the mm-dd-yyyy or mm/dd/yyyy format''' ) ) parser.add_argument( '''date_input''', type=str, help='''Date as a string (mm-dd-yyyy or mm/dd/yyyy)''' ) UpperCamelCase = parser.parse_args() zeller(args.date_input)
333
0
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __a = logging.get_logger(__name__) __a = { 'vocab_file': 'vocab.txt', 'merges_file': 'bpe.codes', } __a = { 'vocab_file': { 'vinai/phobert-base': 'https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt', 'vinai/phobert-large': 'https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt', }, 'merges_file': { 'vinai/phobert-base': 'https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes', 'vinai/phobert-large': 'https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes', }, } __a = { 'vinai/phobert-base': 2_5_6, 'vinai/phobert-large': 2_5_6, } def a ( snake_case__: List[str] ): '''simple docstring''' lowercase_ = set() lowercase_ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase_ = char lowercase_ = set(snake_case__ ) return pairs class lowercase__( UpperCAmelCase ): """simple docstring""" a :Optional[Any] = VOCAB_FILES_NAMES a :List[str] = PRETRAINED_VOCAB_FILES_MAP a :Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : int , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Union[str, Any]="<s>" , SCREAMING_SNAKE_CASE_ : Union[str, Any]="</s>" , SCREAMING_SNAKE_CASE_ : List[Any]="</s>" , SCREAMING_SNAKE_CASE_ : Optional[Any]="<s>" , SCREAMING_SNAKE_CASE_ : List[str]="<unk>" , SCREAMING_SNAKE_CASE_ : List[Any]="<pad>" , SCREAMING_SNAKE_CASE_ : List[Any]="<mask>" , **SCREAMING_SNAKE_CASE_ : Optional[Any] , ) -> Union[str, Any]: super().__init__( bos_token=SCREAMING_SNAKE_CASE_ , eos_token=SCREAMING_SNAKE_CASE_ , unk_token=SCREAMING_SNAKE_CASE_ , sep_token=SCREAMING_SNAKE_CASE_ , cls_token=SCREAMING_SNAKE_CASE_ , pad_token=SCREAMING_SNAKE_CASE_ , mask_token=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ , ) lowercase_ = vocab_file lowercase_ = merges_file lowercase_ = {} lowercase_ = 0 lowercase_ = 1 lowercase_ = 2 lowercase_ = 3 self.add_from_file(SCREAMING_SNAKE_CASE_ ) lowercase_ = {v: k for k, v in self.encoder.items()} with open(SCREAMING_SNAKE_CASE_ , encoding='''utf-8''' ) as merges_handle: lowercase_ = merges_handle.read().split('''\n''' )[:-1] lowercase_ = [tuple(merge.split()[:-1] ) for merge in merges] lowercase_ = dict(zip(SCREAMING_SNAKE_CASE_ , range(len(SCREAMING_SNAKE_CASE_ ) ) ) ) lowercase_ = {} def _lowercase ( self : Any , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase_ = [self.cls_token_id] lowercase_ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _lowercase ( self : List[Any] , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE_ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=SCREAMING_SNAKE_CASE_ , token_ids_a=SCREAMING_SNAKE_CASE_ , already_has_special_tokens=SCREAMING_SNAKE_CASE_ ) if token_ids_a is None: return [1] + ([0] * len(SCREAMING_SNAKE_CASE_ )) + [1] return [1] + ([0] * len(SCREAMING_SNAKE_CASE_ )) + [1, 1] + ([0] * len(SCREAMING_SNAKE_CASE_ )) + [1] def _lowercase ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None ) -> List[int]: lowercase_ = [self.sep_token_id] lowercase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def _lowercase ( self : Any ) -> Any: return len(self.encoder ) def _lowercase ( self : Any ) -> Dict: return dict(self.encoder , **self.added_tokens_encoder ) def _lowercase ( self : Dict , SCREAMING_SNAKE_CASE_ : Any ) -> Any: if token in self.cache: return self.cache[token] lowercase_ = tuple(SCREAMING_SNAKE_CASE_ ) lowercase_ = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) lowercase_ = get_pairs(SCREAMING_SNAKE_CASE_ ) if not pairs: return token while True: lowercase_ = min(SCREAMING_SNAKE_CASE_ , key=lambda SCREAMING_SNAKE_CASE_ : self.bpe_ranks.get(SCREAMING_SNAKE_CASE_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break lowercase_ , lowercase_ = bigram lowercase_ = [] lowercase_ = 0 while i < len(SCREAMING_SNAKE_CASE_ ): try: lowercase_ = word.index(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase_ = j if word[i] == first and i < len(SCREAMING_SNAKE_CASE_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase_ = tuple(SCREAMING_SNAKE_CASE_ ) lowercase_ = new_word if len(SCREAMING_SNAKE_CASE_ ) == 1: break else: lowercase_ = get_pairs(SCREAMING_SNAKE_CASE_ ) lowercase_ = '''@@ '''.join(SCREAMING_SNAKE_CASE_ ) lowercase_ = word[:-4] lowercase_ = word return word def _lowercase ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : Dict ) -> int: lowercase_ = [] lowercase_ = re.findall(R'''\S+\n?''' , SCREAMING_SNAKE_CASE_ ) for token in words: split_tokens.extend(list(self.bpe(SCREAMING_SNAKE_CASE_ ).split(''' ''' ) ) ) return split_tokens def _lowercase ( self : List[str] , SCREAMING_SNAKE_CASE_ : List[Any] ) -> Union[str, Any]: return self.encoder.get(SCREAMING_SNAKE_CASE_ , self.encoder.get(self.unk_token ) ) def _lowercase ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> List[Any]: return self.decoder.get(SCREAMING_SNAKE_CASE_ , self.unk_token ) def _lowercase ( self : Any , SCREAMING_SNAKE_CASE_ : Any ) -> Tuple: lowercase_ = ''' '''.join(SCREAMING_SNAKE_CASE_ ).replace('''@@ ''' , '''''' ).strip() return out_string def _lowercase ( self : List[str] , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase_ = os.path.join( SCREAMING_SNAKE_CASE_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase_ = os.path.join( SCREAMING_SNAKE_CASE_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(SCREAMING_SNAKE_CASE_ ): copyfile(self.vocab_file , SCREAMING_SNAKE_CASE_ ) if os.path.abspath(self.merges_file ) != os.path.abspath(SCREAMING_SNAKE_CASE_ ): copyfile(self.merges_file , SCREAMING_SNAKE_CASE_ ) return out_vocab_file, out_merge_file def _lowercase ( self : List[str] , SCREAMING_SNAKE_CASE_ : List[str] ) -> Dict: if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): try: with open(SCREAMING_SNAKE_CASE_ , '''r''' , encoding='''utf-8''' ) as fd: self.add_from_file(SCREAMING_SNAKE_CASE_ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(f'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return lowercase_ = f.readlines() for lineTmp in lines: lowercase_ = lineTmp.strip() lowercase_ = line.rfind(''' ''' ) if idx == -1: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt>\'''' ) lowercase_ = line[:idx] lowercase_ = len(self.encoder )
30
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 1000 ) -> int: lowerCamelCase__ : str = -1 lowerCamelCase__ : Dict = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowerCamelCase__ : Dict = (n * n - 2 * a * n) // (2 * n - 2 * a) lowerCamelCase__ : Any = n - a - b if c * c == (a * a + b * b): lowerCamelCase__ : Dict = a * b * c if candidate >= product: lowerCamelCase__ : Union[str, Any] = candidate return product if __name__ == "__main__": print(F'{solution() = }')
41
0
'''simple docstring''' import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging from ..processors.glue import glue_convert_examples_to_features, glue_output_modes, glue_processors from ..processors.utils import InputFeatures _lowercase : int = logging.get_logger(__name__) @dataclass class __magic_name__ : UpperCamelCase__ = field(metadata={'''help''': '''The name of the task to train on: ''' + ''', '''.join(glue_processors.keys() )} ) UpperCamelCase__ = field( metadata={'''help''': '''The input data dir. Should contain the .tsv files (or other data files) for the task.'''} ) UpperCamelCase__ = field( default=128, metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) }, ) UpperCamelCase__ = field( default=_UpperCAmelCase, metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def SCREAMING_SNAKE_CASE_ ( self : Dict ): lowercase_ : Dict = self.task_name.lower() class __magic_name__ ( _UpperCAmelCase ): UpperCamelCase__ = '''train''' UpperCamelCase__ = '''dev''' UpperCamelCase__ = '''test''' class __magic_name__ ( _UpperCAmelCase ): UpperCamelCase__ = 42 UpperCamelCase__ = 42 UpperCamelCase__ = 42 def __init__( self : str , lowercase_ : GlueDataTrainingArguments , lowercase_ : PreTrainedTokenizerBase , lowercase_ : Optional[int] = None , lowercase_ : Union[str, Split] = Split.train , lowercase_ : Optional[str] = None , ): warnings.warn( """This dataset will be removed from the library soon, preprocessing should be handled with the 🤗 Datasets """ """library. You can have a look at this example script for pointers: """ """https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py""" , lowercase_ , ) lowercase_ : Union[str, Any] = args lowercase_ : List[Any] = glue_processors[args.task_name]() lowercase_ : Optional[Any] = glue_output_modes[args.task_name] if isinstance(lowercase_ , lowercase_ ): try: lowercase_ : List[Any] = Split[mode] except KeyError: raise KeyError("""mode is not a valid split name""" ) # Load data features from cache or dataset file lowercase_ : str = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{args.task_name}''' , ) lowercase_ : Tuple = self.processor.get_labels() if args.task_name in ["mnli", "mnli-mm"] and tokenizer.__class__.__name__ in ( "RobertaTokenizer", "RobertaTokenizerFast", "XLMRobertaTokenizer", "BartTokenizer", "BartTokenizerFast", ): # HACK(label indices are swapped in RoBERTa pretrained model) lowercase_ : List[str] = label_list[2], label_list[1] lowercase_ : Union[str, Any] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowercase_ : List[Any] = cached_features_file + """.lock""" with FileLock(lowercase_ ): if os.path.exists(lowercase_ ) and not args.overwrite_cache: lowercase_ : Optional[Any] = time.time() lowercase_ : Optional[int] = torch.load(lowercase_ ) logger.info( f'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) else: logger.info(f'''Creating features from dataset file at {args.data_dir}''' ) if mode == Split.dev: lowercase_ : Union[str, Any] = self.processor.get_dev_examples(args.data_dir ) elif mode == Split.test: lowercase_ : Tuple = self.processor.get_test_examples(args.data_dir ) else: lowercase_ : Optional[Any] = self.processor.get_train_examples(args.data_dir ) if limit_length is not None: lowercase_ : Dict = examples[:limit_length] lowercase_ : List[Any] = glue_convert_examples_to_features( lowercase_ , lowercase_ , max_length=args.max_seq_length , label_list=lowercase_ , output_mode=self.output_mode , ) lowercase_ : int = time.time() torch.save(self.features , lowercase_ ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self : List[Any] ): return len(self.features ) def __getitem__( self : int , lowercase_ : Optional[Any] ): return self.features[i] def SCREAMING_SNAKE_CASE_ ( self : int ): return self.label_list
362
'''simple docstring''' def lowerCamelCase ( ) -> Dict: lowercase_ : Union[str, Any] = [] lowercase_ : Tuple = 1 while len(UpperCAmelCase__ ) < 1e6: constant.append(str(UpperCAmelCase__ ) ) i += 1 lowercase_ : int = """""".join(UpperCAmelCase__ ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[99] ) * int(constant[999] ) * int(constant[9999] ) * int(constant[99999] ) * int(constant[999999] ) ) if __name__ == "__main__": print(solution())
21
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class __magic_name__ ( unittest.TestCase): def UpperCAmelCase__ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : Optional[int] = '''ZinengTang/tvlt-base''' UpperCamelCase__ : int = tempfile.mkdtemp() def UpperCAmelCase__ ( self : int , **lowerCamelCase__ : List[str] ) -> List[Any]: '''simple docstring''' return TvltImageProcessor.from_pretrained(self.checkpoint , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : Optional[Any] , **lowerCamelCase__ : Tuple ) -> List[Any]: '''simple docstring''' return TvltFeatureExtractor.from_pretrained(self.checkpoint , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : str ) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCAmelCase__ ( self : Any ) -> int: '''simple docstring''' UpperCamelCase__ : int = self.get_image_processor() UpperCamelCase__ : Union[str, Any] = self.get_feature_extractor() UpperCamelCase__ : List[str] = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase__ : Optional[int] = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , lowerCamelCase__ ) self.assertIsInstance(processor.image_processor , lowerCamelCase__ ) def UpperCAmelCase__ ( self : List[Any] ) -> Tuple: '''simple docstring''' UpperCamelCase__ : str = self.get_image_processor() UpperCamelCase__ : List[Any] = self.get_feature_extractor() UpperCamelCase__ : Dict = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : Any = np.ones([12000] ) UpperCamelCase__ : Union[str, Any] = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ) UpperCamelCase__ : Any = processor(audio=lowerCamelCase__ , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' UpperCamelCase__ : List[Any] = self.get_image_processor() UpperCamelCase__ : Any = self.get_feature_extractor() UpperCamelCase__ : int = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : int = np.ones([3, 224, 224] ) UpperCamelCase__ : List[str] = image_processor(lowerCamelCase__ , return_tensors='''np''' ) UpperCamelCase__ : str = processor(images=lowerCamelCase__ , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : Any = self.get_image_processor() UpperCamelCase__ : Dict = self.get_feature_extractor() UpperCamelCase__ : Union[str, Any] = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : List[str] = np.ones([12000] ) UpperCamelCase__ : Tuple = np.ones([3, 224, 224] ) UpperCamelCase__ : Optional[Any] = processor(audio=lowerCamelCase__ , images=lowerCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(lowerCamelCase__ ): processor() def UpperCAmelCase__ ( self : Dict ) -> int: '''simple docstring''' UpperCamelCase__ : List[str] = self.get_image_processor() UpperCamelCase__ : str = self.get_feature_extractor() UpperCamelCase__ : Tuple = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
146
import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def _a ( SCREAMING_SNAKE_CASE : List[str] ): """simple docstring""" if is_torch_version('''<''' , '''2.0.0''' ) or not hasattr(SCREAMING_SNAKE_CASE , '''_dynamo''' ): return False return isinstance(SCREAMING_SNAKE_CASE , torch._dynamo.eval_frame.OptimizedModule ) def _a ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : bool = True ): """simple docstring""" UpperCamelCase__ : Any = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) UpperCamelCase__ : Optional[int] = is_compiled_module(SCREAMING_SNAKE_CASE ) if is_compiled: UpperCamelCase__ : Optional[int] = model UpperCamelCase__ : Optional[Any] = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): UpperCamelCase__ : Optional[int] = model.module if not keep_fpaa_wrapper: UpperCamelCase__ : Optional[Any] = getattr(SCREAMING_SNAKE_CASE , '''forward''' ) UpperCamelCase__ : Optional[Any] = model.__dict__.pop('''_original_forward''' , SCREAMING_SNAKE_CASE ) if original_forward is not None: while hasattr(SCREAMING_SNAKE_CASE , '''__wrapped__''' ): UpperCamelCase__ : Optional[int] = forward.__wrapped__ if forward == original_forward: break UpperCamelCase__ : Optional[Any] = forward if getattr(SCREAMING_SNAKE_CASE , '''_converted_to_transformer_engine''' , SCREAMING_SNAKE_CASE ): convert_model(SCREAMING_SNAKE_CASE , to_transformer_engine=SCREAMING_SNAKE_CASE ) if is_compiled: UpperCamelCase__ : Tuple = model UpperCamelCase__ : Tuple = compiled_model return model def _a ( ): """simple docstring""" PartialState().wait_for_everyone() def _a ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Dict ): """simple docstring""" if PartialState().distributed_type == DistributedType.TPU: xm.save(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) elif PartialState().local_process_index == 0: torch.save(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) @contextmanager def _a ( **SCREAMING_SNAKE_CASE : Dict ): """simple docstring""" for key, value in kwargs.items(): UpperCamelCase__ : Dict = str(SCREAMING_SNAKE_CASE ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def _a ( SCREAMING_SNAKE_CASE : List[Any] ): """simple docstring""" if not hasattr(SCREAMING_SNAKE_CASE , '''__qualname__''' ) and not hasattr(SCREAMING_SNAKE_CASE , '''__name__''' ): UpperCamelCase__ : str = getattr(SCREAMING_SNAKE_CASE , '''__class__''' , SCREAMING_SNAKE_CASE ) if hasattr(SCREAMING_SNAKE_CASE , '''__qualname__''' ): return obj.__qualname__ if hasattr(SCREAMING_SNAKE_CASE , '''__name__''' ): return obj.__name__ return str(SCREAMING_SNAKE_CASE ) def _a ( SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Any ): """simple docstring""" for key, value in source.items(): if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): UpperCamelCase__ : Optional[Any] = destination.setdefault(SCREAMING_SNAKE_CASE , {} ) merge_dicts(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else: UpperCamelCase__ : List[Any] = value return destination def _a ( SCREAMING_SNAKE_CASE : int = None ): """simple docstring""" if port is None: UpperCamelCase__ : Union[str, Any] = 29500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(('''localhost''', port) ) == 0
146
1
'''simple docstring''' import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class __magic_name__ ( _UpperCamelCase ): def __lowercase ( self : Dict ): _a : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_UpperCAmelCase ,'tf_padding' ) ) self.parent.assertTrue(hasattr(_UpperCAmelCase ,'depth_multiplier' ) ) class __magic_name__ : def __init__( self : Tuple ,_UpperCAmelCase : Tuple ,_UpperCAmelCase : List[str]=13 ,_UpperCAmelCase : str=3 ,_UpperCAmelCase : List[str]=32 ,_UpperCAmelCase : Tuple=0.25 ,_UpperCAmelCase : List[str]=8 ,_UpperCAmelCase : int=8 ,_UpperCAmelCase : str=6 ,_UpperCAmelCase : List[Any]=32 ,_UpperCAmelCase : int=True ,_UpperCAmelCase : List[str]=True ,_UpperCAmelCase : List[Any]=True ,_UpperCAmelCase : Dict="relu6" ,_UpperCAmelCase : Optional[int]=1280 ,_UpperCAmelCase : str=0.1 ,_UpperCAmelCase : Any=0.02 ,_UpperCAmelCase : int=True ,_UpperCAmelCase : Optional[Any]=True ,_UpperCAmelCase : int=10 ,_UpperCAmelCase : List[str]=None ,): _a : List[str] = parent _a : Dict = batch_size _a : Union[str, Any] = num_channels _a : Dict = image_size _a : Optional[int] = depth_multiplier _a : int = depth_divisible_by _a : Union[str, Any] = min_depth _a : str = expand_ratio _a : List[str] = tf_padding _a : List[Any] = output_stride _a : Optional[Any] = first_layer_is_expansion _a : Union[str, Any] = finegrained_output _a : Union[str, Any] = hidden_act _a : Optional[int] = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) _a : str = classifier_dropout_prob _a : Dict = use_labels _a : Union[str, Any] = is_training _a : Any = num_labels _a : Union[str, Any] = initializer_range _a : List[str] = scope def __lowercase ( self : List[Any] ): _a : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _a : Optional[int] = None _a : List[Any] = None if self.use_labels: _a : str = ids_tensor([self.batch_size] ,self.num_labels ) _a : Optional[int] = ids_tensor([self.batch_size, self.image_size, self.image_size] ,self.num_labels ) _a : str = self.get_config() return config, pixel_values, labels, pixel_labels def __lowercase ( self : Dict ): return MobileNetVaConfig( num_channels=self.num_channels ,image_size=self.image_size ,depth_multiplier=self.depth_multiplier ,depth_divisible_by=self.depth_divisible_by ,min_depth=self.min_depth ,expand_ratio=self.expand_ratio ,output_stride=self.output_stride ,first_layer_is_expansion=self.first_layer_is_expansion ,finegrained_output=self.finegrained_output ,hidden_act=self.hidden_act ,tf_padding=self.tf_padding ,classifier_dropout_prob=self.classifier_dropout_prob ,initializer_range=self.initializer_range ,) def __lowercase ( self : Any ,_UpperCAmelCase : List[Any] ,_UpperCAmelCase : Union[str, Any] ,_UpperCAmelCase : str ,_UpperCAmelCase : int ): _a : Tuple = MobileNetVaModel(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _a : Dict = model(_UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape ,( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) ,) self.parent.assertEqual( result.pooler_output.shape ,(self.batch_size, self.last_hidden_size) ,) def __lowercase ( self : str ,_UpperCAmelCase : str ,_UpperCAmelCase : Any ,_UpperCAmelCase : Optional[Any] ,_UpperCAmelCase : Optional[Any] ): _a : str = self.num_labels _a : int = MobileNetVaForImageClassification(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _a : Optional[Any] = model(_UpperCAmelCase ,labels=_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def __lowercase ( self : Any ,_UpperCAmelCase : str ,_UpperCAmelCase : int ,_UpperCAmelCase : Union[str, Any] ,_UpperCAmelCase : int ): _a : Tuple = self.num_labels _a : Optional[int] = MobileNetVaForSemanticSegmentation(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _a : Tuple = model(_UpperCAmelCase ) self.parent.assertEqual( result.logits.shape ,( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) ,) _a : List[Any] = model(_UpperCAmelCase ,labels=_UpperCAmelCase ) self.parent.assertEqual( result.logits.shape ,( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) ,) def __lowercase ( self : Any ): _a : str = self.prepare_config_and_inputs() _a , _a , _a , _a : List[Any] = config_and_inputs _a : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __magic_name__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): lowerCAmelCase : List[str] = ( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) lowerCAmelCase : Tuple = ( { 'feature-extraction': MobileNetVaModel, 'image-classification': MobileNetVaForImageClassification, 'image-segmentation': MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) lowerCAmelCase : List[str] = False lowerCAmelCase : Tuple = False lowerCAmelCase : List[Any] = False lowerCAmelCase : List[Any] = False def __lowercase ( self : List[Any] ): _a : List[str] = MobileNetVaModelTester(self ) _a : Dict = MobileNetVaConfigTester(self ,config_class=_UpperCAmelCase ,has_text_modality=_UpperCAmelCase ) def __lowercase ( self : Optional[int] ): self.config_tester.run_common_tests() @unittest.skip(reason='MobileNetV2 does not use inputs_embeds' ) def __lowercase ( self : str ): pass @unittest.skip(reason='MobileNetV2 does not support input and output embeddings' ) def __lowercase ( self : Union[str, Any] ): pass @unittest.skip(reason='MobileNetV2 does not output attentions' ) def __lowercase ( self : Union[str, Any] ): pass def __lowercase ( self : List[Any] ): _a , _a : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(_UpperCAmelCase ) _a : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : int = [*signature.parameters.keys()] _a : int = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_UpperCAmelCase ) def __lowercase ( self : int ): _a : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCAmelCase ) def __lowercase ( self : int ): def check_hidden_states_output(_UpperCAmelCase : Dict ,_UpperCAmelCase : int ,_UpperCAmelCase : Optional[Any] ): _a : List[Any] = model_class(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() with torch.no_grad(): _a : Dict = model(**self._prepare_for_class(_UpperCAmelCase ,_UpperCAmelCase ) ) _a : List[Any] = outputs.hidden_states _a : List[str] = 16 self.assertEqual(len(_UpperCAmelCase ) ,_UpperCAmelCase ) _a , _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Tuple = True check_hidden_states_output(_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _a : List[Any] = True check_hidden_states_output(_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) def __lowercase ( self : str ): _a : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_UpperCAmelCase ) def __lowercase ( self : Optional[int] ): _a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_UpperCAmelCase ) @slow def __lowercase ( self : Optional[Any] ): for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : List[str] = MobileNetVaModel.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) def __lowerCamelCase ( ) -> Optional[int]: _a : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class __magic_name__ ( unittest.TestCase ): @cached_property def __lowercase ( self : List[Any] ): return ( MobileNetVaImageProcessor.from_pretrained('google/mobilenet_v2_1.0_224' ) if is_vision_available() else None ) @slow def __lowercase ( self : str ): _a : Tuple = MobileNetVaForImageClassification.from_pretrained('google/mobilenet_v2_1.0_224' ).to(_UpperCAmelCase ) _a : str = self.default_image_processor _a : Optional[int] = prepare_img() _a : List[str] = image_processor(images=_UpperCAmelCase ,return_tensors='pt' ).to(_UpperCAmelCase ) # forward pass with torch.no_grad(): _a : List[str] = model(**_UpperCAmelCase ) # verify the logits _a : Tuple = torch.Size((1, 1001) ) self.assertEqual(outputs.logits.shape ,_UpperCAmelCase ) _a : Tuple = torch.tensor([0.24_45, -1.19_93, 0.19_05] ).to(_UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,_UpperCAmelCase ,atol=1E-4 ) ) @slow def __lowercase ( self : Union[str, Any] ): _a : List[str] = MobileNetVaForSemanticSegmentation.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) _a : Optional[Any] = model.to(_UpperCAmelCase ) _a : Optional[int] = MobileNetVaImageProcessor.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) _a : int = prepare_img() _a : int = image_processor(images=_UpperCAmelCase ,return_tensors='pt' ).to(_UpperCAmelCase ) # forward pass with torch.no_grad(): _a : Union[str, Any] = model(**_UpperCAmelCase ) _a : Any = outputs.logits # verify the logits _a : str = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape ,_UpperCAmelCase ) _a : List[str] = torch.tensor( [ [[17.57_90, 17.75_81, 18.33_55], [18.32_57, 18.42_30, 18.89_73], [18.61_69, 18.86_50, 19.21_87]], [[-2.15_95, -2.09_77, -2.37_41], [-2.42_26, -2.30_28, -2.68_35], [-2.78_19, -2.59_91, -2.77_06]], [[4.20_58, 4.83_17, 4.76_38], [4.41_36, 5.03_61, 4.93_83], [4.50_28, 4.96_44, 4.87_34]], ] ,device=_UpperCAmelCase ,) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] ,_UpperCAmelCase ,atol=1E-4 ) )
107
'''simple docstring''' import heapq def __lowerCamelCase ( lowerCAmelCase_ ) -> set[int]: _a : list[list] = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(lowerCAmelCase_ , [-1 * len(lowerCAmelCase_ ), (key, value)] ) # chosen_vertices = set of chosen vertices _a : str = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices _a : Optional[int] = heapq.heappop(lowerCAmelCase_ )[1][0] chosen_vertices.add(lowerCAmelCase_ ) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: _a : Dict = elem[1][1].index(lowerCAmelCase_ ) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(lowerCAmelCase_ ) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() __lowerCAmelCase = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(f"""Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}""")
107
1
'''simple docstring''' lowercase : Optional[int] = {str(digit): digit**5 for digit in range(10)} def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' return sum(DIGITS_FIFTH_POWER[digit] for digit in str(snake_case__ ) ) def lowerCAmelCase_ ( ): '''simple docstring''' return sum( number for number in range(1000 , 100_0000 ) if number == digits_fifth_powers_sum(snake_case__ ) ) if __name__ == "__main__": print(solution())
3
'''simple docstring''' import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors __lowercase = logging.getLogger(__name__) class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : Any = '''sequence-classification''' def __init__( self , __lowerCAmelCase): """simple docstring""" if type(__lowerCAmelCase) == dict: lowerCAmelCase = Namespace(**__lowerCAmelCase) lowerCAmelCase = glue_output_modes[hparams.task] lowerCAmelCase = glue_tasks_num_labels[hparams.task] super().__init__(__lowerCAmelCase , __lowerCAmelCase , self.mode) def a_ ( self , **__lowerCAmelCase): """simple docstring""" return self.model(**__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: lowerCAmelCase = batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None lowerCAmelCase = self(**__lowerCAmelCase) lowerCAmelCase = outputs[0] lowerCAmelCase = self.trainer.lr_schedulers[0]["""scheduler"""] lowerCAmelCase = {"""loss""": loss, """rate""": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def a_ ( self): """simple docstring""" lowerCAmelCase = self.hparams lowerCAmelCase = processors[args.task]() lowerCAmelCase = processor.get_labels() for mode in ["train", "dev"]: lowerCAmelCase = self._feature_file(__lowerCAmelCase) if os.path.exists(__lowerCAmelCase) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , __lowerCAmelCase) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir) lowerCAmelCase = ( processor.get_dev_examples(args.data_dir) if mode == """dev""" else processor.get_train_examples(args.data_dir) ) lowerCAmelCase = convert_examples_to_features( __lowerCAmelCase , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info("""Saving features into cached file %s""" , __lowerCAmelCase) torch.save(__lowerCAmelCase , __lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = False): """simple docstring""" lowerCAmelCase = """dev""" if mode == """test""" else mode lowerCAmelCase = self._feature_file(__lowerCAmelCase) logger.info("""Loading features from cached file %s""" , __lowerCAmelCase) lowerCAmelCase = torch.load(__lowerCAmelCase) lowerCAmelCase = torch.tensor([f.input_ids for f in features] , dtype=torch.long) lowerCAmelCase = torch.tensor([f.attention_mask for f in features] , dtype=torch.long) lowerCAmelCase = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long) if self.hparams.glue_output_mode == "classification": lowerCAmelCase = torch.tensor([f.label for f in features] , dtype=torch.long) elif self.hparams.glue_output_mode == "regression": lowerCAmelCase = torch.tensor([f.label for f in features] , dtype=torch.float) return DataLoader( TensorDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase) , batch_size=__lowerCAmelCase , shuffle=__lowerCAmelCase , ) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: lowerCAmelCase = batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None lowerCAmelCase = self(**__lowerCAmelCase) lowerCAmelCase , lowerCAmelCase = outputs[:2] lowerCAmelCase = logits.detach().cpu().numpy() lowerCAmelCase = inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = torch.stack([x["""val_loss"""] for x in outputs]).mean().detach().cpu().item() lowerCAmelCase = np.concatenate([x["""pred"""] for x in outputs] , axis=0) if self.hparams.glue_output_mode == "classification": lowerCAmelCase = np.argmax(__lowerCAmelCase , axis=1) elif self.hparams.glue_output_mode == "regression": lowerCAmelCase = np.squeeze(__lowerCAmelCase) lowerCAmelCase = np.concatenate([x["""target"""] for x in outputs] , axis=0) lowerCAmelCase = [[] for _ in range(out_label_ids.shape[0])] lowerCAmelCase = [[] for _ in range(out_label_ids.shape[0])] lowerCAmelCase = {**{"""val_loss""": val_loss_mean}, **compute_metrics(self.hparams.task , __lowerCAmelCase , __lowerCAmelCase)} lowerCAmelCase = dict(results.items()) lowerCAmelCase = results return ret, preds_list, out_label_list def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = self._eval_end(__lowerCAmelCase) lowerCAmelCase = ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = self._eval_end(__lowerCAmelCase) lowerCAmelCase = ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def a_ ( __lowerCAmelCase , __lowerCAmelCase): """simple docstring""" BaseTransformer.add_model_specific_args(__lowerCAmelCase , __lowerCAmelCase) parser.add_argument( """--max_seq_length""" , default=128 , type=__lowerCAmelCase , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--task""" , default="""""" , type=__lowerCAmelCase , required=__lowerCAmelCase , help="""The GLUE task to run""" , ) parser.add_argument( """--gpus""" , default=0 , type=__lowerCAmelCase , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""") return parser def snake_case__ ( ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase = argparse.ArgumentParser() add_generic_args(_A , os.getcwd() ) lowerCAmelCase = GLUETransformer.add_model_specific_args(_A , os.getcwd() ) lowerCAmelCase = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: lowerCAmelCase = os.path.join( """./results""" , f"{args.task}_{time.strftime('%Y%m%d_%H%M%S' )}" , ) os.makedirs(args.output_dir ) lowerCAmelCase = GLUETransformer(_A ) lowerCAmelCase = generic_train(_A , _A ) # Optionally, predict on dev set and write to output_dir if args.do_predict: lowerCAmelCase = sorted(glob.glob(os.path.join(args.output_dir , """checkpoint-epoch=*.ckpt""" ) , recursive=_A ) ) lowerCAmelCase = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(_A ) if __name__ == "__main__": main()
272
0
import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase): _a = None _a = BloomTokenizerFast _a = BloomTokenizerFast _a = True _a = False _a = "tokenizer_file" _a = {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"} def SCREAMING_SNAKE_CASE ( self: Dict ): super().setUp() lowercase :Dict = BloomTokenizerFast.from_pretrained("bigscience/tokenizer" ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE ( self: Optional[int] , **_lowerCAmelCase: Optional[Any] ): kwargs.update(self.special_tokens_map ) return BloomTokenizerFast.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self: int ): lowercase :Any = self.get_rust_tokenizer() lowercase :Dict = ['''The quick brown fox</s>''', '''jumps over the lazy dog</s>'''] lowercase :Union[str, Any] = [[21_75, 2_37_14, 7_31_73, 14_42_52, 2], [77, 13_26_19, 34_78, 3_68, 10_95_86, 3_54_33, 2]] lowercase :List[Any] = tokenizer.batch_encode_plus(_UpperCAmelCase )['''input_ids'''] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase :Dict = tokenizer.batch_decode(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _lowerCAmelCase: Tuple=6 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): lowercase :int = self.rust_tokenizer_class.from_pretrained(_UpperCAmelCase , **_UpperCAmelCase ) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input lowercase :Dict = '''This is a simple input''' lowercase :int = ['''This is a simple input 1''', '''This is a simple input 2'''] lowercase :int = ('''This is a simple input''', '''This is a pair''') lowercase :Optional[Any] = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests try: tokenizer_r.encode(_UpperCAmelCase , max_length=_UpperCAmelCase ) tokenizer_r.encode_plus(_UpperCAmelCase , max_length=_UpperCAmelCase ) tokenizer_r.batch_encode_plus(_UpperCAmelCase , max_length=_UpperCAmelCase ) tokenizer_r.encode(_UpperCAmelCase , max_length=_UpperCAmelCase ) tokenizer_r.batch_encode_plus(_UpperCAmelCase , max_length=_UpperCAmelCase ) except ValueError: self.fail("Bloom Tokenizer should be able to deal with padding" ) lowercase :Optional[int] = None # Hotfixing padding = None self.assertRaises(_UpperCAmelCase , tokenizer_r.encode , _UpperCAmelCase , max_length=_UpperCAmelCase , padding="max_length" ) # Simple input self.assertRaises(_UpperCAmelCase , tokenizer_r.encode_plus , _UpperCAmelCase , max_length=_UpperCAmelCase , padding="max_length" ) # Simple input self.assertRaises( _UpperCAmelCase , tokenizer_r.batch_encode_plus , _UpperCAmelCase , max_length=_UpperCAmelCase , padding="max_length" , ) # Pair input self.assertRaises(_UpperCAmelCase , tokenizer_r.encode , _UpperCAmelCase , max_length=_UpperCAmelCase , padding="max_length" ) # Pair input self.assertRaises(_UpperCAmelCase , tokenizer_r.encode_plus , _UpperCAmelCase , max_length=_UpperCAmelCase , padding="max_length" ) # Pair input self.assertRaises( _UpperCAmelCase , tokenizer_r.batch_encode_plus , _UpperCAmelCase , max_length=_UpperCAmelCase , padding="max_length" , ) def SCREAMING_SNAKE_CASE ( self: Optional[int] ): lowercase :Tuple = self.get_rust_tokenizer() lowercase :int = load_dataset("xnli" , "all_languages" , split="test" , streaming=_UpperCAmelCase ) lowercase :str = next(iter(_UpperCAmelCase ) )['''premise'''] # pick up one data lowercase :Any = list(sample_data.values() ) lowercase :Union[str, Any] = list(map(tokenizer.encode , _UpperCAmelCase ) ) lowercase :Union[str, Any] = [tokenizer.decode(_UpperCAmelCase , clean_up_tokenization_spaces=_UpperCAmelCase ) for x in output_tokens] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( self: int ): # The test has to be overriden because BLOOM uses ALiBi positional embeddings that does not have # any sequence length constraints. This test of the parent class will fail since it relies on the # maximum sequence length of the positoonal embeddings. self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map ) , 1 ) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values() )[0] ) , 1 )
352
import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCAmelCase : str = { "facebook/mask2former-swin-small-coco-instance": ( "https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json" ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } _UpperCAmelCase : Dict = logging.get_logger(__name__) class __lowerCAmelCase ( lowerCAmelCase): _a = '''mask2former''' _a = ['''swin'''] _a = {'''hidden_size''': '''hidden_dim'''} def __init__( self: List[str] , _lowerCAmelCase: Optional[Dict] = None , _lowerCAmelCase: int = 2_56 , _lowerCAmelCase: int = 2_56 , _lowerCAmelCase: int = 2_56 , _lowerCAmelCase: int = 10_24 , _lowerCAmelCase: str = "relu" , _lowerCAmelCase: int = 6 , _lowerCAmelCase: int = 10 , _lowerCAmelCase: int = 8 , _lowerCAmelCase: float = 0.0 , _lowerCAmelCase: int = 20_48 , _lowerCAmelCase: bool = False , _lowerCAmelCase: bool = False , _lowerCAmelCase: int = 4 , _lowerCAmelCase: int = 2_55 , _lowerCAmelCase: int = 1_00 , _lowerCAmelCase: float = 0.1 , _lowerCAmelCase: float = 2.0 , _lowerCAmelCase: float = 5.0 , _lowerCAmelCase: float = 5.0 , _lowerCAmelCase: int = 1_25_44 , _lowerCAmelCase: float = 3.0 , _lowerCAmelCase: float = 0.75 , _lowerCAmelCase: float = 0.02 , _lowerCAmelCase: float = 1.0 , _lowerCAmelCase: bool = True , _lowerCAmelCase: List[int] = [4, 8, 16, 32] , _lowerCAmelCase: bool = None , **_lowerCAmelCase: List[str] , ): if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `Swin` backbone." ) lowercase :Optional[int] = CONFIG_MAPPING["swin"]( image_size=2_24 , in_channels=3 , patch_size=4 , embed_dim=96 , depths=[2, 2, 18, 2] , num_heads=[3, 6, 12, 24] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=_lowerCAmelCase , out_features=["stage1", "stage2", "stage3", "stage4"] , ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ): lowercase :List[str] = backbone_config.pop("model_type" ) lowercase :Tuple = CONFIG_MAPPING[backbone_model_type] lowercase :int = config_class.from_dict(_lowerCAmelCase ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F"Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. " F"Supported model types: {','.join(self.backbones_supported )}" ) lowercase :Optional[Any] = backbone_config lowercase :Union[str, Any] = feature_size lowercase :Any = mask_feature_size lowercase :List[Any] = hidden_dim lowercase :Optional[int] = encoder_feedforward_dim lowercase :Dict = activation_function lowercase :Tuple = encoder_layers lowercase :List[str] = decoder_layers lowercase :Optional[Any] = num_attention_heads lowercase :Optional[Any] = dropout lowercase :Any = dim_feedforward lowercase :List[Any] = pre_norm lowercase :List[Any] = enforce_input_projection lowercase :Optional[int] = common_stride lowercase :List[Any] = ignore_value lowercase :Optional[int] = num_queries lowercase :List[str] = no_object_weight lowercase :Dict = class_weight lowercase :Union[str, Any] = mask_weight lowercase :List[Any] = dice_weight lowercase :Dict = train_num_points lowercase :Optional[int] = oversample_ratio lowercase :List[Any] = importance_sample_ratio lowercase :Dict = init_std lowercase :Union[str, Any] = init_xavier_std lowercase :Optional[Any] = use_auxiliary_loss lowercase :Any = feature_strides lowercase :int = output_auxiliary_logits lowercase :Dict = decoder_layers super().__init__(**_lowerCAmelCase ) @classmethod def SCREAMING_SNAKE_CASE ( cls: Tuple , _lowerCAmelCase: PretrainedConfig , **_lowerCAmelCase: str ): return cls( backbone_config=_lowerCAmelCase , **_lowerCAmelCase , ) def SCREAMING_SNAKE_CASE ( self: int ): lowercase :str = copy.deepcopy(self.__dict__ ) lowercase :Optional[Any] = self.backbone_config.to_dict() lowercase :Union[str, Any] = self.__class__.model_type return output
158
0
import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def __SCREAMING_SNAKE_CASE ( snake_case_ ): # picklable for multiprocessing '''simple docstring''' return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def __SCREAMING_SNAKE_CASE ( ): '''simple docstring''' with parallel_backend("spark" ): assert ParallelBackendConfig.backend_name == "spark" _UpperCAmelCase = [1, 2, 3] with pytest.raises(snake_case_ ): with parallel_backend("unsupported backend" ): map_nested(snake_case_ , snake_case_ , num_proc=2 ) with pytest.raises(snake_case_ ): with parallel_backend("unsupported backend" ): map_nested(snake_case_ , snake_case_ , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize("num_proc" , [2, -1] ) def __SCREAMING_SNAKE_CASE ( snake_case_ ): '''simple docstring''' _UpperCAmelCase = [1, 2] _UpperCAmelCase = {"a": 1, "b": 2} _UpperCAmelCase = {"a": [1, 2], "b": [3, 4]} _UpperCAmelCase = {"a": {"1": 1}, "b": 2} _UpperCAmelCase = {"a": 1, "b": 2, "c": 3, "d": 4} _UpperCAmelCase = [2, 3] _UpperCAmelCase = {"a": 2, "b": 3} _UpperCAmelCase = {"a": [2, 3], "b": [4, 5]} _UpperCAmelCase = {"a": {"1": 2}, "b": 3} _UpperCAmelCase = {"a": 2, "b": 3, "c": 4, "d": 5} with parallel_backend("spark" ): assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa assert map_nested(snake_case_ , snake_case_ , num_proc=snake_case_ ) == expected_map_nested_sa
133
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class __lowerCAmelCase ( unittest.TestCase ): snake_case_ : Tuple = StableDiffusionLDMaDPipeline snake_case_ : Optional[int] = TEXT_TO_IMAGE_PARAMS snake_case_ : str = TEXT_TO_IMAGE_BATCH_PARAMS snake_case_ : List[str] = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase ( self : Optional[int] ): """simple docstring""" torch.manual_seed(0 ) _UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) _UpperCAmelCase = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , ) torch.manual_seed(0 ) _UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=6 , out_channels=6 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) _UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) _UpperCAmelCase = CLIPTextModel(snake_case__ ) _UpperCAmelCase = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _UpperCAmelCase = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCamelCase ( self : Optional[int] , snake_case__ : Dict , snake_case__ : Optional[int]=0 ): """simple docstring""" if str(snake_case__ ).startswith("mps" ): _UpperCAmelCase = torch.manual_seed(snake_case__ ) else: _UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) _UpperCAmelCase = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCamelCase ( self : Optional[Any] ): """simple docstring""" _UpperCAmelCase = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = StableDiffusionLDMaDPipeline(**snake_case__ ) _UpperCAmelCase = ldmad_pipe.to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb[0, -3:, -3:, -1] _UpperCAmelCase = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) _UpperCAmelCase = np.array( [0.37_338_176, 0.70_247, 0.74_203_193, 0.51_643_604, 0.58_256_793, 0.60_932_136, 0.4_181_095, 0.48_355_877, 0.46_535_262] ) _UpperCAmelCase = np.array([103.46_727, 85.812_004, 87.849_236] ) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth ).max() < 1e-2 def UpperCamelCase ( self : List[str] ): """simple docstring""" _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = StableDiffusionLDMaDPipeline(**snake_case__ ) _UpperCAmelCase = ldmad_pipe.to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) _UpperCAmelCase = 3 * [inputs["prompt"]] # forward _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb_slice_a[0, -3:, -3:, -1] _UpperCAmelCase = depth_slice_a[0, -3:, -1] _UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) _UpperCAmelCase = 3 * [inputs.pop("prompt" )] _UpperCAmelCase = ldmad_pipe.tokenizer( snake_case__ , padding="max_length" , max_length=ldmad_pipe.tokenizer.model_max_length , truncation=snake_case__ , return_tensors="pt" , ) _UpperCAmelCase = text_inputs["input_ids"].to(snake_case__ ) _UpperCAmelCase = ldmad_pipe.text_encoder(snake_case__ )[0] _UpperCAmelCase = prompt_embeds # forward _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb_slice_a[0, -3:, -3:, -1] _UpperCAmelCase = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten() ).max() < 1e-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten() ).max() < 1e-4 def UpperCamelCase ( self : List[str] ): """simple docstring""" _UpperCAmelCase = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = PNDMScheduler(skip_prk_steps=snake_case__ ) _UpperCAmelCase = StableDiffusionLDMaDPipeline(**snake_case__ ) _UpperCAmelCase = ldmad_pipe.to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) _UpperCAmelCase = "french fries" _UpperCAmelCase = ldmad_pipe(**snake_case__ , negative_prompt=snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb[0, -3:, -3:, -1] _UpperCAmelCase = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) _UpperCAmelCase = np.array( [0.37_044, 0.71_811_503, 0.7_223_251, 0.48_603_675, 0.5_638_391, 0.6_364_948, 0.42_833_704, 0.4_901_315, 0.47_926_217] ) _UpperCAmelCase = np.array([107.84_738, 84.62_802, 89.962_135] ) assert np.abs(rgb_slice.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(depth_slice.flatten() - expected_slice_depth ).max() < 1e-2 @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def UpperCamelCase ( self : Tuple ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self : str , snake_case__ : Optional[int] , snake_case__ : Tuple="cpu" , snake_case__ : Any=torch.floataa , snake_case__ : Dict=0 ): """simple docstring""" _UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) _UpperCAmelCase = np.random.RandomState(snake_case__ ).standard_normal((1, 4, 64, 64) ) _UpperCAmelCase = torch.from_numpy(snake_case__ ).to(device=snake_case__ , dtype=snake_case__ ) _UpperCAmelCase = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCamelCase ( self : Any ): """simple docstring""" _UpperCAmelCase = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ) _UpperCAmelCase = ldmad_pipe.to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_inputs(snake_case__ ) _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb[0, -3:, -3:, -1].flatten() _UpperCAmelCase = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 512, 512, 3) assert depth.shape == (1, 512, 512) _UpperCAmelCase = np.array( [0.53_805_465, 0.56_707_305, 0.5_486_515, 0.57_012_236, 0.5_814_511, 0.56_253_487, 0.54_843_014, 0.55_092_263, 0.6_459_706] ) _UpperCAmelCase = np.array( [0.9_263_781, 0.6_678_672, 0.5_486_515, 0.92_202_145, 0.67_831_135, 0.56_253_487, 0.9_241_694, 0.7_551_478, 0.6_459_706] ) assert np.abs(rgb_slice - expected_slice_rgb ).max() < 3e-3 assert np.abs(depth_slice - expected_slice_depth ).max() < 3e-3 @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def UpperCamelCase ( self : Dict ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self : Any , snake_case__ : Optional[Any] , snake_case__ : int="cpu" , snake_case__ : Optional[Any]=torch.floataa , snake_case__ : Optional[Any]=0 ): """simple docstring""" _UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) _UpperCAmelCase = np.random.RandomState(snake_case__ ).standard_normal((1, 4, 64, 64) ) _UpperCAmelCase = torch.from_numpy(snake_case__ ).to(device=snake_case__ , dtype=snake_case__ ) _UpperCAmelCase = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 50, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCamelCase ( self : Union[str, Any] ): """simple docstring""" _UpperCAmelCase = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ).to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_inputs(snake_case__ ) _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = 0.495_586 _UpperCAmelCase = 0.33_795_515 _UpperCAmelCase = 112.48_518 _UpperCAmelCase = 98.489_746 assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3 def UpperCamelCase ( self : Tuple ): """simple docstring""" _UpperCAmelCase = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d-4c" ).to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_inputs(snake_case__ ) _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = 0.4_194_127 _UpperCAmelCase = 0.35_375_586 _UpperCAmelCase = 0.5_638_502 _UpperCAmelCase = 0.34_686_103 assert rgb.shape == (1, 512, 512, 3) assert depth.shape == (1, 512, 512, 1) assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3
133
1
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
366
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase = "cpu" , __lowerCamelCase = None ) -> None: lowercase__ : List[str] = torch.load(__lowerCamelCase , map_location=__lowerCamelCase ) for k, v in tqdm(state_dict.items() ): if not isinstance(__lowerCamelCase , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) lowercase__ : List[Any] = v.half() if save_path is None: # overwrite src_path lowercase__ : Any = src_path torch.save(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": fire.Fire(convert)
302
0
"""simple docstring""" import unittest import numpy as np from transformers import MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING, TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING from transformers.pipelines import AudioClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_torchaudio, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _UpperCAmelCase ( unittest.TestCase): __a : List[Any] = MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING __a : Dict = TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING def __snake_case ( self , _A , _A , _A ) -> Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = AudioClassificationPipeline(model=_A , feature_extractor=_A ) # test with a raw waveform _UpperCAmelCase : Optional[Any] = np.zeros((3_40_00,) ) _UpperCAmelCase : Tuple = np.zeros((1_40_00,) ) return audio_classifier, [audioa, audio] def __snake_case ( self , _A , _A ) -> Tuple: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = examples _UpperCAmelCase : Tuple = audio_classifier(_A ) # by default a model is initialized with num_labels=2 self.assertEqual( _A , [ {"""score""": ANY(_A ), """label""": ANY(_A )}, {"""score""": ANY(_A ), """label""": ANY(_A )}, ] , ) _UpperCAmelCase : List[str] = audio_classifier(_A , top_k=1 ) self.assertEqual( _A , [ {"""score""": ANY(_A ), """label""": ANY(_A )}, ] , ) self.run_torchaudio(_A ) @require_torchaudio def __snake_case ( self , _A ) -> Union[str, Any]: '''simple docstring''' import datasets # test with a local file _UpperCAmelCase : int = datasets.load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) _UpperCAmelCase : Dict = dataset[0]["""audio"""]["""array"""] _UpperCAmelCase : Optional[int] = audio_classifier(_A ) self.assertEqual( _A , [ {"""score""": ANY(_A ), """label""": ANY(_A )}, {"""score""": ANY(_A ), """label""": ANY(_A )}, ] , ) @require_torch def __snake_case ( self ) -> List[str]: '''simple docstring''' _UpperCAmelCase : Dict = """anton-l/wav2vec2-random-tiny-classifier""" _UpperCAmelCase : int = pipeline("""audio-classification""" , model=_A ) _UpperCAmelCase : str = np.ones((80_00,) ) _UpperCAmelCase : Optional[Any] = audio_classifier(_A , top_k=4 ) _UpperCAmelCase : Tuple = [ {"""score""": 0.0842, """label""": """no"""}, {"""score""": 0.0838, """label""": """up"""}, {"""score""": 0.0837, """label""": """go"""}, {"""score""": 0.0834, """label""": """right"""}, ] _UpperCAmelCase : Any = [ {"""score""": 0.0845, """label""": """stop"""}, {"""score""": 0.0844, """label""": """on"""}, {"""score""": 0.0841, """label""": """right"""}, {"""score""": 0.0834, """label""": """left"""}, ] self.assertIn(nested_simplify(_A , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) _UpperCAmelCase : Dict = {"""array""": np.ones((80_00,) ), """sampling_rate""": audio_classifier.feature_extractor.sampling_rate} _UpperCAmelCase : Union[str, Any] = audio_classifier(_A , top_k=4 ) self.assertIn(nested_simplify(_A , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) @require_torch @slow def __snake_case ( self ) -> Optional[int]: '''simple docstring''' import datasets _UpperCAmelCase : Dict = """superb/wav2vec2-base-superb-ks""" _UpperCAmelCase : Dict = pipeline("""audio-classification""" , model=_A ) _UpperCAmelCase : List[Any] = datasets.load_dataset("""anton-l/superb_dummy""" , """ks""" , split="""test""" ) _UpperCAmelCase : Tuple = np.array(dataset[3]["""speech"""] , dtype=np.floataa ) _UpperCAmelCase : List[Any] = audio_classifier(_A , top_k=4 ) self.assertEqual( nested_simplify(_A , decimals=3 ) , [ {"""score""": 0.981, """label""": """go"""}, {"""score""": 0.007, """label""": """up"""}, {"""score""": 0.006, """label""": """_unknown_"""}, {"""score""": 0.001, """label""": """down"""}, ] , ) @require_tf @unittest.skip("""Audio classification is not implemented for TF""" ) def __snake_case ( self ) -> Optional[Any]: '''simple docstring''' pass
246
"""simple docstring""" import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class _UpperCAmelCase ( unittest.TestCase): __a : Tuple = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def __snake_case ( self , _A , _A , _A ) -> Union[str, Any]: '''simple docstring''' _UpperCAmelCase : str = hf_hub_download( repo_id="""nateraw/video-demo""" , filename="""archery.mp4""" , repo_type="""dataset""" ) _UpperCAmelCase : Tuple = VideoClassificationPipeline(model=_A , image_processor=_A , top_k=2 ) _UpperCAmelCase : List[str] = [ example_video_filepath, """https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4""", ] return video_classifier, examples def __snake_case ( self , _A , _A ) -> Optional[int]: '''simple docstring''' for example in examples: _UpperCAmelCase : str = video_classifier(_A ) self.assertEqual( _A , [ {"""score""": ANY(_A ), """label""": ANY(_A )}, {"""score""": ANY(_A ), """label""": ANY(_A )}, ] , ) @require_torch def __snake_case ( self ) -> List[str]: '''simple docstring''' _UpperCAmelCase : List[Any] = """hf-internal-testing/tiny-random-VideoMAEForVideoClassification""" _UpperCAmelCase : Optional[Any] = VideoMAEFeatureExtractor( size={"""shortest_edge""": 10} , crop_size={"""height""": 10, """width""": 10} ) _UpperCAmelCase : List[str] = pipeline( """video-classification""" , model=_A , feature_extractor=_A , frame_sampling_rate=4 ) _UpperCAmelCase : Union[str, Any] = hf_hub_download(repo_id="""nateraw/video-demo""" , filename="""archery.mp4""" , repo_type="""dataset""" ) _UpperCAmelCase : Union[str, Any] = video_classifier(_A , top_k=2 ) self.assertEqual( nested_simplify(_A , decimals=4 ) , [{"""score""": 0.5199, """label""": """LABEL_0"""}, {"""score""": 0.4801, """label""": """LABEL_1"""}] , ) _UpperCAmelCase : int = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(_A , decimals=4 ) , [ [{"""score""": 0.5199, """label""": """LABEL_0"""}, {"""score""": 0.4801, """label""": """LABEL_1"""}], [{"""score""": 0.5199, """label""": """LABEL_0"""}, {"""score""": 0.4801, """label""": """LABEL_1"""}], ] , ) @require_tf def __snake_case ( self ) -> Any: '''simple docstring''' pass
246
1
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): __SCREAMING_SNAKE_CASE : Dict = 'pt' elif is_tf_available(): __SCREAMING_SNAKE_CASE : Tuple = 'tf' else: __SCREAMING_SNAKE_CASE : Any = 'jax' class lowercase_ ( __snake_case , unittest.TestCase ): _lowerCamelCase = ByTaTokenizer _lowerCamelCase = False def UpperCamelCase ( self ): super().setUp() _snake_case : int = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase ( self ): return ByTaTokenizer.from_pretrained("google/byt5-small" ) def UpperCamelCase ( self , **lowercase_ ): return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_=False , lowercase_=20 , lowercase_=5 ): # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for ByT5 because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. _snake_case : int = [] for i in range(len(lowercase_ ) ): try: _snake_case : List[Any] = tokenizer.decode([i] , clean_up_tokenization_spaces=lowercase_ ) except UnicodeDecodeError: pass toks.append((i, tok) ) _snake_case : List[str] = list(filter(lambda lowercase_ : re.match(R"^[ a-zA-Z]+$" , t[1] ) , lowercase_ ) ) _snake_case : Optional[Any] = list(filter(lambda lowercase_ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=lowercase_ ) , lowercase_ ) ) if max_length is not None and len(lowercase_ ) > max_length: _snake_case : Tuple = toks[:max_length] if min_length is not None and len(lowercase_ ) < min_length and len(lowercase_ ) > 0: while len(lowercase_ ) < min_length: _snake_case : str = toks + toks # toks_str = [t[1] for t in toks] _snake_case : Optional[int] = [t[0] for t in toks] # Ensure consistency _snake_case : List[str] = tokenizer.decode(lowercase_ , clean_up_tokenization_spaces=lowercase_ ) if " " not in output_txt and len(lowercase_ ) > 1: _snake_case : Dict = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=lowercase_ ) + " " + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=lowercase_ ) ) if with_prefix_space: _snake_case : Optional[Any] = " " + output_txt _snake_case : List[Any] = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) return output_txt, output_ids def UpperCamelCase ( self ): _snake_case : int = self.ta_base_tokenizer _snake_case : str = tokenizer(["hi</s>", "I went to the gym</s>", "</s>"] ) _snake_case : Tuple = tokenizer(["hi", "I went to the gym", ""] ) self.assertListEqual(batch_with_eos_added["input_ids"] , batch_without_eos_added["input_ids"] ) def UpperCamelCase ( self ): _snake_case : Dict = self.ta_base_tokenizer _snake_case : List[str] = "Unicode €." _snake_case : Tuple = tokenizer(lowercase_ ) _snake_case : Tuple = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded["input_ids"] , lowercase_ ) # decoding _snake_case : str = tokenizer.decode(lowercase_ ) self.assertEqual(lowercase_ , "Unicode €.</s>" ) _snake_case : List[Any] = tokenizer("e è é ê ë" ) _snake_case : Optional[int] = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded["input_ids"] , lowercase_ ) # decoding _snake_case : Union[str, Any] = tokenizer.decode(lowercase_ ) self.assertEqual(lowercase_ , "e è é ê ë</s>" ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("e è é ê ë" ) ) , "e è é ê ë</s>" ) def UpperCamelCase ( self ): _snake_case : List[str] = self.ta_base_tokenizer _snake_case : List[str] = ["A long paragraph for summarization.", "Another paragraph for summarization."] # fmt: off _snake_case : Union[str, Any] = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on _snake_case : Optional[Any] = tokenizer(lowercase_ , padding=lowercase_ , return_tensors=lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) if FRAMEWORK != "jax": _snake_case : List[str] = list(batch.input_ids.numpy()[0] ) else: _snake_case : Optional[int] = list(batch.input_ids.tolist()[0] ) self.assertListEqual(lowercase_ , lowercase_ ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def UpperCamelCase ( self ): _snake_case : Any = self.ta_base_tokenizer _snake_case : List[Any] = ["A long paragraph for summarization.", "Another paragraph for summarization."] _snake_case : Optional[int] = tokenizer(lowercase_ , padding=lowercase_ , return_tensors=lowercase_ ) # check if input_ids are returned and no decoder_input_ids self.assertIn("input_ids" , lowercase_ ) self.assertIn("attention_mask" , lowercase_ ) self.assertNotIn("decoder_input_ids" , lowercase_ ) self.assertNotIn("decoder_attention_mask" , lowercase_ ) def UpperCamelCase ( self ): _snake_case : int = self.ta_base_tokenizer _snake_case : Any = [ "Summary of the text.", "Another summary.", ] _snake_case : Any = tokenizer( text_target=lowercase_ , max_length=32 , padding="max_length" , truncation=lowercase_ , return_tensors=lowercase_ ) self.assertEqual(32 , targets["input_ids"].shape[1] ) def UpperCamelCase ( self ): _snake_case : str = self.ta_base_tokenizer _snake_case : Tuple = ["A long paragraph for summarization. </s>"] _snake_case : Optional[int] = ["Summary of the text. </s>"] # fmt: off _snake_case : List[str] = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] _snake_case : int = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on _snake_case : List[str] = tokenizer(lowercase_ , text_target=lowercase_ ) self.assertEqual(lowercase_ , batch["input_ids"][0] ) self.assertEqual(lowercase_ , batch["labels"][0] ) def UpperCamelCase ( self ): # safety check on max_len default value so we are sure the test works _snake_case : List[str] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test _snake_case : Union[str, Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc _snake_case : Any = tempfile.mkdtemp() _snake_case : Dict = " He is very happy, UNwant\u00E9d,running" _snake_case : Tuple = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) tokenizer.save_pretrained(lowercase_ ) _snake_case : Tuple = tokenizer.__class__.from_pretrained(lowercase_ ) _snake_case : List[str] = after_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) shutil.rmtree(lowercase_ ) _snake_case : Optional[Any] = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc _snake_case : Union[str, Any] = tempfile.mkdtemp() _snake_case : Any = " He is very happy, UNwant\u00E9d,running" tokenizer.add_tokens(["bim", "bambam"] ) _snake_case : str = tokenizer.additional_special_tokens additional_special_tokens.append("new_additional_special_token" ) tokenizer.add_special_tokens({"additional_special_tokens": additional_special_tokens} ) _snake_case : Tuple = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) tokenizer.save_pretrained(lowercase_ ) _snake_case : str = tokenizer.__class__.from_pretrained(lowercase_ ) _snake_case : Tuple = after_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) self.assertIn("new_additional_special_token" , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) _snake_case : Dict = tokenizer.__class__.from_pretrained(lowercase_ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(lowercase_ ) def UpperCamelCase ( self ): _snake_case : Optional[int] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(lowercase_ ) with open(os.path.join(lowercase_ , "special_tokens_map.json" ) , encoding="utf-8" ) as json_file: _snake_case : int = json.load(lowercase_ ) with open(os.path.join(lowercase_ , "tokenizer_config.json" ) , encoding="utf-8" ) as json_file: _snake_case : Tuple = json.load(lowercase_ ) _snake_case : Union[str, Any] = [f"""<extra_id_{i}>""" for i in range(125 )] _snake_case : Union[str, Any] = added_tokens_extra_ids + [ "an_additional_special_token" ] _snake_case : Any = added_tokens_extra_ids + [ "an_additional_special_token" ] with open(os.path.join(lowercase_ , "special_tokens_map.json" ) , "w" , encoding="utf-8" ) as outfile: json.dump(lowercase_ , lowercase_ ) with open(os.path.join(lowercase_ , "tokenizer_config.json" ) , "w" , encoding="utf-8" ) as outfile: json.dump(lowercase_ , lowercase_ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files _snake_case : str = tokenizer_class.from_pretrained( lowercase_ , ) self.assertIn( "an_additional_special_token" , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ["an_additional_special_token"] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["an_additional_special_token"] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained _snake_case : str = added_tokens_extra_ids + [AddedToken("a_new_additional_special_token" , lstrip=lowercase_ )] _snake_case : Dict = tokenizer_class.from_pretrained( lowercase_ , additional_special_tokens=lowercase_ , ) self.assertIn("a_new_additional_special_token" , tokenizer.additional_special_tokens ) self.assertEqual( ["a_new_additional_special_token"] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["a_new_additional_special_token"] ) ) , ) def UpperCamelCase ( self ): _snake_case : int = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(lowercase_ ) _snake_case : Optional[Any] = tokenizer_class.from_pretrained(lowercase_ ) self.assertTrue(tokenizer.decode([255] ) == "" ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): # The default common tokenizer tests uses invalid tokens for ByT5 that can only accept one-character strings # and special added tokens as tokens _snake_case : Dict = self.get_tokenizers(fast=lowercase_ , do_lower_case=lowercase_ ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): _snake_case : List[str] = ["t", "h", "i", "s", " ", "i", "s", " ", "a", " ", "t", "e", "x", "t", "</s>"] _snake_case : Optional[int] = tokenizer.convert_tokens_to_string(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) def UpperCamelCase ( self ): _snake_case : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): _snake_case : List[str] = [ "bos_token", "eos_token", "unk_token", "sep_token", "pad_token", "cls_token", "mask_token", ] _snake_case : Dict = 0 _snake_case : Dict = tokenizer.convert_ids_to_tokens( lowercase_ , skip_special_tokens=lowercase_ ) for attr in attributes_list: setattr(lowercase_ , attr + "_id" , lowercase_ ) self.assertEqual(getattr(lowercase_ , lowercase_ ) , lowercase_ ) self.assertEqual(getattr(lowercase_ , attr + "_id" ) , lowercase_ ) setattr(lowercase_ , attr + "_id" , lowercase_ ) self.assertEqual(getattr(lowercase_ , lowercase_ ) , lowercase_ ) self.assertEqual(getattr(lowercase_ , attr + "_id" ) , lowercase_ ) setattr(lowercase_ , "additional_special_tokens_ids" , [] ) self.assertListEqual(getattr(lowercase_ , "additional_special_tokens" ) , [] ) self.assertListEqual(getattr(lowercase_ , "additional_special_tokens_ids" ) , [] ) setattr(lowercase_ , "additional_special_tokens_ids" , [token_id_to_test_setters] ) self.assertListEqual(getattr(lowercase_ , "additional_special_tokens" ) , [token_to_test_setters] ) self.assertListEqual(getattr(lowercase_ , "additional_special_tokens_ids" ) , [token_id_to_test_setters] )
284
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem __SCREAMING_SNAKE_CASE : Any = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 __SCREAMING_SNAKE_CASE : List[compression.BaseCompressedFileFileSystem] = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(F'''A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.''') fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def snake_case (__lowercase ) -> str: '''simple docstring''' if "://" in dataset_path: _snake_case : Tuple = dataset_path.split("://" )[1] return dataset_path def snake_case (__lowercase ) -> bool: '''simple docstring''' if fs is not None and fs.protocol != "file": return True else: return False def snake_case (__lowercase , __lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : Optional[int] = not is_remote_filesystem(__lowercase ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(__lowercase ) , fs._strip_protocol(__lowercase ) ) else: fs.mv(__lowercase , __lowercase , recursive=__lowercase ) def snake_case () -> None: '''simple docstring''' if hasattr(fsspec.asyn , "reset_lock" ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: _snake_case : List[str] = None _snake_case : str = None _snake_case : List[Any] = threading.Lock()
284
1
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer A : List[str] = logging.get_logger(__name__) A : Dict = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } A : Dict = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } A : Optional[int] = { 'facebook/blenderbot_small-90M': 5_1_2, } class A (__UpperCAmelCase ): '''simple docstring''' __lowerCamelCase : int = VOCAB_FILES_NAMES __lowerCamelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase : List[Any] = BlenderbotSmallTokenizer def __init__( self : Optional[int] , __lowerCAmelCase : Dict=None , __lowerCAmelCase : List[Any]=None , __lowerCAmelCase : Optional[int]="<|endoftext|>" , __lowerCAmelCase : Tuple="<|endoftext|>" , __lowerCAmelCase : Optional[int]="<|endoftext|>" , __lowerCAmelCase : Any=False , __lowerCAmelCase : Dict=True , **__lowerCAmelCase : Any , ) -> Optional[int]: """simple docstring""" super().__init__( ByteLevelBPETokenizer( vocab=_lowerCamelCase , merges=_lowerCamelCase , add_prefix_space=_lowerCamelCase , trim_offsets=_lowerCamelCase , ) , bos_token=_lowerCamelCase , eos_token=_lowerCamelCase , unk_token=_lowerCamelCase , **_lowerCamelCase , ) A__ = add_prefix_space def a_ ( self : int , __lowerCAmelCase : Tuple , __lowerCAmelCase : List[Any]=None ) -> int: """simple docstring""" A__ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def a_ ( self : Tuple , __lowerCAmelCase : List[int] , __lowerCAmelCase : Optional[List[int]] = None ) -> Optional[int]: """simple docstring""" A__ = [self.sep_token_id] A__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
274
"""simple docstring""" import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None ): """simple docstring""" assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" A_ : Dict = nn.Parameter(_UpperCAmelCase ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" A_ : Optional[Any] = nn.Parameter(_UpperCAmelCase ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Optional[int] = np.asarray(weights[0] ) A_ : Optional[Any] = np.asarray(weights[1] ) A_ : Union[str, Any] = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(_UpperCAmelCase ).transpose(1 , 2 ).contiguous().view(-1 , _UpperCAmelCase ) , ) set_param( torch_layer.self_attention.value , torch.tensor(_UpperCAmelCase ).transpose(1 , 2 ).contiguous().view(-1 , _UpperCAmelCase ) , ) set_param( torch_layer.output.dense , torch.tensor(_UpperCAmelCase ).view(-1 , _UpperCAmelCase ).contiguous().transpose(0 , 1 ) , ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : int = np.asarray(weights[0] ) A_ : Optional[int] = np.asarray(weights[1] ) A_ : int = np.asarray(weights[2] ) A_ : List[str] = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(_UpperCAmelCase ).transpose(1 , 2 ).contiguous().view(-1 , _UpperCAmelCase ) , ) set_param( torch_layer.self_attention.key , torch.tensor(_UpperCAmelCase ).transpose(1 , 2 ).contiguous().view(-1 , _UpperCAmelCase ) , ) set_param( torch_layer.self_attention.value , torch.tensor(_UpperCAmelCase ).transpose(1 , 2 ).contiguous().view(-1 , _UpperCAmelCase ) , ) set_param( torch_layer.output.dense , torch.tensor(_UpperCAmelCase ).view(-1 , _UpperCAmelCase ).contiguous().transpose(0 , 1 ) , ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[Any] = weights[0][0][0] A_ : Any = np.asarray(layer_norm_a[0] ) A_ : List[str] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(_UpperCAmelCase ) , torch.tensor(_UpperCAmelCase ) , ) # lsh weights + output A_ : List[str] = weights[0][1] if len(_UpperCAmelCase ) < 4: set_layer_weights_in_torch_lsh(_UpperCAmelCase , torch_block.attention , _UpperCAmelCase ) else: set_layer_weights_in_torch_local(_UpperCAmelCase , torch_block.attention , _UpperCAmelCase ) # intermediate weighs A_ : Dict = weights[2][0][1][2] # Chunked Feed Forward if len(_UpperCAmelCase ) == 4: A_ : Tuple = intermediate_weights[2] # layernorm 2 A_ : List[Any] = np.asarray(intermediate_weights[0][0] ) A_ : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(_UpperCAmelCase ) , torch.tensor(_UpperCAmelCase ) , ) # intermediate dense A_ : Optional[int] = np.asarray(intermediate_weights[1][0] ) A_ : List[str] = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(_UpperCAmelCase ).transpose(0 , 1 ).contiguous() , torch.tensor(_UpperCAmelCase ) , ) # intermediate out A_ : List[str] = np.asarray(intermediate_weights[4][0] ) A_ : int = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(_UpperCAmelCase ).transpose(0 , 1 ).contiguous() , torch.tensor(_UpperCAmelCase ) , ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Any = torch_model.reformer # word embeds A_ : str = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(_UpperCAmelCase ) , ) if isinstance(weights[3] , _UpperCAmelCase ): A_ : Tuple = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): A_ : Tuple = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" A_ : Tuple = nn.Parameter(torch.tensor(_UpperCAmelCase ) ) A_ : str = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( _UpperCAmelCase ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): A_ : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # output layer norm A_ : int = np.asarray(weights[7][0] ) A_ : str = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(_UpperCAmelCase ) , torch.tensor(_UpperCAmelCase ) , ) # output embeddings A_ : Optional[Any] = np.asarray(weights[9][0] ) A_ : Tuple = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(_UpperCAmelCase ).transpose(0 , 1 ).contiguous() , torch.tensor(_UpperCAmelCase ) , ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = ReformerConfig.from_json_file(_UpperCAmelCase ) print(f"""Building PyTorch model from configuration: {config}""" ) A_ : Optional[Any] = ReformerModelWithLMHead(_UpperCAmelCase ) with open(_UpperCAmelCase , '''rb''' ) as f: A_ : Union[str, Any] = pickle.load(_UpperCAmelCase )['''weights'''] set_model_weights_in_torch(_UpperCAmelCase , _UpperCAmelCase , config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _UpperCAmelCase ) if __name__ == "__main__": _lowerCamelCase : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--trax_model_pkl_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained Reformer model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) _lowerCamelCase : Dict = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
167
0
"""simple docstring""" import json import logging import math import os import sys from dataclasses import dataclass, field from typing import Optional from datasets import Dataset, load_dataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_MASKED_LM_MAPPING, AutoConfig, AutoModelForMaskedLM, AutoTokenizer, DataCollatorForWholeWordMask, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process a_ = logging.getLogger(__name__) a_ = list(MODEL_FOR_MASKED_LM_MAPPING.keys()) a_ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class __snake_case : """simple docstring""" _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={ """help""": ( """The model checkpoint for weights initialization.Don't set if you want to train a model from scratch.""" ) } , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """If training from scratch, pass a model type from the list: """ + """, """.join(SCREAMING_SNAKE_CASE__ )} , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={ """help""": ( """Override some existing default config settings when a model is trained from scratch. Example: """ """n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index""" ) } , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) _lowerCamelCase = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) def UpperCamelCase__( self ): '''simple docstring''' if self.config_overrides is not None and (self.config_name is not None or self.model_name_or_path is not None): raise ValueError( '''--config_overrides can\'t be used in combination with --config_name or --model_name_or_path''' ) @dataclass class __snake_case : """simple docstring""" _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) _lowerCamelCase = field(default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """The input training data file (a text file)."""} ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """An optional input train ref data file for whole word masking in Chinese."""} , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """An optional input validation ref data file for whole word masking in Chinese."""} , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) _lowerCamelCase = field( default=5 , metadata={ """help""": """The percentage of the train set used as validation set in case there's no validation split""" } , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated. Default to the max input length of the model.""" ) } , ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) _lowerCamelCase = field( default=0.1_5 , metadata={"""help""": """Ratio of tokens to mask for masked language modeling loss"""} ) _lowerCamelCase = field( default=SCREAMING_SNAKE_CASE__ , metadata={ """help""": ( """Whether to pad all samples to `max_seq_length`. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch.""" ) } , ) def UpperCamelCase__( self ): '''simple docstring''' if self.train_file is not None: __A : List[Any] = self.train_file.split('''.''' )[-1] assert extension in ["csv", "json", "txt"], "`train_file` should be a csv, a json or a txt file." if self.validation_file is not None: __A : Optional[Any] = self.validation_file.split('''.''' )[-1] assert extension in ["csv", "json", "txt"], "`validation_file` should be a csv, a json or a txt file." def __lowercase ( snake_case_ : List[Any] ,snake_case_ : Dict ) ->Union[str, Any]: '''simple docstring''' with open(lowercase__ ,'''r''' ,encoding='''utf-8''' ) as f: __A : Dict = [json.loads(lowercase__ ) for line in f.read().splitlines() if (len(lowercase__ ) > 0 and not line.isspace())] assert len(lowercase__ ) == len(lowercase__ ) __A : Tuple = {c: dataset[c] for c in dataset.column_names} __A : Tuple = refs return Dataset.from_dict(lowercase__ ) def __lowercase ( ) ->str: '''simple docstring''' __A : List[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. __A : List[Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: __A : Optional[int] = parser.parse_args_into_dataclasses() # Detecting last checkpoint. __A : Dict = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: __A : Optional[int] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' ,datefmt='''%m/%d/%Y %H:%M:%S''' ,handlers=[logging.StreamHandler(sys.stdout )] ,) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' ,lowercase__ ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. __A : List[Any] = load_dataset(data_args.dataset_name ,data_args.dataset_config_name ) if "validation" not in datasets.keys(): __A : Optional[int] = load_dataset( data_args.dataset_name ,data_args.dataset_config_name ,split=F"""train[:{data_args.validation_split_percentage}%]""" ,) __A : List[str] = load_dataset( data_args.dataset_name ,data_args.dataset_config_name ,split=F"""train[{data_args.validation_split_percentage}%:]""" ,) else: __A : Union[str, Any] = {} if data_args.train_file is not None: __A : Union[str, Any] = data_args.train_file if data_args.validation_file is not None: __A : List[str] = data_args.validation_file __A : List[str] = data_args.train_file.split('''.''' )[-1] if extension == "txt": __A : int = 'text' __A : Tuple = load_dataset(lowercase__ ,data_files=lowercase__ ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. __A : Union[str, Any] = { 'cache_dir': model_args.cache_dir, 'revision': model_args.model_revision, 'use_auth_token': True if model_args.use_auth_token else None, } if model_args.config_name: __A : Dict = AutoConfig.from_pretrained(model_args.config_name ,**lowercase__ ) elif model_args.model_name_or_path: __A : List[Any] = AutoConfig.from_pretrained(model_args.model_name_or_path ,**lowercase__ ) else: __A : List[str] = CONFIG_MAPPING[model_args.model_type]() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.config_overrides is not None: logger.info(F"""Overriding config: {model_args.config_overrides}""" ) config.update_from_string(model_args.config_overrides ) logger.info(F"""New config: {config}""" ) __A : int = { 'cache_dir': model_args.cache_dir, 'use_fast': model_args.use_fast_tokenizer, 'revision': model_args.model_revision, 'use_auth_token': True if model_args.use_auth_token else None, } if model_args.tokenizer_name: __A : Dict = AutoTokenizer.from_pretrained(model_args.tokenizer_name ,**lowercase__ ) elif model_args.model_name_or_path: __A : Optional[Any] = AutoTokenizer.from_pretrained(model_args.model_name_or_path ,**lowercase__ ) else: raise ValueError( '''You are instantiating a new tokenizer from scratch. This is not supported by this script.''' '''You can do it from another script, save it, and load it from here, using --tokenizer_name.''' ) if model_args.model_name_or_path: __A : Optional[int] = AutoModelForMaskedLM.from_pretrained( model_args.model_name_or_path ,from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) ,config=lowercase__ ,cache_dir=model_args.cache_dir ,revision=model_args.model_revision ,use_auth_token=True if model_args.use_auth_token else None ,) else: logger.info('''Training new model from scratch''' ) __A : List[Any] = AutoModelForMaskedLM.from_config(lowercase__ ) model.resize_token_embeddings(len(lowercase__ ) ) # Preprocessing the datasets. # First we tokenize all the texts. if training_args.do_train: __A : Union[str, Any] = datasets['train'].column_names else: __A : str = datasets['validation'].column_names __A : List[Any] = 'text' if 'text' in column_names else column_names[0] __A : Optional[Any] = 'max_length' if data_args.pad_to_max_length else False def tokenize_function(snake_case_ : str ): # Remove empty lines __A : Optional[Any] = [line for line in examples['text'] if len(lowercase__ ) > 0 and not line.isspace()] return tokenizer(examples['''text'''] ,padding=lowercase__ ,truncation=lowercase__ ,max_length=data_args.max_seq_length ) __A : List[Any] = datasets.map( lowercase__ ,batched=lowercase__ ,num_proc=data_args.preprocessing_num_workers ,remove_columns=[text_column_name] ,load_from_cache_file=not data_args.overwrite_cache ,) # Add the chinese references if provided if data_args.train_ref_file is not None: __A : Optional[Any] = add_chinese_references(tokenized_datasets['''train'''] ,data_args.train_ref_file ) if data_args.validation_ref_file is not None: __A : Dict = add_chinese_references( tokenized_datasets['''validation'''] ,data_args.validation_ref_file ) # If we have ref files, need to avoid it removed by trainer __A : int = data_args.train_ref_file or data_args.validation_ref_file if has_ref: __A : str = False # Data collator # This one will take care of randomly masking the tokens. __A : Union[str, Any] = DataCollatorForWholeWordMask(tokenizer=lowercase__ ,mlm_probability=data_args.mlm_probability ) # Initialize our Trainer __A : str = Trainer( model=lowercase__ ,args=lowercase__ ,train_dataset=tokenized_datasets['''train'''] if training_args.do_train else None ,eval_dataset=tokenized_datasets['''validation'''] if training_args.do_eval else None ,tokenizer=lowercase__ ,data_collator=lowercase__ ,) # Training if training_args.do_train: if last_checkpoint is not None: __A : Any = last_checkpoint elif model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ): __A : Optional[Any] = model_args.model_name_or_path else: __A : str = None __A : Optional[int] = trainer.train(resume_from_checkpoint=lowercase__ ) trainer.save_model() # Saves the tokenizer too for easy upload __A : List[Any] = os.path.join(training_args.output_dir ,'''train_results.txt''' ) if trainer.is_world_process_zero(): with open(lowercase__ ,'''w''' ) as writer: logger.info('''***** Train results *****''' ) for key, value in sorted(train_result.metrics.items() ): logger.info(F""" {key} = {value}""" ) writer.write(F"""{key} = {value}\n""" ) # Need to save the state, since Trainer.save_model saves only the tokenizer with the model trainer.state.save_to_json(os.path.join(training_args.output_dir ,'''trainer_state.json''' ) ) # Evaluation __A : List[Any] = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) __A : int = trainer.evaluate() __A : int = math.exp(eval_output['''eval_loss'''] ) __A : int = perplexity __A : Optional[Any] = os.path.join(training_args.output_dir ,'''eval_results_mlm_wwm.txt''' ) if trainer.is_world_process_zero(): with open(lowercase__ ,'''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in sorted(results.items() ): logger.info(F""" {key} = {value}""" ) writer.write(F"""{key} = {value}\n""" ) return results def __lowercase ( snake_case_ : List[str] ) ->Tuple: '''simple docstring''' main() if __name__ == "__main__": main()
362
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available a_ = { """configuration_vivit""": ["""VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """VivitConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ["""VivitImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """VivitModel""", """VivitPreTrainedModel""", """VivitForVideoClassification""", ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
291
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class a__ ( unittest.TestCase ): def __init__( self : List[Any],_A : Any,_A : Union[str, Any]=7,_A : Optional[Any]=3,_A : Optional[int]=10,_A : Optional[int]=18,_A : Optional[Any]=30,_A : Optional[Any]=400,_A : int=True,_A : Dict=None,_A : Any=True,_A : List[Any]=[0.5, 0.5, 0.5],_A : Any=[0.5, 0.5, 0.5],_A : str=None,): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = size if size is not None else {"shortest_edge": 18} SCREAMING_SNAKE_CASE_ : str = crop_size if crop_size is not None else {"height": 18, "width": 18} SCREAMING_SNAKE_CASE_ : List[Any] = parent SCREAMING_SNAKE_CASE_ : Union[str, Any] = batch_size SCREAMING_SNAKE_CASE_ : Any = num_channels SCREAMING_SNAKE_CASE_ : Any = num_frames SCREAMING_SNAKE_CASE_ : Tuple = image_size SCREAMING_SNAKE_CASE_ : Any = min_resolution SCREAMING_SNAKE_CASE_ : Dict = max_resolution SCREAMING_SNAKE_CASE_ : Union[str, Any] = do_resize SCREAMING_SNAKE_CASE_ : int = size SCREAMING_SNAKE_CASE_ : Dict = do_normalize SCREAMING_SNAKE_CASE_ : Optional[Any] = image_mean SCREAMING_SNAKE_CASE_ : Tuple = image_std SCREAMING_SNAKE_CASE_ : List[str] = crop_size def __UpperCamelCase ( self : Optional[Any] ): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class a__ ( A__ , unittest.TestCase ): A = VivitImageProcessor if is_vision_available() else None def __UpperCamelCase ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = VivitImageProcessingTester(self ) @property def __UpperCamelCase ( self : Union[str, Any] ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def __UpperCamelCase ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A,"image_mean" ) ) self.assertTrue(hasattr(_A,"image_std" ) ) self.assertTrue(hasattr(_A,"do_normalize" ) ) self.assertTrue(hasattr(_A,"do_resize" ) ) self.assertTrue(hasattr(_A,"do_center_crop" ) ) self.assertTrue(hasattr(_A,"size" ) ) def __UpperCamelCase ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size,{"shortest_edge": 18} ) self.assertEqual(image_processor.crop_size,{"height": 18, "width": 18} ) SCREAMING_SNAKE_CASE_ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict,size=42,crop_size=84 ) self.assertEqual(image_processor.size,{"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size,{"height": 84, "width": 84} ) def __UpperCamelCase ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos SCREAMING_SNAKE_CASE_ : List[str] = prepare_video_inputs(self.image_processor_tester,equal_resolution=_A ) for video in video_inputs: self.assertIsInstance(_A,_A ) self.assertIsInstance(video[0],Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE_ : int = image_processing(video_inputs[0],return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape,( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ),) # Test batched SCREAMING_SNAKE_CASE_ : Tuple = image_processing(_A,return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape,( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ),) def __UpperCamelCase ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE_ : str = prepare_video_inputs(self.image_processor_tester,equal_resolution=_A,numpify=_A ) for video in video_inputs: self.assertIsInstance(_A,_A ) self.assertIsInstance(video[0],np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE_ : Optional[int] = image_processing(video_inputs[0],return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape,( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ),) # Test batched SCREAMING_SNAKE_CASE_ : List[str] = image_processing(_A,return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape,( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ),) def __UpperCamelCase ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE_ : Dict = prepare_video_inputs(self.image_processor_tester,equal_resolution=_A,torchify=_A ) for video in video_inputs: self.assertIsInstance(_A,_A ) self.assertIsInstance(video[0],torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE_ : List[str] = image_processing(video_inputs[0],return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape,( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ),) # Test batched SCREAMING_SNAKE_CASE_ : Optional[int] = image_processing(_A,return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape,( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ),)
18
def A_ ( snake_case : str ) -> int: '''simple docstring''' assert column_title.isupper() __UpperCamelCase = 0 __UpperCamelCase = len(snake_case ) - 1 __UpperCamelCase = 0 while index >= 0: __UpperCamelCase = (ord(column_title[index] ) - 64) * pow(26 , snake_case ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
328
0
'''simple docstring''' from pathlib import Path import numpy as np from PIL import Image def a_ ( lowerCamelCase : np.ndarray ): lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.2_989 * r + 0.5_870 * g + 0.1_140 * b def a_ ( lowerCamelCase : np.ndarray ): return (gray > 127) & (gray <= 255) def a_ ( lowerCamelCase : np.ndarray , lowerCamelCase : np.ndarray ): lowerCAmelCase = np.zeros_like(lowerCamelCase ) lowerCAmelCase = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image lowerCAmelCase = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): lowerCAmelCase = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() lowerCAmelCase = int(summation > 0 ) return output if __name__ == "__main__": # read original image __snake_case =Path(__file__).resolve().parent / """image_data""" / """lena.jpg""" __snake_case =np.array(Image.open(lena_path)) # kernel to be applied __snake_case =np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) __snake_case =dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image __snake_case =Image.fromarray(output).convert("""RGB""") pil_img.save("""result_dilation.png""")
55
'''simple docstring''' from __future__ import annotations def a_ ( lowerCamelCase : list[float] , lowerCamelCase : list[float] ): lowerCAmelCase = sorted(numsa + numsa ) lowerCAmelCase , lowerCAmelCase = divmod(len(lowerCamelCase ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() __snake_case =[float(x) for x in input("""Enter the elements of first array: """).split()] __snake_case =[float(x) for x in input("""Enter the elements of second array: """).split()] print(F'''The median of two arrays is: {median_of_two_arrays(array_a, array_a)}''')
55
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCAmelCase_ = { 'configuration_vision_encoder_decoder': ['VisionEncoderDecoderConfig', 'VisionEncoderDecoderOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['VisionEncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['TFVisionEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ = ['FlaxVisionEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys UpperCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
12
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union UpperCAmelCase : Union[str, Any] = TypeVar("T") UpperCAmelCase : Dict = Union[List[T], Tuple[T, ...]] UpperCAmelCase : int = Union[T, List[T], Dict[str, T]] UpperCAmelCase : Tuple = Union[str, bytes, os.PathLike]
136
0
"""simple docstring""" import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _A = logging.get_logger('transformers.models.speecht5') def UpperCAmelCase ( a_, a_, a_ ): '''simple docstring''' hf_model.apply_weight_norm() lowerCamelCase : str = checkpoint['input_conv.weight_g'] lowerCamelCase : int = checkpoint['input_conv.weight_v'] lowerCamelCase : Optional[Any] = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): lowerCamelCase : Tuple = checkpoint[F"""upsamples.{i}.1.weight_g"""] lowerCamelCase : Any = checkpoint[F"""upsamples.{i}.1.weight_v"""] lowerCamelCase : List[str] = checkpoint[F"""upsamples.{i}.1.bias"""] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): lowerCamelCase : Optional[Any] = checkpoint[F"""blocks.{i}.convs1.{j}.1.weight_g"""] lowerCamelCase : Tuple = checkpoint[F"""blocks.{i}.convs1.{j}.1.weight_v"""] lowerCamelCase : Union[str, Any] = checkpoint[F"""blocks.{i}.convs1.{j}.1.bias"""] lowerCamelCase : Dict = checkpoint[F"""blocks.{i}.convs2.{j}.1.weight_g"""] lowerCamelCase : str = checkpoint[F"""blocks.{i}.convs2.{j}.1.weight_v"""] lowerCamelCase : Optional[Any] = checkpoint[F"""blocks.{i}.convs2.{j}.1.bias"""] lowerCamelCase : Any = checkpoint['output_conv.1.weight_g'] lowerCamelCase : Tuple = checkpoint['output_conv.1.weight_v'] lowerCamelCase : int = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def UpperCAmelCase ( a_, a_, a_, a_=None, a_=None, ): '''simple docstring''' if config_path is not None: lowerCamelCase : str = SpeechTaHifiGanConfig.from_pretrained(a_ ) else: lowerCamelCase : Dict = SpeechTaHifiGanConfig() lowerCamelCase : int = SpeechTaHifiGan(a_ ) lowerCamelCase : Optional[Any] = torch.load(a_ ) load_weights(orig_checkpoint['model']['generator'], a_, a_ ) lowerCamelCase : Tuple = np.load(a_ ) lowerCamelCase : str = stats[0].reshape(-1 ) lowerCamelCase : Optional[int] = stats[1].reshape(-1 ) lowerCamelCase : Dict = torch.from_numpy(a_ ).float() lowerCamelCase : Optional[int] = torch.from_numpy(a_ ).float() model.save_pretrained(a_ ) if repo_id: print('Pushing to the hub...' ) model.push_to_hub(a_ ) if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--stats_path', required=True, default=None, type=str, help='Path to stats.npy file') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) _A = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
205
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer _A = logging.get_logger(__name__) _A = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} # See all BART models at https://huggingface.co/models?filter=bart _A = { 'vocab_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/vocab.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/vocab.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json', }, 'merges_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/merges.txt', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/merges.txt', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt', }, 'tokenizer_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json', }, } _A = { 'facebook/bart-base': 1_0_2_4, 'facebook/bart-large': 1_0_2_4, 'facebook/bart-large-mnli': 1_0_2_4, 'facebook/bart-large-cnn': 1_0_2_4, 'facebook/bart-large-xsum': 1_0_2_4, 'yjernite/bart_eli5': 1_0_2_4, } class _lowercase ( __UpperCAmelCase ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = ['input_ids', 'attention_mask'] lowercase_ = BartTokenizer def __init__( self , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_="replace" , UpperCAmelCase_="<s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="</s>" , UpperCAmelCase_="<s>" , UpperCAmelCase_="<unk>" , UpperCAmelCase_="<pad>" , UpperCAmelCase_="<mask>" , UpperCAmelCase_=False , UpperCAmelCase_=True , **UpperCAmelCase_ , ) -> Union[str, Any]: super().__init__( UpperCAmelCase_ , UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , errors=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , add_prefix_space=UpperCAmelCase_ , trim_offsets=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCamelCase : List[Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , UpperCAmelCase_ ) != add_prefix_space: lowerCamelCase : Tuple = getattr(UpperCAmelCase_ , pre_tok_state.pop('type' ) ) lowerCamelCase : Optional[Any] = add_prefix_space lowerCamelCase : str = pre_tok_class(**UpperCAmelCase_ ) lowerCamelCase : Optional[Any] = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowerCamelCase : Dict = 'post_processor' lowerCamelCase : str = getattr(self.backend_tokenizer , UpperCAmelCase_ , UpperCAmelCase_ ) if tokenizer_component_instance: lowerCamelCase : Optional[Any] = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCamelCase : int = tuple(state['sep'] ) if "cls" in state: lowerCamelCase : str = tuple(state['cls'] ) lowerCamelCase : Optional[Any] = False if state.get('add_prefix_space' , UpperCAmelCase_ ) != add_prefix_space: lowerCamelCase : Dict = add_prefix_space lowerCamelCase : Tuple = True if state.get('trim_offsets' , UpperCAmelCase_ ) != trim_offsets: lowerCamelCase : Tuple = trim_offsets lowerCamelCase : Dict = True if changes_to_apply: lowerCamelCase : Optional[int] = getattr(UpperCAmelCase_ , state.pop('type' ) ) lowerCamelCase : Any = component_class(**UpperCAmelCase_ ) setattr(self.backend_tokenizer , UpperCAmelCase_ , UpperCAmelCase_ ) @property def _UpperCamelCase ( self ) -> str: if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def _UpperCamelCase ( self , UpperCAmelCase_ ) -> List[Any]: lowerCamelCase : Optional[int] = AddedToken(UpperCAmelCase_ , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else value lowerCamelCase : int = value def _UpperCamelCase ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ) -> BatchEncoding: lowerCamelCase : str = kwargs.get('is_split_into_words' , UpperCAmelCase_ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ 'to use it with pretokenized inputs.' ) return super()._batch_encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def _UpperCamelCase ( self , *UpperCAmelCase_ , **UpperCAmelCase_ ) -> BatchEncoding: lowerCamelCase : Optional[Any] = kwargs.get('is_split_into_words' , UpperCAmelCase_ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ 'to use it with pretokenized inputs.' ) return super()._encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def _UpperCamelCase ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ) -> Tuple[str]: lowerCamelCase : Any = self._tokenizer.model.save(UpperCAmelCase_ , name=UpperCAmelCase_ ) return tuple(UpperCAmelCase_ ) def _UpperCamelCase ( self , UpperCAmelCase_ , UpperCAmelCase_=None ) -> List[Any]: lowerCamelCase : Union[str, Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _UpperCamelCase ( self , UpperCAmelCase_ , UpperCAmelCase_ = None ) -> List[int]: lowerCamelCase : List[Any] = [self.sep_token_id] lowerCamelCase : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
205
1