code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class UpperCamelCase_ ( UpperCamelCase , unittest.TestCase): """simple docstring""" snake_case__ : Dict = BioGptTokenizer snake_case__ : List[str] = False def UpperCAmelCase_ ( self : Dict ) -> int: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __SCREAMING_SNAKE_CASE = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "w</w>", "r</w>", "t</w>", "lo", "low", "er</w>", "low</w>", "lowest</w>", "newer</w>", "wider</w>", "<unk>", ] __SCREAMING_SNAKE_CASE = dict(zip(UpperCAmelCase__ , range(len(UpperCAmelCase__ ) ) ) ) __SCREAMING_SNAKE_CASE = ["l o 123", "lo w 1456", "e r</w> 1789", ""] __SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) __SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" ) as fp: fp.write(json.dumps(UpperCAmelCase__ ) ) with open(self.merges_file , "w" ) as fp: fp.write("\n".join(UpperCAmelCase__ ) ) def UpperCAmelCase_ ( self : Any , UpperCAmelCase__ : str ) -> Dict: __SCREAMING_SNAKE_CASE = "lower newer" __SCREAMING_SNAKE_CASE = "lower newer" return input_text, output_text def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = BioGptTokenizer(self.vocab_file , self.merges_file ) __SCREAMING_SNAKE_CASE = "lower" __SCREAMING_SNAKE_CASE = ["low", "er</w>"] __SCREAMING_SNAKE_CASE = tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokens + ["<unk>"] __SCREAMING_SNAKE_CASE = [1_4, 1_5, 2_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , UpperCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Dict ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = BioGptTokenizer.from_pretrained("microsoft/biogpt" ) __SCREAMING_SNAKE_CASE = tokenizer.encode("sequence builders" , add_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.encode("multi-sequence build" , add_special_tokens=UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
54
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase : Tuple = {"configuration_plbart": ["PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP", "PLBartConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[Any] = ["PLBartTokenizer"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Dict = [ "PLBART_PRETRAINED_MODEL_ARCHIVE_LIST", "PLBartForCausalLM", "PLBartForConditionalGeneration", "PLBartForSequenceClassification", "PLBartModel", "PLBartPreTrainedModel", ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys UpperCAmelCase : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure)
136
0
'''simple docstring''' from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
83
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class snake_case__ : """simple docstring""" def __init__( self : Dict , UpperCamelCase__ : int , UpperCamelCase__ : MutableSequence[float] ) -> None: """simple docstring""" if len(UpperCamelCase__ ) != degree + 1: raise ValueError( '''The number of coefficients should be equal to the degree + 1.''' ) snake_case : list[float] = list(UpperCamelCase__ ) snake_case : int = degree def __add__( self : int , UpperCamelCase__ : Polynomial ) -> Polynomial: """simple docstring""" if self.degree > polynomial_a.degree: snake_case : Tuple = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , UpperCamelCase__ ) else: snake_case : List[Any] = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , UpperCamelCase__ ) def __sub__( self : Tuple , UpperCamelCase__ : Polynomial ) -> Polynomial: """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : int ) -> Polynomial: """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Union[str, Any] , UpperCamelCase__ : Polynomial ) -> Polynomial: """simple docstring""" snake_case : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , UpperCamelCase__ ) def lowerCAmelCase ( self : List[str] , UpperCamelCase__ : int | float ) -> int | float: """simple docstring""" snake_case : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Dict ) -> str: """simple docstring""" snake_case : List[Any] = '''''' for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(UpperCamelCase__ ) return polynomial def __repr__( self : List[str] ) -> str: """simple docstring""" return self.__str__() def lowerCAmelCase ( self : Any ) -> Polynomial: """simple docstring""" snake_case : list[float] = [0] * self.degree for i in range(self.degree ): snake_case : Dict = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , UpperCamelCase__ ) def lowerCAmelCase ( self : int , UpperCamelCase__ : int | float = 0 ) -> Polynomial: """simple docstring""" snake_case : list[float] = [0] * (self.degree + 2) snake_case : Union[str, Any] = constant for i in range(self.degree + 1 ): snake_case : Optional[int] = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , UpperCamelCase__ ) def __eq__( self : Any , UpperCamelCase__ : object ) -> bool: """simple docstring""" if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : Union[str, Any] , UpperCamelCase__ : object ) -> bool: """simple docstring""" return not self.__eq__(UpperCamelCase__ )
83
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { "facebook/data2vec-text-base": "https://huggingface.co/data2vec/resolve/main/config.json", } class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = 'data2vec-text' def __init__( self , _a=30_522 , _a=768 , _a=12 , _a=12 , _a=3_072 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=2 , _a=0.02 , _a=1E-12 , _a=1 , _a=0 , _a=2 , _a="absolute" , _a=True , _a=None , **_a , ): super().__init__(pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a ) __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = hidden_act __a = intermediate_size __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = max_position_embeddings __a = type_vocab_size __a = initializer_range __a = layer_norm_eps __a = position_embedding_type __a = use_cache __a = classifier_dropout class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def __UpperCAmelCase ( self ): if self.task == "multiple-choice": __a = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: __a = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
45
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available A: Optional[Any] = { "configuration_tapas": ["TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP", "TapasConfig"], "tokenization_tapas": ["TapasTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A: Optional[int] = [ "TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST", "TapasForMaskedLM", "TapasForQuestionAnswering", "TapasForSequenceClassification", "TapasModel", "TapasPreTrainedModel", "load_tf_weights_in_tapas", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A: List[str] = [ "TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST", "TFTapasForMaskedLM", "TFTapasForQuestionAnswering", "TFTapasForSequenceClassification", "TFTapasModel", "TFTapasPreTrainedModel", ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys A: Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
109
0
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @property def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' __lowerCamelCase = self.dummy_uncond_unet __lowerCamelCase = KarrasVeScheduler() __lowerCamelCase = KarrasVePipeline(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe(num_inference_steps=2 , generator=lowerCamelCase__ , output_type='numpy' ).images __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe(num_inference_steps=2 , generator=lowerCamelCase__ , output_type='numpy' , return_dict=lowerCamelCase__ )[0] __lowerCamelCase = image[0, -3:, -3:, -1] __lowerCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self ) -> int: '''simple docstring''' __lowerCamelCase = 'google/ncsnpp-celebahq-256' __lowerCamelCase = UNetaDModel.from_pretrained(lowerCamelCase__ ) __lowerCamelCase = KarrasVeScheduler() __lowerCamelCase = KarrasVePipeline(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe(num_inference_steps=20 , generator=lowerCamelCase__ , output_type='numpy' ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) __lowerCamelCase = np.array([0.5_78, 0.58_11, 0.59_24, 0.58_09, 0.5_87, 0.58_86, 0.58_61, 0.58_02, 0.5_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
348
import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __A = { "facebook/mask2former-swin-small-coco-instance": ( "https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json" ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } __A = logging.get_logger(__name__) class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" snake_case_ = '''mask2former''' snake_case_ = ['''swin'''] snake_case_ = {'''hidden_size''': '''hidden_dim'''} def __init__( self , lowerCamelCase__ = None , lowerCamelCase__ = 256 , lowerCamelCase__ = 256 , lowerCamelCase__ = 256 , lowerCamelCase__ = 1_024 , lowerCamelCase__ = "relu" , lowerCamelCase__ = 6 , lowerCamelCase__ = 10 , lowerCamelCase__ = 8 , lowerCamelCase__ = 0.0 , lowerCamelCase__ = 2_048 , lowerCamelCase__ = False , lowerCamelCase__ = False , lowerCamelCase__ = 4 , lowerCamelCase__ = 255 , lowerCamelCase__ = 100 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 2.0 , lowerCamelCase__ = 5.0 , lowerCamelCase__ = 5.0 , lowerCamelCase__ = 12_544 , lowerCamelCase__ = 3.0 , lowerCamelCase__ = 0.75 , lowerCamelCase__ = 0.02 , lowerCamelCase__ = 1.0 , lowerCamelCase__ = True , lowerCamelCase__ = [4, 8, 16, 32] , lowerCamelCase__ = None , **lowerCamelCase__ , ) -> Tuple: '''simple docstring''' if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `Swin` backbone.' ) __lowerCamelCase = CONFIG_MAPPING['swin']( image_size=224 , in_channels=3 , patch_size=4 , embed_dim=96 , depths=[2, 2, 18, 2] , num_heads=[3, 6, 12, 24] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=lowerCamelCase__ , out_features=['stage1', 'stage2', 'stage3', 'stage4'] , ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ): __lowerCamelCase = backbone_config.pop('model_type' ) __lowerCamelCase = CONFIG_MAPPING[backbone_model_type] __lowerCamelCase = config_class.from_dict(lowerCamelCase__ ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( f"""Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. """ f"""Supported model types: {','.join(self.backbones_supported )}""" ) __lowerCamelCase = backbone_config __lowerCamelCase = feature_size __lowerCamelCase = mask_feature_size __lowerCamelCase = hidden_dim __lowerCamelCase = encoder_feedforward_dim __lowerCamelCase = activation_function __lowerCamelCase = encoder_layers __lowerCamelCase = decoder_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = dropout __lowerCamelCase = dim_feedforward __lowerCamelCase = pre_norm __lowerCamelCase = enforce_input_projection __lowerCamelCase = common_stride __lowerCamelCase = ignore_value __lowerCamelCase = num_queries __lowerCamelCase = no_object_weight __lowerCamelCase = class_weight __lowerCamelCase = mask_weight __lowerCamelCase = dice_weight __lowerCamelCase = train_num_points __lowerCamelCase = oversample_ratio __lowerCamelCase = importance_sample_ratio __lowerCamelCase = init_std __lowerCamelCase = init_xavier_std __lowerCamelCase = use_auxiliary_loss __lowerCamelCase = feature_strides __lowerCamelCase = output_auxiliary_logits __lowerCamelCase = decoder_layers super().__init__(**lowerCamelCase__ ) @classmethod def lowercase_ ( cls , lowerCamelCase__ , **lowerCamelCase__ ) -> Any: '''simple docstring''' return cls( backbone_config=lowerCamelCase__ , **lowerCamelCase__ , ) def lowercase_ ( self ) -> Dict[str, any]: '''simple docstring''' __lowerCamelCase = copy.deepcopy(self.__dict__ ) __lowerCamelCase = self.backbone_config.to_dict() __lowerCamelCase = self.__class__.model_type return output
348
1
'''simple docstring''' from functools import lru_cache @lru_cache def __snake_case( _lowerCAmelCase ) -> Any: if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
35
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ = logging.get_logger(__name__) snake_case_ = { """transfo-xl-wt103""": """https://huggingface.co/transfo-xl-wt103/resolve/main/config.json""", } class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __UpperCamelCase = """transfo-xl""" __UpperCamelCase = ["""mems"""] __UpperCamelCase = { """n_token""": """vocab_size""", """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self :List[Any] , lowercase_ :Optional[int]=26_77_35 , lowercase_ :Union[str, Any]=[2_00_00, 4_00_00, 20_00_00] , lowercase_ :List[Any]=10_24 , lowercase_ :Optional[Any]=10_24 , lowercase_ :Tuple=16 , lowercase_ :Tuple=64 , lowercase_ :Any=40_96 , lowercase_ :int=4 , lowercase_ :List[str]=False , lowercase_ :Union[str, Any]=18 , lowercase_ :Optional[Any]=16_00 , lowercase_ :Dict=10_00 , lowercase_ :Optional[int]=True , lowercase_ :Tuple=True , lowercase_ :Dict=0 , lowercase_ :Tuple=-1 , lowercase_ :Optional[int]=True , lowercase_ :Optional[int]=0.1 , lowercase_ :str=0.0 , lowercase_ :List[str]=True , lowercase_ :int="normal" , lowercase_ :Dict=0.01 , lowercase_ :Optional[Any]=0.01 , lowercase_ :Dict=0.02 , lowercase_ :Tuple=1E-5 , lowercase_ :str=0 , **lowercase_ :Tuple , ) -> List[str]: UpperCAmelCase = vocab_size UpperCAmelCase = [] self.cutoffs.extend(lowercase_ ) if proj_share_all_but_first: UpperCAmelCase = [False] + [True] * len(self.cutoffs ) else: UpperCAmelCase = [False] + [False] * len(self.cutoffs ) UpperCAmelCase = d_model UpperCAmelCase = d_embed UpperCAmelCase = d_head UpperCAmelCase = d_inner UpperCAmelCase = div_val UpperCAmelCase = pre_lnorm UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = mem_len UpperCAmelCase = same_length UpperCAmelCase = attn_type UpperCAmelCase = clamp_len UpperCAmelCase = sample_softmax UpperCAmelCase = adaptive UpperCAmelCase = dropout UpperCAmelCase = dropatt UpperCAmelCase = untie_r UpperCAmelCase = init UpperCAmelCase = init_range UpperCAmelCase = proj_init_std UpperCAmelCase = init_std UpperCAmelCase = layer_norm_epsilon super().__init__(eos_token_id=lowercase_ , **lowercase_ ) @property def UpperCAmelCase__ ( self :Union[str, Any] ) -> Any: # Message copied from Transformer-XL documentation logger.info(f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def UpperCAmelCase__ ( self :Union[str, Any] , lowercase_ :Any ) -> Tuple: # Message copied from Transformer-XL documentation raise NotImplementedError( f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
78
0
class lowercase : def __init__( self ,A__): lowercase = len(A__) lowercase = [0] * len_array if len_array > 0: lowercase = array[0] for i in range(1 ,A__): lowercase = self.prefix_sum[i - 1] + array[i] def A__ ( self ,A__ ,A__): if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def A__ ( self ,A__): lowercase = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(A__) return False if __name__ == "__main__": import doctest doctest.testmod()
97
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available from ...utils import OptionalDependencyNotAvailable lowercase__ :Union[str, Any] = {"configuration_dpt": ["DPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DPTConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[Any] = ["DPTFeatureExtractor"] lowercase__ :List[Any] = ["DPTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Optional[int] = [ "DPT_PRETRAINED_MODEL_ARCHIVE_LIST", "DPTForDepthEstimation", "DPTForSemanticSegmentation", "DPTModel", "DPTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_dpt import DPT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPTConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_dpt import DPTFeatureExtractor from .image_processing_dpt import DPTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dpt import ( DPT_PRETRAINED_MODEL_ARCHIVE_LIST, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel, DPTPreTrainedModel, ) else: import sys lowercase__ :List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
97
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available A__ : Optional[int] ={ """configuration_gpt_neo""": ["""GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoConfig""", """GPTNeoOnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : str =[ """GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTNeoForCausalLM""", """GPTNeoForQuestionAnswering""", """GPTNeoForSequenceClassification""", """GPTNeoForTokenClassification""", """GPTNeoModel""", """GPTNeoPreTrainedModel""", """load_tf_weights_in_gpt_neo""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : List[Any] =[ """FlaxGPTNeoForCausalLM""", """FlaxGPTNeoModel""", """FlaxGPTNeoPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys A__ : Optional[Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
70
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) a_ : Union[str, Any] = { """configuration_funnel""": ["""FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FunnelConfig"""], """convert_funnel_original_tf_checkpoint_to_pytorch""": [], """tokenization_funnel""": ["""FunnelTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : List[str] = ["""FunnelTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Optional[int] = [ """FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST""", """FunnelBaseModel""", """FunnelForMaskedLM""", """FunnelForMultipleChoice""", """FunnelForPreTraining""", """FunnelForQuestionAnswering""", """FunnelForSequenceClassification""", """FunnelForTokenClassification""", """FunnelModel""", """FunnelPreTrainedModel""", """load_tf_weights_in_funnel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Optional[Any] = [ """TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFFunnelBaseModel""", """TFFunnelForMaskedLM""", """TFFunnelForMultipleChoice""", """TFFunnelForPreTraining""", """TFFunnelForQuestionAnswering""", """TFFunnelForSequenceClassification""", """TFFunnelForTokenClassification""", """TFFunnelModel""", """TFFunnelPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_funnel import FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP, FunnelConfig from .tokenization_funnel import FunnelTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_funnel_fast import FunnelTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_funnel import ( FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, FunnelBaseModel, FunnelForMaskedLM, FunnelForMultipleChoice, FunnelForPreTraining, FunnelForQuestionAnswering, FunnelForSequenceClassification, FunnelForTokenClassification, FunnelModel, FunnelPreTrainedModel, load_tf_weights_in_funnel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_funnel import ( TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, TFFunnelPreTrainedModel, ) else: import sys a_ : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
75
0
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCamelCase : def __init__( self: str , _SCREAMING_SNAKE_CASE: Dict , _SCREAMING_SNAKE_CASE: Optional[Any]=3 , _SCREAMING_SNAKE_CASE: Tuple=32 , _SCREAMING_SNAKE_CASE: Any=3 , _SCREAMING_SNAKE_CASE: Optional[Any]=10 , _SCREAMING_SNAKE_CASE: str=[10, 20, 30, 40] , _SCREAMING_SNAKE_CASE: int=[1, 1, 2, 1] , _SCREAMING_SNAKE_CASE: List[str]=True , _SCREAMING_SNAKE_CASE: Optional[int]=True , _SCREAMING_SNAKE_CASE: List[str]="relu" , _SCREAMING_SNAKE_CASE: List[Any]=3 , _SCREAMING_SNAKE_CASE: int=None , ) -> int: """simple docstring""" UpperCamelCase_ = parent UpperCamelCase_ = batch_size UpperCamelCase_ = image_size UpperCamelCase_ = num_channels UpperCamelCase_ = embeddings_size UpperCamelCase_ = hidden_sizes UpperCamelCase_ = depths UpperCamelCase_ = is_training UpperCamelCase_ = use_labels UpperCamelCase_ = hidden_act UpperCamelCase_ = num_labels UpperCamelCase_ = scope UpperCamelCase_ = len(_SCREAMING_SNAKE_CASE ) def lowercase ( self: Any ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase_ = None if self.use_labels: UpperCamelCase_ = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase_ = self.get_config() return config, pixel_values, labels def lowercase ( self: Dict ) -> int: """simple docstring""" return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def lowercase ( self: List[str] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: Optional[Any] ) -> List[str]: """simple docstring""" UpperCamelCase_ = TFResNetModel(config=_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = model(_SCREAMING_SNAKE_CASE ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowercase ( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: List[str] ) -> Dict: """simple docstring""" UpperCamelCase_ = self.num_labels UpperCamelCase_ = TFResNetForImageClassification(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase ( self: Optional[Any] ) -> Tuple: """simple docstring""" UpperCamelCase_ = self.prepare_config_and_inputs() UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = config_and_inputs UpperCamelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): _UpperCamelCase : Union[str, Any] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () _UpperCamelCase : Union[str, Any] = ( {'''feature-extraction''': TFResNetModel, '''image-classification''': TFResNetForImageClassification} if is_tf_available() else {} ) _UpperCamelCase : Optional[int] = False _UpperCamelCase : List[str] = False _UpperCamelCase : int = False _UpperCamelCase : str = False _UpperCamelCase : List[str] = False def lowercase ( self: str ) -> str: """simple docstring""" UpperCamelCase_ = TFResNetModelTester(self ) UpperCamelCase_ = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE ) def lowercase ( self: str ) -> Optional[Any]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase ( self: List[Any] ) -> Union[str, Any]: """simple docstring""" return @unittest.skip(reason="ResNet does not use inputs_embeds" ) def lowercase ( self: List[Any] ) -> str: """simple docstring""" pass @unittest.skip(reason="ResNet does not support input and output embeddings" ) def lowercase ( self: Any ) -> Any: """simple docstring""" pass def lowercase ( self: Union[str, Any] ) -> Tuple: """simple docstring""" UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase_ = model_class(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase_ = [*signature.parameters.keys()] UpperCamelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def lowercase ( self: List[Any] ) -> Dict: """simple docstring""" UpperCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def lowercase ( self: Any ) -> Optional[int]: """simple docstring""" def check_hidden_states_output(_SCREAMING_SNAKE_CASE: Optional[int] , _SCREAMING_SNAKE_CASE: str , _SCREAMING_SNAKE_CASE: str ): UpperCamelCase_ = model_class(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) UpperCamelCase_ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCamelCase_ = self.model_tester.num_stages self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase_ = ["basic", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: UpperCamelCase_ = layer_type UpperCamelCase_ = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase_ = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def lowercase ( self: Optional[int] ) -> Dict: """simple docstring""" UpperCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @slow def lowercase ( self: Tuple ) -> Tuple: """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase_ = TFResNetModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( ) -> str: UpperCamelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class _UpperCamelCase ( unittest.TestCase ): @cached_property def lowercase ( self: str ) -> List[str]: """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowercase ( self: Union[str, Any] ) -> str: """simple docstring""" UpperCamelCase_ = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCamelCase_ = self.default_image_processor UpperCamelCase_ = prepare_img() UpperCamelCase_ = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="tf" ) # forward pass UpperCamelCase_ = model(**_SCREAMING_SNAKE_CASE ) # verify the logits UpperCamelCase_ = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) UpperCamelCase_ = tf.constant([-11.10_69, -9.78_77, -8.37_77] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , _SCREAMING_SNAKE_CASE , atol=1e-4 ) )
357
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): @register_to_config def __init__( self: Any , _SCREAMING_SNAKE_CASE: int = 768 , ) -> Tuple: """simple docstring""" super().__init__() UpperCamelCase_ = nn.Parameter(torch.zeros(1 , _SCREAMING_SNAKE_CASE ) ) UpperCamelCase_ = nn.Parameter(torch.ones(1 , _SCREAMING_SNAKE_CASE ) ) def lowercase ( self: List[Any] , _SCREAMING_SNAKE_CASE: Optional[Union[str, torch.device]] = None , _SCREAMING_SNAKE_CASE: Optional[torch.dtype] = None , ) -> List[Any]: """simple docstring""" UpperCamelCase_ = nn.Parameter(self.mean.to(_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) ) UpperCamelCase_ = nn.Parameter(self.std.to(_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) ) return self def lowercase ( self: str , _SCREAMING_SNAKE_CASE: Dict ) -> Optional[int]: """simple docstring""" UpperCamelCase_ = (embeds - self.mean) * 1.0 / self.std return embeds def lowercase ( self: List[Any] , _SCREAMING_SNAKE_CASE: Union[str, Any] ) -> Dict: """simple docstring""" UpperCamelCase_ = (embeds * self.std) + self.mean return embeds
328
0
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def a ( snake_case__: str = "" ): '''simple docstring''' lowercase_ = url or '''https://www.imdb.com/chart/top/?ref_=nv_mv_250''' lowercase_ = BeautifulSoup(requests.get(snake_case_ ).text , '''html.parser''' ) lowercase_ = soup.find_all('''td''' , attrs='''titleColumn''' ) lowercase_ = soup.find_all('''td''' , class_='''ratingColumn imdbRating''' ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(snake_case_ , snake_case_ ) } def a ( snake_case__: str = "IMDb_Top_250_Movies.csv" ): '''simple docstring''' lowercase_ = get_imdb_top_aaa_movies() with open(snake_case_ , '''w''' , newline='''''' ) as out_file: lowercase_ = csv.writer(snake_case_ ) writer.writerow(['''Movie title''', '''IMDb rating'''] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
30
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) snake_case_ = logging.getLogger(__name__) @dataclass(frozen=_UpperCAmelCase ) class SCREAMING_SNAKE_CASE__ : A_ : str A_ : str A_ : Optional[str] = None A_ : Optional[str] = None A_ : Optional[str] = None @dataclass(frozen=_UpperCAmelCase ) class SCREAMING_SNAKE_CASE__ : A_ : List[int] A_ : Optional[List[int]] = None A_ : Optional[List[int]] = None A_ : Optional[Union[int, float]] = None A_ : Optional[int] = None if is_torch_available(): import torch from torch.utils.data import Dataset class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): A_ : List[InputFeatures] def __init__(self : int , a__ : str , a__ : PreTrainedTokenizer , a__ : str , a__ : Optional[int] = None , a__ : List[Any]=False , a__ : bool = False , ): """simple docstring""" __snake_case = hans_processors[task]() __snake_case = os.path.join( a__ , '''cached_{}_{}_{}_{}'''.format( '''dev''' if evaluate else '''train''' , tokenizer.__class__.__name__ , str(a__ ) , a__ , ) , ) __snake_case = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) __snake_case , __snake_case = label_list[2], label_list[1] __snake_case = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __snake_case = cached_features_file + '''.lock''' with FileLock(a__ ): if os.path.exists(a__ ) and not overwrite_cache: logger.info(f"""Loading features from cached file {cached_features_file}""" ) __snake_case = torch.load(a__ ) else: logger.info(f"""Creating features from dataset file at {data_dir}""" ) __snake_case = ( processor.get_dev_examples(a__ ) if evaluate else processor.get_train_examples(a__ ) ) logger.info('''Training examples: %s''' , len(a__ ) ) __snake_case = hans_convert_examples_to_features(a__ , a__ , a__ , a__ ) logger.info('''Saving features into cached file %s''' , a__ ) torch.save(self.features , a__ ) def __len__(self : int ): """simple docstring""" return len(self.features ) def __getitem__(self : Dict , a__ : List[Any] ): """simple docstring""" return self.features[i] def a (self : List[Any] ): """simple docstring""" return self.label_list if is_tf_available(): import tensorflow as tf class SCREAMING_SNAKE_CASE__ : A_ : List[InputFeatures] def __init__(self : Tuple , a__ : str , a__ : PreTrainedTokenizer , a__ : str , a__ : Optional[int] = 128 , a__ : Any=False , a__ : bool = False , ): """simple docstring""" __snake_case = hans_processors[task]() __snake_case = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) __snake_case , __snake_case = label_list[2], label_list[1] __snake_case = label_list __snake_case = processor.get_dev_examples(a__ ) if evaluate else processor.get_train_examples(a__ ) __snake_case = hans_convert_examples_to_features(a__ , a__ , a__ , a__ ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc='''convert examples to features''' ): if ex_index % 1_0000 == 0: logger.info('''Writing example %d of %d''' % (ex_index, len(a__ )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) __snake_case = tf.data.Dataset.from_generator( a__ , ( { '''example_id''': tf.intaa, '''input_ids''': tf.intaa, '''attention_mask''': tf.intaa, '''token_type_ids''': tf.intaa, }, tf.intaa, ) , ( { '''example_id''': tf.TensorShape([] ), '''input_ids''': tf.TensorShape([None, None] ), '''attention_mask''': tf.TensorShape([None, None] ), '''token_type_ids''': tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def a (self : Union[str, Any] ): """simple docstring""" return self.dataset def __len__(self : Dict ): """simple docstring""" return len(self.features ) def __getitem__(self : Any , a__ : Dict ): """simple docstring""" return self.features[i] def a (self : str ): """simple docstring""" return self.label_list class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): def a (self : Dict , a__ : Dict ): """simple docstring""" return self._create_examples(self._read_tsv(os.path.join(a__ , '''heuristics_train_set.txt''' ) ) , '''train''' ) def a (self : Optional[int] , a__ : Tuple ): """simple docstring""" return self._create_examples(self._read_tsv(os.path.join(a__ , '''heuristics_evaluation_set.txt''' ) ) , '''dev''' ) def a (self : int ): """simple docstring""" return ["contradiction", "entailment", "neutral"] def a (self : Any , a__ : Optional[int] , a__ : List[Any] ): """simple docstring""" __snake_case = [] for i, line in enumerate(a__ ): if i == 0: continue __snake_case = '''%s-%s''' % (set_type, line[0]) __snake_case = line[5] __snake_case = line[6] __snake_case = line[7][2:] if line[7].startswith('''ex''' ) else line[7] __snake_case = line[0] examples.append(InputExample(guid=a__ , text_a=a__ , text_b=a__ , label=a__ , pairID=a__ ) ) return examples def lowerCamelCase__ ( snake_case_ : List[InputExample] , snake_case_ : List[str] , snake_case_ : int , snake_case_ : PreTrainedTokenizer , ) -> List[str]: __snake_case = {label: i for i, label in enumerate(snake_case_ )} __snake_case = [] for ex_index, example in tqdm.tqdm(enumerate(snake_case_ ) , desc='''convert examples to features''' ): if ex_index % 1_0000 == 0: logger.info('''Writing example %d''' % (ex_index) ) __snake_case = tokenizer( example.text_a , example.text_b , add_special_tokens=snake_case_ , max_length=snake_case_ , padding='''max_length''' , truncation=snake_case_ , return_overflowing_tokens=snake_case_ , ) __snake_case = label_map[example.label] if example.label in label_map else 0 __snake_case = int(example.pairID ) features.append(InputFeatures(**snake_case_ , label=snake_case_ , pairID=snake_case_ ) ) for i, example in enumerate(examples[:5] ): logger.info('''*** Example ***''' ) logger.info(f"""guid: {example}""" ) logger.info(f"""features: {features[i]}""" ) return features snake_case_ = { 'hans': 3, } snake_case_ = { 'hans': HansProcessor, }
24
0
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, BertPooler, BertPreTrainedModel, ) def a__ ( a__ ): """simple docstring""" __SCREAMING_SNAKE_CASE = torch.exp(a__ ) __SCREAMING_SNAKE_CASE = torch.sum(a__ , dim=1 ) # sum of exp(x_i) __SCREAMING_SNAKE_CASE = torch.sum(x * exp_x , dim=1 ) # sum of x_i * exp(x_i) return torch.log(a__ ) - B / A class lowerCAmelCase__ ( nn.Module ): """simple docstring""" def __init__( self : Dict , __SCREAMING_SNAKE_CASE : Optional[int] ) -> str: """simple docstring""" super().__init__() __SCREAMING_SNAKE_CASE = config.output_attentions __SCREAMING_SNAKE_CASE = config.output_hidden_states __SCREAMING_SNAKE_CASE = nn.ModuleList([BertLayer(__SCREAMING_SNAKE_CASE ) for _ in range(config.num_hidden_layers )] ) __SCREAMING_SNAKE_CASE = nn.ModuleList([BertHighway(__SCREAMING_SNAKE_CASE ) for _ in range(config.num_hidden_layers )] ) __SCREAMING_SNAKE_CASE = [-1 for _ in range(config.num_hidden_layers )] def UpperCAmelCase__ ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> int: """simple docstring""" if (type(__SCREAMING_SNAKE_CASE ) is float) or (type(__SCREAMING_SNAKE_CASE ) is int): for i in range(len(self.early_exit_entropy ) ): __SCREAMING_SNAKE_CASE = x else: __SCREAMING_SNAKE_CASE = x def UpperCAmelCase__ ( self : Tuple , __SCREAMING_SNAKE_CASE : Dict ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = pooler.state_dict() for highway in self.highway: for name, param in highway.pooler.state_dict().items(): param.copy_(loaded_model[name] ) def UpperCAmelCase__ ( self : Any , __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : str=None , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Any=None , ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = () __SCREAMING_SNAKE_CASE = () __SCREAMING_SNAKE_CASE = () for i, layer_module in enumerate(self.layer ): if self.output_hidden_states: __SCREAMING_SNAKE_CASE = all_hidden_states + (hidden_states,) __SCREAMING_SNAKE_CASE = layer_module( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , head_mask[i] , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = layer_outputs[0] if self.output_attentions: __SCREAMING_SNAKE_CASE = all_attentions + (layer_outputs[1],) __SCREAMING_SNAKE_CASE = (hidden_states,) if self.output_hidden_states: __SCREAMING_SNAKE_CASE = current_outputs + (all_hidden_states,) if self.output_attentions: __SCREAMING_SNAKE_CASE = current_outputs + (all_attentions,) __SCREAMING_SNAKE_CASE = self.highway[i](__SCREAMING_SNAKE_CASE ) # logits, pooled_output if not self.training: __SCREAMING_SNAKE_CASE = highway_exit[0] __SCREAMING_SNAKE_CASE = entropy(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = highway_exit + (highway_entropy,) # logits, hidden_states(?), entropy __SCREAMING_SNAKE_CASE = all_highway_exits + (highway_exit,) if highway_entropy < self.early_exit_entropy[i]: __SCREAMING_SNAKE_CASE = (highway_logits,) + current_outputs[1:] + (all_highway_exits,) raise HighwayException(__SCREAMING_SNAKE_CASE , i + 1 ) else: __SCREAMING_SNAKE_CASE = all_highway_exits + (highway_exit,) # Add last layer if self.output_hidden_states: __SCREAMING_SNAKE_CASE = all_hidden_states + (hidden_states,) __SCREAMING_SNAKE_CASE = (hidden_states,) if self.output_hidden_states: __SCREAMING_SNAKE_CASE = outputs + (all_hidden_states,) if self.output_attentions: __SCREAMING_SNAKE_CASE = outputs + (all_attentions,) __SCREAMING_SNAKE_CASE = outputs + (all_highway_exits,) return outputs # last-layer hidden state, (all hidden states), (all attentions), all highway exits @add_start_docstrings( "The Bert Model transformer with early exiting (DeeBERT). " , a , ) class lowerCAmelCase__ ( a ): """simple docstring""" def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : str ) -> Dict: """simple docstring""" super().__init__(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = config __SCREAMING_SNAKE_CASE = BertEmbeddings(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = DeeBertEncoder(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = BertPooler(__SCREAMING_SNAKE_CASE ) self.init_weights() def UpperCAmelCase__ ( self : Dict ) -> Optional[Any]: """simple docstring""" self.encoder.init_highway_pooler(self.pooler ) def UpperCAmelCase__ ( self : Tuple ) -> Tuple: """simple docstring""" return self.embeddings.word_embeddings def UpperCAmelCase__ ( self : int , __SCREAMING_SNAKE_CASE : int ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE = value def UpperCAmelCase__ ( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> List[str]: """simple docstring""" for layer, heads in heads_to_prune.items(): self.encoder.layer[layer].attention.prune_heads(__SCREAMING_SNAKE_CASE ) @add_start_docstrings_to_model_forward(__SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self : List[str] , __SCREAMING_SNAKE_CASE : Union[str, Any]=None , __SCREAMING_SNAKE_CASE : List[Any]=None , __SCREAMING_SNAKE_CASE : Optional[Any]=None , __SCREAMING_SNAKE_CASE : Union[str, Any]=None , __SCREAMING_SNAKE_CASE : Optional[int]=None , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : List[str]=None , ) -> Optional[int]: """simple docstring""" if input_ids is not None and inputs_embeds is not None: raise ValueError("""You cannot specify both input_ids and inputs_embeds at the same time""" ) elif input_ids is not None: __SCREAMING_SNAKE_CASE = input_ids.size() elif inputs_embeds is not None: __SCREAMING_SNAKE_CASE = inputs_embeds.size()[:-1] else: raise ValueError("""You have to specify either input_ids or inputs_embeds""" ) __SCREAMING_SNAKE_CASE = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: __SCREAMING_SNAKE_CASE = torch.ones(__SCREAMING_SNAKE_CASE , device=__SCREAMING_SNAKE_CASE ) if encoder_attention_mask is None: __SCREAMING_SNAKE_CASE = torch.ones(__SCREAMING_SNAKE_CASE , device=__SCREAMING_SNAKE_CASE ) if token_type_ids is None: __SCREAMING_SNAKE_CASE = torch.zeros(__SCREAMING_SNAKE_CASE , dtype=torch.long , device=__SCREAMING_SNAKE_CASE ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. __SCREAMING_SNAKE_CASE = self.get_extended_attention_mask(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if encoder_attention_mask.dim() == 3: __SCREAMING_SNAKE_CASE = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.dim() == 2: __SCREAMING_SNAKE_CASE = encoder_attention_mask[:, None, None, :] __SCREAMING_SNAKE_CASE = encoder_extended_attention_mask.to( dtype=next(self.parameters() ).dtype ) # fp16 compatibility __SCREAMING_SNAKE_CASE = (1.0 - encoder_extended_attention_mask) * -10000.0 # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] __SCREAMING_SNAKE_CASE = self.get_head_mask(__SCREAMING_SNAKE_CASE , self.config.num_hidden_layers ) __SCREAMING_SNAKE_CASE = self.embeddings( input_ids=__SCREAMING_SNAKE_CASE , position_ids=__SCREAMING_SNAKE_CASE , token_type_ids=__SCREAMING_SNAKE_CASE , inputs_embeds=__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = self.encoder( __SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , head_mask=__SCREAMING_SNAKE_CASE , encoder_hidden_states=__SCREAMING_SNAKE_CASE , encoder_attention_mask=__SCREAMING_SNAKE_CASE , ) __SCREAMING_SNAKE_CASE = encoder_outputs[0] __SCREAMING_SNAKE_CASE = self.pooler(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = ( sequence_output, pooled_output, ) + encoder_outputs[ 1: ] # add hidden_states and attentions if they are here return outputs # sequence_output, pooled_output, (hidden_states), (attentions), highway exits class lowerCAmelCase__ ( a ): """simple docstring""" def __init__( self : Dict , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : str ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE = message __SCREAMING_SNAKE_CASE = exit_layer # start from 1! class lowerCAmelCase__ ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] , __SCREAMING_SNAKE_CASE : Tuple ) -> Optional[Any]: """simple docstring""" super().__init__() __SCREAMING_SNAKE_CASE = BertPooler(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = nn.Dropout(config.hidden_dropout_prob ) __SCREAMING_SNAKE_CASE = nn.Linear(config.hidden_size , config.num_labels ) def UpperCAmelCase__ ( self : Any , __SCREAMING_SNAKE_CASE : Optional[int] ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = encoder_outputs[0] __SCREAMING_SNAKE_CASE = self.pooler(__SCREAMING_SNAKE_CASE ) # "return" pooler_output # BertModel __SCREAMING_SNAKE_CASE = (pooler_input, pooler_output) + encoder_outputs[1:] # "return" bmodel_output # Dropout and classification __SCREAMING_SNAKE_CASE = bmodel_output[1] __SCREAMING_SNAKE_CASE = self.dropout(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = self.classifier(__SCREAMING_SNAKE_CASE ) return logits, pooled_output @add_start_docstrings( "Bert Model (with early exiting - DeeBERT) with a classifier on top,\n also takes care of multi-layer training. " , a , ) class lowerCAmelCase__ ( a ): """simple docstring""" def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : Tuple ) -> Union[str, Any]: """simple docstring""" super().__init__(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = config.num_labels __SCREAMING_SNAKE_CASE = config.num_hidden_layers __SCREAMING_SNAKE_CASE = DeeBertModel(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = nn.Dropout(config.hidden_dropout_prob ) __SCREAMING_SNAKE_CASE = nn.Linear(config.hidden_size , self.config.num_labels ) self.init_weights() @add_start_docstrings_to_model_forward(__SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self : Optional[int] , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : List[Any]=None , __SCREAMING_SNAKE_CASE : List[Any]=None , __SCREAMING_SNAKE_CASE : str=None , __SCREAMING_SNAKE_CASE : Optional[Any]=None , __SCREAMING_SNAKE_CASE : str=None , __SCREAMING_SNAKE_CASE : str=None , __SCREAMING_SNAKE_CASE : Optional[int]=-1 , __SCREAMING_SNAKE_CASE : str=False , ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = self.num_layers try: __SCREAMING_SNAKE_CASE = self.bert( __SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , token_type_ids=__SCREAMING_SNAKE_CASE , position_ids=__SCREAMING_SNAKE_CASE , head_mask=__SCREAMING_SNAKE_CASE , inputs_embeds=__SCREAMING_SNAKE_CASE , ) # sequence_output, pooled_output, (hidden_states), (attentions), highway exits __SCREAMING_SNAKE_CASE = outputs[1] __SCREAMING_SNAKE_CASE = self.dropout(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = self.classifier(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: __SCREAMING_SNAKE_CASE = e.message __SCREAMING_SNAKE_CASE = e.exit_layer __SCREAMING_SNAKE_CASE = outputs[0] if not self.training: __SCREAMING_SNAKE_CASE = entropy(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = [] if labels is not None: if self.num_labels == 1: # We are doing regression __SCREAMING_SNAKE_CASE = MSELoss() __SCREAMING_SNAKE_CASE = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: __SCREAMING_SNAKE_CASE = CrossEntropyLoss() __SCREAMING_SNAKE_CASE = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits __SCREAMING_SNAKE_CASE = [] for highway_exit in outputs[-1]: __SCREAMING_SNAKE_CASE = highway_exit[0] if not self.training: highway_logits_all.append(__SCREAMING_SNAKE_CASE ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression __SCREAMING_SNAKE_CASE = MSELoss() __SCREAMING_SNAKE_CASE = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: __SCREAMING_SNAKE_CASE = CrossEntropyLoss() __SCREAMING_SNAKE_CASE = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(__SCREAMING_SNAKE_CASE ) if train_highway: __SCREAMING_SNAKE_CASE = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: __SCREAMING_SNAKE_CASE = (loss,) + outputs if not self.training: __SCREAMING_SNAKE_CASE = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: __SCREAMING_SNAKE_CASE = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), (highway_exits)
331
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self : List[Any] ) -> List[str]: """simple docstring""" __SCREAMING_SNAKE_CASE = [[1, 2, 4], [1, 2, 3, 4]] __SCREAMING_SNAKE_CASE = DisjunctiveConstraint(__SCREAMING_SNAKE_CASE ) self.assertTrue(isinstance(dc.token_ids , __SCREAMING_SNAKE_CASE ) ) with self.assertRaises(__SCREAMING_SNAKE_CASE ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(__SCREAMING_SNAKE_CASE ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def UpperCAmelCase__ ( self : Union[str, Any] ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(__SCREAMING_SNAKE_CASE ): DisjunctiveConstraint(__SCREAMING_SNAKE_CASE ) # fails here def UpperCAmelCase__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = [[1, 2, 3], [1, 2, 4]] __SCREAMING_SNAKE_CASE = DisjunctiveConstraint(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(1 ) __SCREAMING_SNAKE_CASE = stepped is True and completed is False and reset is False self.assertTrue(__SCREAMING_SNAKE_CASE ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(2 ) __SCREAMING_SNAKE_CASE = stepped is True and completed is False and reset is False self.assertTrue(__SCREAMING_SNAKE_CASE ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(3 ) __SCREAMING_SNAKE_CASE = stepped is True and completed is True and reset is False self.assertTrue(__SCREAMING_SNAKE_CASE ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def UpperCAmelCase__ ( self : int ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] __SCREAMING_SNAKE_CASE = DisjunctiveConstraint(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
331
1
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse("3.8"): import importlib_metadata else: import importlib.metadata as importlib_metadata def a ( _UpperCAmelCase : str , _UpperCAmelCase : int=False ): '''simple docstring''' try: __UpperCAmelCase : Dict = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __UpperCAmelCase : Union[str, Any] = default else: # KEY is set, convert it to True or False. try: __UpperCAmelCase : Union[str, Any] = strtobool(_UpperCAmelCase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'If set, {key} must be yes or no.' ) return _value __A =parse_flag_from_env("RUN_SLOW", default=False) __A =parse_flag_from_env("RUN_REMOTE", default=False) __A =parse_flag_from_env("RUN_LOCAL", default=True) __A =parse_flag_from_env("RUN_PACKAGED", default=True) # Compression __A =pytest.mark.skipif(not config.LZ4_AVAILABLE, reason="test requires lz4") __A =pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason="test requires py7zr") __A =pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason="test requires zstandard") # Audio __A =pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec("soundfile") is None or version.parse(importlib_metadata.version("soundfile")) < version.parse("0.12.0"), reason="test requires sndfile>=0.12.1: 'pip install \"soundfile>=0.12.1\"'; ", ) # Beam __A =pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse("0.3.2"), reason="test requires apache-beam and a compatible dill version", ) # Dill-cloudpickle compatibility __A =pytest.mark.skipif( config.DILL_VERSION <= version.parse("0.3.2"), reason="test requires dill>0.3.2 for cloudpickle compatibility", ) # Windows __A =pytest.mark.skipif( sys.platform == "win32", reason="test should not be run on Windows", ) def a ( _UpperCAmelCase : Any ): '''simple docstring''' try: import faiss # noqa except ImportError: __UpperCAmelCase : Optional[Any] = unittest.skip('''test requires faiss''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : Tuple ): '''simple docstring''' try: import regex # noqa except ImportError: __UpperCAmelCase : Union[str, Any] = unittest.skip('''test requires regex''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : Union[str, Any] ): '''simple docstring''' try: import elasticsearch # noqa except ImportError: __UpperCAmelCase : List[Any] = unittest.skip('''test requires elasticsearch''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : Tuple ): '''simple docstring''' try: import sqlalchemy # noqa except ImportError: __UpperCAmelCase : str = unittest.skip('''test requires sqlalchemy''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : int ): '''simple docstring''' if not config.TORCH_AVAILABLE: __UpperCAmelCase : Any = unittest.skip('''test requires PyTorch''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : Dict ): '''simple docstring''' if not config.TF_AVAILABLE: __UpperCAmelCase : List[str] = unittest.skip('''test requires TensorFlow''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : int ): '''simple docstring''' if not config.JAX_AVAILABLE: __UpperCAmelCase : Any = unittest.skip('''test requires JAX''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : List[str] ): '''simple docstring''' if not config.PIL_AVAILABLE: __UpperCAmelCase : Union[str, Any] = unittest.skip('''test requires Pillow''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : Optional[int] ): '''simple docstring''' try: import transformers # noqa F401 except ImportError: return unittest.skip('''test requires transformers''' )(_UpperCAmelCase ) else: return test_case def a ( _UpperCAmelCase : Tuple ): '''simple docstring''' try: import tiktoken # noqa F401 except ImportError: return unittest.skip('''test requires tiktoken''' )(_UpperCAmelCase ) else: return test_case def a ( _UpperCAmelCase : List[Any] ): '''simple docstring''' try: import spacy # noqa F401 except ImportError: return unittest.skip('''test requires spacy''' )(_UpperCAmelCase ) else: return test_case def a ( _UpperCAmelCase : List[str] ): '''simple docstring''' def _require_spacy_model(_UpperCAmelCase : Union[str, Any] ): try: import spacy # noqa F401 spacy.load(_UpperCAmelCase ) except ImportError: return unittest.skip('''test requires spacy''' )(_UpperCAmelCase ) except OSError: return unittest.skip('''test requires spacy model \'{}\''''.format(_UpperCAmelCase ) )(_UpperCAmelCase ) else: return test_case return _require_spacy_model def a ( _UpperCAmelCase : Tuple ): '''simple docstring''' try: import pyspark # noqa F401 except ImportError: return unittest.skip('''test requires pyspark''' )(_UpperCAmelCase ) else: return test_case def a ( _UpperCAmelCase : Any ): '''simple docstring''' try: import joblibspark # noqa F401 except ImportError: return unittest.skip('''test requires joblibspark''' )(_UpperCAmelCase ) else: return test_case def a ( _UpperCAmelCase : List[str] ): '''simple docstring''' if not _run_slow_tests or _run_slow_tests == 0: __UpperCAmelCase : int = unittest.skip('''test is slow''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : Tuple ): '''simple docstring''' if not _run_local_tests or _run_local_tests == 0: __UpperCAmelCase : Any = unittest.skip('''test is local''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : int ): '''simple docstring''' if not _run_packaged_tests or _run_packaged_tests == 0: __UpperCAmelCase : Union[str, Any] = unittest.skip('''test is packaged''' )(_UpperCAmelCase ) return test_case def a ( _UpperCAmelCase : str ): '''simple docstring''' if not _run_remote_tests or _run_remote_tests == 0: __UpperCAmelCase : int = unittest.skip('''test requires remote''' )(_UpperCAmelCase ) return test_case def a ( *_UpperCAmelCase : Dict ): '''simple docstring''' def decorate(cls : Tuple ): for name, fn in cls.__dict__.items(): if callable(_UpperCAmelCase ) and name.startswith('''test''' ): for decorator in decorators: __UpperCAmelCase : Dict = decorator(_UpperCAmelCase ) setattr(cls , _UpperCAmelCase , _UpperCAmelCase ) return cls return decorate class UpperCAmelCase__ ( __UpperCamelCase ): '''simple docstring''' pass class UpperCAmelCase__ ( __UpperCamelCase ): '''simple docstring''' UpperCamelCase = 0 UpperCamelCase = 1 UpperCamelCase = 2 @contextmanager def a ( _UpperCAmelCase : List[str]=OfflineSimulationMode.CONNECTION_FAILS , _UpperCAmelCase : List[str]=1e-16 ): '''simple docstring''' __UpperCAmelCase : List[Any] = requests.Session().request def timeout_request(_UpperCAmelCase : Any , _UpperCAmelCase : Tuple , _UpperCAmelCase : Optional[Any] , **_UpperCAmelCase : Union[str, Any] ): # Change the url to an invalid url so that the connection hangs __UpperCAmelCase : int = '''https://10.255.255.1''' if kwargs.get('''timeout''' ) is None: raise RequestWouldHangIndefinitelyError( f'Tried a call to {url} in offline mode with no timeout set. Please set a timeout.' ) __UpperCAmelCase : List[Any] = timeout try: return online_request(_UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier __UpperCAmelCase : List[str] = url __UpperCAmelCase : Optional[int] = e.args[0] __UpperCAmelCase : str = (max_retry_error.args[0].replace('''10.255.255.1''' , f'OfflineMock[{url}]' ),) __UpperCAmelCase : Optional[Any] = (max_retry_error,) raise def raise_connection_error(_UpperCAmelCase : str , _UpperCAmelCase : int , **_UpperCAmelCase : Union[str, Any] ): raise requests.ConnectionError('''Offline mode is enabled.''' , request=_UpperCAmelCase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch('''requests.Session.send''' , _UpperCAmelCase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch('''requests.Session.request''' , _UpperCAmelCase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch('''datasets.config.HF_DATASETS_OFFLINE''' , _UpperCAmelCase ): yield else: raise ValueError('''Please use a value from the OfflineSimulationMode enum.''' ) @contextmanager def a ( *_UpperCAmelCase : int , **_UpperCAmelCase : Tuple ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = str(Path().resolve() ) with tempfile.TemporaryDirectory(*_UpperCAmelCase , **_UpperCAmelCase ) as tmp_dir: try: os.chdir(_UpperCAmelCase ) yield finally: os.chdir(_UpperCAmelCase ) @contextmanager def a ( ): '''simple docstring''' import gc gc.collect() __UpperCAmelCase : List[Any] = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def a ( ): '''simple docstring''' import gc gc.collect() __UpperCAmelCase : str = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def a ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Any ): '''simple docstring''' return deepcopy(_UpperCAmelCase ).integers(0 , 1_00 , 10 ).tolist() == deepcopy(_UpperCAmelCase ).integers(0 , 1_00 , 10 ).tolist() def a ( _UpperCAmelCase : Any ): '''simple docstring''' import decorator from requests.exceptions import HTTPError def _wrapper(_UpperCAmelCase : Optional[Any] , *_UpperCAmelCase : Optional[Any] , **_UpperCAmelCase : List[str] ): try: return func(*_UpperCAmelCase , **_UpperCAmelCase ) except HTTPError as err: if str(_UpperCAmelCase ).startswith('''500''' ) or str(_UpperCAmelCase ).startswith('''502''' ): pytest.xfail(str(_UpperCAmelCase ) ) raise err return decorator.decorator(_wrapper , _UpperCAmelCase ) class UpperCAmelCase__ : '''simple docstring''' def __init__( self : Tuple , a_ : Optional[int] , a_ : Any , a_ : Optional[Any] ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = returncode __UpperCAmelCase : Tuple = stdout __UpperCAmelCase : Optional[int] = stderr async def a ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[Any] ): '''simple docstring''' while True: __UpperCAmelCase : List[Any] = await stream.readline() if line: callback(_UpperCAmelCase ) else: break async def a ( _UpperCAmelCase : int , _UpperCAmelCase : Dict=None , _UpperCAmelCase : Optional[int]=None , _UpperCAmelCase : int=None , _UpperCAmelCase : Any=False , _UpperCAmelCase : Union[str, Any]=False ): '''simple docstring''' if echo: print('''\nRunning: ''' , ''' '''.join(_UpperCAmelCase ) ) __UpperCAmelCase : Tuple = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=_UpperCAmelCase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=_UpperCAmelCase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __UpperCAmelCase : str = [] __UpperCAmelCase : Optional[Any] = [] def tee(_UpperCAmelCase : List[Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any]="" ): __UpperCAmelCase : Optional[int] = line.decode('''utf-8''' ).rstrip() sink.append(_UpperCAmelCase ) if not quiet: print(_UpperCAmelCase , _UpperCAmelCase , file=_UpperCAmelCase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda _UpperCAmelCase : tee(_UpperCAmelCase , _UpperCAmelCase , sys.stdout , label='''stdout:''' ) ), _read_stream(p.stderr , lambda _UpperCAmelCase : tee(_UpperCAmelCase , _UpperCAmelCase , sys.stderr , label='''stderr:''' ) ), ] , timeout=_UpperCAmelCase , ) return _RunOutput(await p.wait() , _UpperCAmelCase , _UpperCAmelCase ) def a ( _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any]=None , _UpperCAmelCase : Optional[Any]=None , _UpperCAmelCase : Dict=1_80 , _UpperCAmelCase : str=False , _UpperCAmelCase : Dict=True ): '''simple docstring''' __UpperCAmelCase : int = asyncio.get_event_loop() __UpperCAmelCase : List[Any] = loop.run_until_complete( _stream_subprocess(_UpperCAmelCase , env=_UpperCAmelCase , stdin=_UpperCAmelCase , timeout=_UpperCAmelCase , quiet=_UpperCAmelCase , echo=_UpperCAmelCase ) ) __UpperCAmelCase : List[str] = ''' '''.join(_UpperCAmelCase ) if result.returncode > 0: __UpperCAmelCase : str = '''\n'''.join(result.stderr ) raise RuntimeError( f'\'{cmd_str}\' failed with returncode {result.returncode}\n\n' f'The combined stderr from workers follows:\n{stderr}' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f'\'{cmd_str}\' produced no output.' ) return result def a ( ): '''simple docstring''' __UpperCAmelCase : List[str] = os.environ.get('''PYTEST_XDIST_WORKER''' , '''gw0''' ) __UpperCAmelCase : Optional[Any] = re.sub(R'''^gw''' , '''''' , _UpperCAmelCase , 0 , re.M ) return int(_UpperCAmelCase ) def a ( ): '''simple docstring''' __UpperCAmelCase : str = 2_95_00 __UpperCAmelCase : str = pytest_xdist_worker_id() return port + uniq_delta
226
import numpy as np class UpperCAmelCase__ : '''simple docstring''' def __init__( self : List[Any] ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = (0, 0) __UpperCAmelCase : List[str] = None __UpperCAmelCase : int = 0 __UpperCAmelCase : Union[str, Any] = 0 __UpperCAmelCase : List[Any] = 0 def __eq__( self : Optional[int] , a_ : List[Any] ): '''simple docstring''' return self.position == cell.position def snake_case__ ( self : List[Any] ): '''simple docstring''' print(self.position ) class UpperCAmelCase__ : '''simple docstring''' def __init__( self : str , a_ : Dict=(5, 5) ): '''simple docstring''' __UpperCAmelCase : str = np.zeros(a_ ) __UpperCAmelCase : List[Any] = world_size[0] __UpperCAmelCase : List[str] = world_size[1] def snake_case__ ( self : Tuple ): '''simple docstring''' print(self.w ) def snake_case__ ( self : Union[str, Any] , a_ : Tuple ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] __UpperCAmelCase : str = cell.position[0] __UpperCAmelCase : List[Any] = cell.position[1] __UpperCAmelCase : str = [] for n in neughbour_cord: __UpperCAmelCase : Optional[Any] = current_x + n[0] __UpperCAmelCase : Any = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: __UpperCAmelCase : int = Cell() __UpperCAmelCase : List[str] = (x, y) __UpperCAmelCase : List[Any] = cell neighbours.append(a_ ) return neighbours def a ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : Any , _UpperCAmelCase : int ): '''simple docstring''' __UpperCAmelCase : List[Any] = [] __UpperCAmelCase : Optional[Any] = [] _open.append(_UpperCAmelCase ) while _open: __UpperCAmelCase : str = np.argmin([n.f for n in _open] ) __UpperCAmelCase : Optional[Any] = _open[min_f] _closed.append(_open.pop(_UpperCAmelCase ) ) if current == goal: break for n in world.get_neigbours(_UpperCAmelCase ): for c in _closed: if c == n: continue __UpperCAmelCase : Optional[Any] = current.g + 1 __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = n.position __UpperCAmelCase , __UpperCAmelCase : Tuple = goal.position __UpperCAmelCase : Union[str, Any] = (ya - ya) ** 2 + (xa - xa) ** 2 __UpperCAmelCase : Optional[Any] = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(_UpperCAmelCase ) __UpperCAmelCase : str = [] while current.parent is not None: path.append(current.position ) __UpperCAmelCase : int = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": __A =Gridworld() # Start position and goal __A =Cell() __A =(0, 0) __A =Cell() __A =(4, 4) print(f'''path from {start.position} to {goal.position}''') __A =astar(world, start, goal) # Just for visual reasons. for i in s: __A =1 print(world.w)
226
1
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer lowerCamelCase_ : List[str] = logging.get_logger(__name__) # pylint: disable=invalid-name lowerCamelCase_ : Any = """ Examples: ```py >>> from PIL import Image >>> import torch >>> from diffusers import DiffusionPipeline >>> from diffusers.utils import export_to_gif, load_image >>> device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\") >>> repo = \"openai/shap-e-img2img\" >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16) >>> pipe = pipe.to(device) >>> guidance_scale = 3.0 >>> image_url = \"https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png\" >>> image = load_image(image_url).convert(\"RGB\") >>> images = pipe( ... image, ... guidance_scale=guidance_scale, ... num_inference_steps=64, ... frame_size=256, ... ).images >>> gif_path = export_to_gif(images[0], \"corgi_3d.gif\") ``` """ @dataclass class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" __lowerCAmelCase = 42 class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" def __init__( self , __A , __A , __A , __A , __A , ) -> Union[str, Any]: super().__init__() self.register_modules( prior=__A , image_encoder=__A , image_processor=__A , scheduler=__A , renderer=__A , ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A , __A , __A , __A ) -> Union[str, Any]: if latents is None: a =randn_tensor(__A , generator=__A , device=__A , dtype=__A ) else: if latents.shape != shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) a =latents.to(__A ) a =latents * scheduler.init_noise_sigma return latents def SCREAMING_SNAKE_CASE ( self , __A=0 ) -> Optional[Any]: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) a =torch.device(f'''cuda:{gpu_id}''' ) a =[self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__A , __A ) @property def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: if self.device != torch.device('''meta''' ) or not hasattr(self.image_encoder , '''_hf_hook''' ): return self.device for module in self.image_encoder.modules(): if ( hasattr(__A , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def SCREAMING_SNAKE_CASE ( self , __A , __A , __A , __A , ) -> Optional[Any]: if isinstance(__A , __A ) and isinstance(image[0] , torch.Tensor ): a =torch.cat(__A , axis=0 ) if image[0].ndim == 4 else torch.stack(__A , axis=0 ) if not isinstance(__A , torch.Tensor ): a =self.image_processor(__A , return_tensors='''pt''' ).pixel_values[0].unsqueeze(0 ) a =image.to(dtype=self.image_encoder.dtype , device=__A ) a =self.image_encoder(__A )['''last_hidden_state'''] a =image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 a =image_embeds.repeat_interleave(__A , dim=0 ) if do_classifier_free_guidance: a =torch.zeros_like(__A ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes a =torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(__A ) def __call__( self , __A , __A = 1 , __A = 25 , __A = None , __A = None , __A = 4.0 , __A = 64 , __A = "pil" , __A = True , ) -> Any: if isinstance(__A , PIL.Image.Image ): a =1 elif isinstance(__A , torch.Tensor ): a =image.shape[0] elif isinstance(__A , __A ) and isinstance(image[0] , (torch.Tensor, PIL.Image.Image) ): a =len(__A ) else: raise ValueError( f'''`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(__A )}''' ) a =self._execution_device a =batch_size * num_images_per_prompt a =guidance_scale > 1.0 a =self._encode_image(__A , __A , __A , __A ) # prior self.scheduler.set_timesteps(__A , device=__A ) a =self.scheduler.timesteps a =self.prior.config.num_embeddings a =self.prior.config.embedding_dim a =self.prepare_latents( (batch_size, num_embeddings * embedding_dim) , image_embeds.dtype , __A , __A , __A , self.scheduler , ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim a =latents.reshape(latents.shape[0] , __A , __A ) for i, t in enumerate(self.progress_bar(__A ) ): # expand the latents if we are doing classifier free guidance a =torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents a =self.scheduler.scale_model_input(__A , __A ) a =self.prior( __A , timestep=__A , proj_embedding=__A , ).predicted_image_embedding # remove the variance a , a =noise_pred.split( scaled_model_input.shape[2] , dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: a , a =noise_pred.chunk(2 ) a =noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) a =self.scheduler.step( __A , timestep=__A , sample=__A , ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=__A ) a =[] for i, latent in enumerate(__A ): print() a =self.renderer.decode( latent[None, :] , __A , size=__A , ray_batch_size=4096 , n_coarse_samples=64 , n_fine_samples=128 , ) images.append(__A ) a =torch.stack(__A ) if output_type not in ["np", "pil"]: raise ValueError(f'''Only the output types `pil` and `np` are supported not output_type={output_type}''' ) a =images.cpu().numpy() if output_type == "pil": a =[self.numpy_to_pil(__A ) for image in images] # Offload last model to CPU if hasattr(self , '''final_offload_hook''' ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=__A )
215
"""simple docstring""" import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __A ( _SCREAMING_SNAKE_CASE, unittest.TestCase ): """simple docstring""" __lowerCAmelCase = RoFormerTokenizer __lowerCAmelCase = RoFormerTokenizerFast __lowerCAmelCase = True __lowerCAmelCase = True def SCREAMING_SNAKE_CASE ( self ) -> List[str]: super().setUp() def SCREAMING_SNAKE_CASE ( self , **__A ) -> Optional[int]: return self.tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__A ) def SCREAMING_SNAKE_CASE ( self , **__A ) -> List[Any]: return self.rust_tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__A ) def SCREAMING_SNAKE_CASE ( self ) -> Dict: a ='''永和服装饰品有限公司,今天天气非常好''' a ='''永和 服装 饰品 有限公司 , 今 天 天 气 非常 好''' return input_text, output_text def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a =self.get_tokenizer() a , a =self.get_chinese_input_output_texts() a =tokenizer.tokenize(__A ) self.assertListEqual(__A , output_text.split() ) a =tokens + [tokenizer.unk_token] a =[2_2943, 2_1332, 3_4431, 4_5904, 117, 306, 1231, 1231, 2653, 3_3994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) def SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: a =self.get_rust_tokenizer() a , a =self.get_chinese_input_output_texts() a =tokenizer.tokenize(__A ) self.assertListEqual(__A , output_text.split() ) a =tokens + [tokenizer.unk_token] a =[2_2943, 2_1332, 3_4431, 4_5904, 117, 306, 1231, 1231, 2653, 3_3994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) def SCREAMING_SNAKE_CASE ( self ) -> Tuple: pass def SCREAMING_SNAKE_CASE ( self ) -> Tuple: pass def SCREAMING_SNAKE_CASE ( self ) -> int: pass
215
1
# coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers UpperCamelCase_ = '3' print('''Python version:''', sys.version) print('''transformers version:''', transformers.__version__) try: import torch print('''Torch version:''', torch.__version__) print('''Cuda available:''', torch.cuda.is_available()) print('''Cuda version:''', torch.version.cuda) print('''CuDNN version:''', torch.backends.cudnn.version()) print('''Number of GPUs available:''', torch.cuda.device_count()) print('''NCCL version:''', torch.cuda.nccl.version()) except ImportError: print('''Torch version:''', None) try: import deepspeed print('''DeepSpeed version:''', deepspeed.__version__) except ImportError: print('''DeepSpeed version:''', None) try: import tensorflow as tf print('''TensorFlow version:''', tf.__version__) print('''TF GPUs available:''', bool(tf.config.list_physical_devices('''GPU'''))) print('''Number of TF GPUs available:''', len(tf.config.list_physical_devices('''GPU'''))) except ImportError: print('''TensorFlow version:''', None)
345
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class lowercase__ ( unittest.TestCase): def __A ( self : str , UpperCamelCase__ : int ): '''simple docstring''' for model_result in results.values(): for batch_size, sequence_length in zip(model_result['''bs'''] , model_result['''ss'''] ): SCREAMING_SNAKE_CASE : Union[str, Any] = model_result['''result'''][batch_size][sequence_length] self.assertIsNotNone(UpperCamelCase__ ) def __A ( self : Optional[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : List[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : str = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = '''sgugger/tiny-distilbert-classification''' SCREAMING_SNAKE_CASE : Union[str, Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , only_pretrain_model=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : List[str] = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : Any ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : int = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Union[str, Any] = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : Optional[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : int = AutoConfig.from_pretrained(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[int] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : str = TensorFlowBenchmark(UpperCamelCase__ , [config] ) SCREAMING_SNAKE_CASE : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : int ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : Optional[int] = AutoConfig.from_pretrained(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : List[str] = TensorFlowBenchmark(UpperCamelCase__ , [config] ) SCREAMING_SNAKE_CASE : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : List[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : int = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Optional[int] = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self : Optional[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : int = AutoConfig.from_pretrained(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Tuple = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Dict = TensorFlowBenchmark(UpperCamelCase__ , [config] ) SCREAMING_SNAKE_CASE : str = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = '''patrickvonplaten/t5-tiny-random''' SCREAMING_SNAKE_CASE : Union[str, Any] = AutoConfig.from_pretrained(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Union[str, Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Dict = TensorFlowBenchmark(UpperCamelCase__ , configs=[config] ) SCREAMING_SNAKE_CASE : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices('''GPU''' ) ) == 0 , '''Cannot do xla on CPU.''' ) def __A ( self : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : Tuple = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , use_xla=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Any = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = '''sshleifer/tiny-gpt2''' with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE : Tuple = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCamelCase__ , save_to_csv=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCamelCase__ , '''inf_time.csv''' ) , inference_memory_csv_file=os.path.join(UpperCamelCase__ , '''inf_mem.csv''' ) , env_info_csv_file=os.path.join(UpperCamelCase__ , '''env.csv''' ) , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Optional[Any] = TensorFlowBenchmark(UpperCamelCase__ ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCamelCase__ , '''inf_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , '''inf_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , '''env.csv''' ) ).exists() ) def __A ( self : int ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = '''sshleifer/tiny-gpt2''' def _check_summary_is_not_empty(UpperCamelCase__ : Dict ): self.assertTrue(hasattr(UpperCamelCase__ , '''sequential''' ) ) self.assertTrue(hasattr(UpperCamelCase__ , '''cumulative''' ) ) self.assertTrue(hasattr(UpperCamelCase__ , '''current''' ) ) self.assertTrue(hasattr(UpperCamelCase__ , '''total''' ) ) with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE : List[str] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCamelCase__ , '''log.txt''' ) , log_print=UpperCamelCase__ , trace_memory_line_by_line=UpperCamelCase__ , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : List[Any] = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Dict = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , '''log.txt''' ) ).exists() )
182
0
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=__magic_name__ ) class lowerCamelCase__ ( __magic_name__ ): '''simple docstring''' lowerCamelCase = field(default='''language-modeling''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) lowerCamelCase = Features({'''text''': Value('''string''' )} ) lowerCamelCase = Features({} ) lowerCamelCase = "text" @property def _lowerCAmelCase ( self ) -> Dict[str, str]: return {self.text_column: "text"}
341
"""simple docstring""" def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase ) -> int: return int((input_a, input_a).count(1 ) != 0 ) def _lowerCamelCase() -> None: assert or_gate(0 , 0 ) == 0 assert or_gate(0 , 1 ) == 1 assert or_gate(1 , 0 ) == 1 assert or_gate(1 , 1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
341
1
import numpy as np _lowerCAmelCase : Optional[Any] = [ ["a", "b", "c", "d", "e"], ["f", "g", "h", "i", "k"], ["l", "m", "n", "o", "p"], ["q", "r", "s", "t", "u"], ["v", "w", "x", "y", "z"], ] class __magic_name__ : def __init__( self ) -> Optional[Any]: '''simple docstring''' __a =np.array(__snake_case ) def __magic_name__ ( self , __snake_case ) -> Any: '''simple docstring''' __a , __a =np.where(letter == self.SQUARE ) __a =np.concatenate([indexa + 1, indexa + 1] ) return indexes def __magic_name__ ( self , __snake_case , __snake_case ) -> List[Any]: '''simple docstring''' __a =self.SQUARE[indexa - 1, indexa - 1] return letter def __magic_name__ ( self , __snake_case ) -> Optional[int]: '''simple docstring''' __a =message.lower() __a =message.replace(' ' , '' ) __a =message.replace('j' , 'i' ) __a =np.empty((2, len(__snake_case )) ) for letter_index in range(len(__snake_case ) ): __a =self.letter_to_numbers(message[letter_index] ) __a =numbers[0] __a =numbers[1] __a =first_step.reshape(2 * len(__snake_case ) ) __a ='' for numbers_index in range(len(__snake_case ) ): __a =int(second_step[numbers_index * 2] ) __a =int(second_step[(numbers_index * 2) + 1] ) __a =self.numbers_to_letter(__snake_case , __snake_case ) __a =encoded_message + letter return encoded_message def __magic_name__ ( self , __snake_case ) -> Tuple: '''simple docstring''' __a =message.lower() message.replace(' ' , '' ) __a =np.empty(2 * len(__snake_case ) ) for letter_index in range(len(__snake_case ) ): __a =self.letter_to_numbers(message[letter_index] ) __a =numbers[0] __a =numbers[1] __a =first_step.reshape((2, len(__snake_case )) ) __a ='' for numbers_index in range(len(__snake_case ) ): __a =int(second_step[0, numbers_index] ) __a =int(second_step[1, numbers_index] ) __a =self.numbers_to_letter(__snake_case , __snake_case ) __a =decoded_message + letter return decoded_message
218
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __A ( )-> Tuple: """simple docstring""" raise RuntimeError('CUDA out of memory.' ) class __lowerCamelCase ( nn.Module): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() _UpperCAmelCase = nn.Linear(3 , 4 ) _UpperCAmelCase = nn.BatchNormad(4 ) _UpperCAmelCase = nn.Linear(4 , 5 ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase ) ) ) class __lowerCamelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase ): nonlocal batch_sizes batch_sizes.append(UpperCAmelCase ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga _UpperCAmelCase , _UpperCAmelCase = mock_training_loop_function('hello' ) self.assertListEqual(UpperCAmelCase , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, 'hello'] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(UpperCAmelCase ): pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function(128 , 'hello' , 'world' ) self.assertIn('Batch size was passed into `f`' , cm.exception.args[0] ) self.assertIn('`f(arg1=\'hello\', arg2=\'world\')' , cm.exception.args[0] ) def UpperCamelCase ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(UpperCAmelCase ): raise ValueError('Oops, we had an error!' ) with self.assertRaises(UpperCAmelCase ) as cm: mock_training_loop_function() self.assertIn('Oops, we had an error!' , cm.exception.args[0] ) @require_cuda def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase = torch.cuda.memory_allocated() _UpperCAmelCase = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , UpperCAmelCase ) _UpperCAmelCase = release_memory(UpperCAmelCase ) self.assertEqual(torch.cuda.memory_allocated() , UpperCAmelCase )
39
0
from collections import deque from math import floor from random import random from time import time class __a : def __init__( self ) -> Dict: '''simple docstring''' lowercase__: Dict = {} def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=1 ) -> Optional[int]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: lowercase__: int = [[w, v]] if not self.graph.get(lowerCAmelCase__ ): lowercase__: Union[str, Any] = [] def SCREAMING_SNAKE_CASE__ ( self ) -> Any: '''simple docstring''' return list(self.graph ) def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-2 , lowerCAmelCase__=-1 ) -> Union[str, Any]: '''simple docstring''' if s == d: return [] lowercase__: Tuple = [] lowercase__: Tuple = [] if s == -2: lowercase__: Any = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) lowercase__: Dict = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowercase__: Dict = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) lowercase__: Union[str, Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: lowercase__: Optional[int] = stack[len(lowerCAmelCase__ ) - 1] else: lowercase__: Dict = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-1 ) -> List[str]: '''simple docstring''' if c == -1: lowercase__: int = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): lowercase__: str = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-2 ) -> Dict: '''simple docstring''' lowercase__: int = deque() lowercase__: Dict = [] if s == -2: lowercase__: Optional[int] = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: lowercase__: str = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ ) -> Dict: '''simple docstring''' lowercase__: Tuple = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ ) -> Union[str, Any]: '''simple docstring''' return len(self.graph[u] ) def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-2 ) -> Optional[Any]: '''simple docstring''' lowercase__: Tuple = [] lowercase__: str = [] if s == -2: lowercase__: Dict = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) lowercase__: List[Any] = s lowercase__: Any = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowercase__: List[str] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowercase__: Dict = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(lowerCAmelCase__ ) != 0: lowercase__: int = stack[len(lowerCAmelCase__ ) - 1] else: lowercase__: Optional[int] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return sorted_nodes def SCREAMING_SNAKE_CASE__ ( self ) -> Any: '''simple docstring''' lowercase__: List[Any] = [] lowercase__: int = [] lowercase__: List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) lowercase__: Dict = -2 lowercase__: Union[str, Any] = [] lowercase__: List[str] = s lowercase__: Dict = False lowercase__: Union[str, Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowercase__: Any = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): lowercase__: List[Any] = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowercase__: Union[str, Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() lowercase__: Any = True if len(lowerCAmelCase__ ) != 0: lowercase__: Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: lowercase__: Union[str, Any] = False indirect_parents.append(lowerCAmelCase__ ) lowercase__: int = s lowercase__: str = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[Any]: '''simple docstring''' lowercase__: Any = [] lowercase__: int = [] lowercase__: Dict = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) lowercase__: Optional[int] = -2 lowercase__: List[Any] = [] lowercase__: List[str] = s lowercase__: List[Any] = False lowercase__: str = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowercase__: Dict = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): lowercase__: Any = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowercase__: List[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() lowercase__: Optional[Any] = True if len(lowerCAmelCase__ ) != 0: lowercase__: Any = stack[len(lowerCAmelCase__ ) - 1] else: lowercase__: Optional[Any] = False indirect_parents.append(lowerCAmelCase__ ) lowercase__: Dict = s lowercase__: Dict = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-2 , lowerCAmelCase__=-1 ) -> Dict: '''simple docstring''' lowercase__: Union[str, Any] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase__: Optional[Any] = time() return end - begin def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-2 ) -> List[str]: '''simple docstring''' lowercase__: str = time() self.bfs(lowerCAmelCase__ ) lowercase__: List[str] = time() return end - begin class __a : def __init__( self ) -> Tuple: '''simple docstring''' lowercase__: Dict = {} def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=1 ) -> List[Any]: '''simple docstring''' # check if the u exists if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist lowercase__: str = [[w, v]] # add the other way if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist lowercase__: Union[str, Any] = [[w, u]] def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) # the other way round if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(lowerCAmelCase__ ) def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-2 , lowerCAmelCase__=-1 ) -> List[str]: '''simple docstring''' if s == d: return [] lowercase__: str = [] lowercase__: int = [] if s == -2: lowercase__: Tuple = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) lowercase__: int = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowercase__: int = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) lowercase__: List[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: lowercase__: Union[str, Any] = stack[len(lowerCAmelCase__ ) - 1] else: lowercase__: Optional[Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-1 ) -> Optional[Any]: '''simple docstring''' if c == -1: lowercase__: Any = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): lowercase__: Optional[Any] = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-2 ) -> List[Any]: '''simple docstring''' lowercase__: str = deque() lowercase__: List[Any] = [] if s == -2: lowercase__: str = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: lowercase__: Union[str, Any] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ ) -> List[Any]: '''simple docstring''' return len(self.graph[u] ) def SCREAMING_SNAKE_CASE__ ( self ) -> str: '''simple docstring''' lowercase__: str = [] lowercase__: Dict = [] lowercase__: Optional[int] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) lowercase__: Dict = -2 lowercase__: Dict = [] lowercase__: List[Any] = s lowercase__: Union[str, Any] = False lowercase__: List[str] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowercase__: Optional[int] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): lowercase__: Any = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowercase__: List[str] = node[1] break # check if all the children are visited if s == ss: stack.pop() lowercase__: str = True if len(lowerCAmelCase__ ) != 0: lowercase__: Dict = stack[len(lowerCAmelCase__ ) - 1] else: lowercase__: int = False indirect_parents.append(lowerCAmelCase__ ) lowercase__: Tuple = s lowercase__: List[Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def SCREAMING_SNAKE_CASE__ ( self ) -> Any: '''simple docstring''' lowercase__: Tuple = [] lowercase__: Optional[int] = [] lowercase__: Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) lowercase__: Tuple = -2 lowercase__: Any = [] lowercase__: int = s lowercase__: Optional[int] = False lowercase__: List[Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowercase__: Optional[int] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): lowercase__: Union[str, Any] = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowercase__: Any = node[1] break # check if all the children are visited if s == ss: stack.pop() lowercase__: List[str] = True if len(lowerCAmelCase__ ) != 0: lowercase__: List[str] = stack[len(lowerCAmelCase__ ) - 1] else: lowercase__: Dict = False indirect_parents.append(lowerCAmelCase__ ) lowercase__: Optional[Any] = s lowercase__: Optional[int] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def SCREAMING_SNAKE_CASE__ ( self ) -> List[Any]: '''simple docstring''' return list(self.graph ) def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-2 , lowerCAmelCase__=-1 ) -> Union[str, Any]: '''simple docstring''' lowercase__: Dict = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase__: List[Any] = time() return end - begin def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__=-2 ) -> List[Any]: '''simple docstring''' lowercase__: str = time() self.bfs(lowerCAmelCase__ ) lowercase__: List[str] = time() return end - begin
288
import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging __lowerCAmelCase = logging.get_logger(__name__) class __a ( __UpperCamelCase ): __lowercase : Tuple = ['input_values', 'attention_mask'] def __init__( self , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 16_000 , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = False , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 16 , lowerCAmelCase__ = 64 , lowerCAmelCase__ = "hann_window" , lowerCAmelCase__ = 1.0 , lowerCAmelCase__ = 80 , lowerCAmelCase__ = 7_600 , lowerCAmelCase__ = 1E-10 , lowerCAmelCase__ = 2 , lowerCAmelCase__ = True , **lowerCAmelCase__ , ) -> List[Any]: '''simple docstring''' super().__init__(feature_size=lowerCAmelCase__ , sampling_rate=lowerCAmelCase__ , padding_value=lowerCAmelCase__ , **lowerCAmelCase__ ) lowercase__: Dict = do_normalize lowercase__: Optional[Any] = return_attention_mask lowercase__: str = num_mel_bins lowercase__: Dict = hop_length lowercase__: Dict = win_length lowercase__: Optional[int] = win_function lowercase__: Any = frame_signal_scale lowercase__: Tuple = fmin lowercase__: Tuple = fmax lowercase__: Dict = mel_floor lowercase__: int = reduction_factor lowercase__: List[Any] = win_length * sampling_rate // 1_000 lowercase__: Optional[Any] = hop_length * sampling_rate // 1_000 lowercase__: Optional[int] = optimal_fft_length(self.sample_size ) lowercase__: Optional[Any] = (self.n_fft // 2) + 1 lowercase__: str = window_function(window_length=self.sample_size , name=self.win_function , periodic=lowerCAmelCase__ ) lowercase__: Optional[Any] = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm='slaney' , mel_scale='slaney' , ) if frame_signal_scale != 1.0: warnings.warn( 'The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers' , lowerCAmelCase__ , ) if reduction_factor != 2.0: warnings.warn( 'The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers' , lowerCAmelCase__ , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def SCREAMING_SNAKE_CASE__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 0.0 ) -> List[np.ndarray]: '''simple docstring''' if attention_mask is not None: lowercase__: List[str] = np.array(lowerCAmelCase__ , np.intaa ) lowercase__: Tuple = [] for vector, length in zip(lowerCAmelCase__ , attention_mask.sum(-1 ) ): lowercase__: int = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: lowercase__: Tuple = padding_value normed_input_values.append(lowerCAmelCase__ ) else: lowercase__: Union[str, Any] = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , ) -> np.ndarray: '''simple docstring''' lowercase__: List[str] = spectrogram( lowerCAmelCase__ , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel='log10' , ) return log_mel_spec.T def __call__( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> BatchFeature: '''simple docstring''' if audio is None and audio_target is None: raise ValueError('You must provide either `audio` or `audio_target` values.' ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'The model corresponding to this feature extractor: {self} was trained using a sampling rate of' F' {self.sampling_rate}. Please make sure that the provided audio input was sampled with' F' {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( 'It is strongly recommended to pass the ``sampling_rate`` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) if audio is not None: lowercase__: Dict = self._process_audio( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ , ) else: lowercase__: str = None if audio_target is not None: lowercase__: List[str] = self._process_audio( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ , ) if inputs is None: return inputs_target else: lowercase__: int = inputs_target['input_values'] lowercase__: List[str] = inputs_target.get('attention_mask' ) if decoder_attention_mask is not None: lowercase__: Optional[int] = decoder_attention_mask return inputs def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> BatchFeature: '''simple docstring''' lowercase__: int = isinstance(lowerCAmelCase__ , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(F'Only mono-channel audio is supported for input to {self}' ) lowercase__: Tuple = is_batched_numpy or ( isinstance(lowerCAmelCase__ , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowercase__: Dict = [np.asarray(lowerCAmelCase__ , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(lowerCAmelCase__ , np.ndarray ): lowercase__: Optional[Any] = np.asarray(lowerCAmelCase__ , dtype=np.floataa ) elif isinstance(lowerCAmelCase__ , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): lowercase__: Optional[Any] = speech.astype(np.floataa ) # always return batch if not is_batched: lowercase__: Optional[int] = [speech] # needed to make pad() work on spectrogram inputs lowercase__: str = self.feature_size # convert into correct format for padding if is_target: lowercase__: int = [self._extract_mel_features(lowerCAmelCase__ ) for waveform in speech] lowercase__: Dict = BatchFeature({'input_values': features} ) lowercase__: Union[str, Any] = self.num_mel_bins else: lowercase__: Union[str, Any] = BatchFeature({'input_values': speech} ) lowercase__: Dict = self.pad( lowerCAmelCase__ , padding=lowerCAmelCase__ , max_length=lowerCAmelCase__ , truncation=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , **lowerCAmelCase__ , ) lowercase__: List[str] = feature_size_hack # convert input values to correct format lowercase__: Union[str, Any] = padded_inputs['input_values'] if not isinstance(input_values[0] , np.ndarray ): lowercase__: List[Any] = [np.asarray(lowerCAmelCase__ , dtype=np.floataa ) for array in input_values] elif ( not isinstance(lowerCAmelCase__ , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): lowercase__: Dict = [array.astype(np.floataa ) for array in input_values] elif isinstance(lowerCAmelCase__ , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): lowercase__: Tuple = input_values.astype(np.floataa ) # convert attention_mask to correct format lowercase__: Tuple = padded_inputs.get('attention_mask' ) if attention_mask is not None: lowercase__: str = [np.asarray(lowerCAmelCase__ , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: lowercase__: Tuple = ( attention_mask if self._get_padding_strategies(lowerCAmelCase__ , max_length=lowerCAmelCase__ ) is not PaddingStrategy.DO_NOT_PAD else None ) lowercase__: str = self.zero_mean_unit_var_norm( padded_inputs['input_values'] , attention_mask=lowerCAmelCase__ , padding_value=self.padding_value ) if return_tensors is not None: lowercase__: Union[str, Any] = padded_inputs.convert_to_tensors(lowerCAmelCase__ ) return padded_inputs def SCREAMING_SNAKE_CASE__ ( self ) -> Dict[str, Any]: '''simple docstring''' lowercase__: int = super().to_dict() # Don't serialize these as they are derived from the other properties. lowercase__: str = ['window', 'mel_filters', 'sample_size', 'sample_stride', 'n_fft', 'n_freqs'] for name in names: if name in output: del output[name] return output
288
1
'''simple docstring''' from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function snake_case_ : int = 1.0_54_57_18_17e-34 # unit of ℏ : J * s snake_case_ : Optional[int] = 3e8 # unit of c : m * s^-1 def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): if (force, area, distance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if force < 0: raise ValueError('Magnitude of force can not be negative' ) if distance < 0: raise ValueError('Distance can not be negative' ) if area < 0: raise ValueError('Area can not be negative' ) if force == 0: _UpperCamelCase : Tuple = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 2_4_0 * (distance) ** 4 ) return {"force": force} elif area == 0: _UpperCamelCase : str = (2_4_0 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: _UpperCamelCase : Tuple = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (2_4_0 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError('One and only one argument must be 0' ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
83
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class lowercase__ ( lowercase ): lowercase__ = """openai/whisper-base""" lowercase__ = ( """This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the """ """transcribed text.""" ) lowercase__ = """transcriber""" lowercase__ = WhisperProcessor lowercase__ = WhisperForConditionalGeneration lowercase__ = ["""audio"""] lowercase__ = ["""text"""] def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Optional[int] ): '''simple docstring''' return self.pre_processor(lowerCamelCase__ ,return_tensors='pt' ).input_features def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : Tuple ): '''simple docstring''' return self.model.generate(inputs=lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' return self.pre_processor.batch_decode(lowerCamelCase__ ,skip_special_tokens=lowerCamelCase__ )[0]
83
1
import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" with open(lowerCamelCase__ ) as metadata_file: lowercase__ : Dict = json.load(lowerCamelCase__ ) lowercase__ : str = LukeConfig(use_entity_aware_attention=lowerCamelCase__ , **metadata["model_config"] ) # Load in the weights from the checkpoint_path lowercase__ : str = torch.load(lowerCamelCase__ , map_location="cpu" ) # Load the entity vocab file lowercase__ : List[Any] = load_entity_vocab(lowerCamelCase__ ) lowercase__ : Optional[Any] = RobertaTokenizer.from_pretrained(metadata["model_config"]["bert_model_name"] ) # Add special tokens to the token vocabulary for downstream tasks lowercase__ : int = AddedToken("<ent>" , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) lowercase__ : str = AddedToken("<ent2>" , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) tokenizer.add_special_tokens({"additional_special_tokens": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowerCamelCase__ ) with open(os.path.join(lowerCamelCase__ , LukeTokenizer.vocab_files_names["entity_vocab_file"] ) , "w" ) as f: json.dump(lowerCamelCase__ , lowerCamelCase__ ) lowercase__ : Optional[int] = LukeTokenizer.from_pretrained(lowerCamelCase__ ) # Initialize the embeddings of the special tokens lowercase__ : int = state_dict["embeddings.word_embeddings.weight"] lowercase__ : Dict = word_emb[tokenizer.convert_tokens_to_ids(["@"] )[0]].unsqueeze(0 ) lowercase__ : Any = word_emb[tokenizer.convert_tokens_to_ids(["#"] )[0]].unsqueeze(0 ) lowercase__ : str = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: lowercase__ : List[str] = F"""encoder.layer.{layer_index}.attention.self.""" lowercase__ : List[str] = state_dict[prefix + matrix_name] lowercase__ : List[Any] = state_dict[prefix + matrix_name] lowercase__ : Optional[Any] = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks lowercase__ : Any = state_dict["entity_embeddings.entity_embeddings.weight"] lowercase__ : Dict = entity_emb[entity_vocab["[MASK]"]] lowercase__ : Any = LukeModel(config=lowerCamelCase__ ).eval() lowercase__ , lowercase__ : Dict = model.load_state_dict(lowerCamelCase__ , strict=lowerCamelCase__ ) if not (len(lowerCamelCase__ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {", ".join(lowerCamelCase__ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith("entity_predictions" ) or key.startswith("lm_head" ) for key in unexpected_keys )): raise ValueError( "Unexpected keys" F""" {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}""" ) # Check outputs lowercase__ : List[str] = LukeTokenizer.from_pretrained(lowerCamelCase__ , task="entity_classification" ) lowercase__ : Union[str, Any] = ( "Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the" " new world number one avoid a humiliating second- round exit at Wimbledon ." ) lowercase__ : Tuple = (39, 42) lowercase__ : str = tokenizer(lowerCamelCase__ , entity_spans=[span] , add_prefix_space=lowerCamelCase__ , return_tensors="pt" ) lowercase__ : int = model(**lowerCamelCase__ ) # Verify word hidden states if model_size == "large": lowercase__ : Union[str, Any] = torch.Size((1, 42, 1_024) ) lowercase__ : Dict = torch.tensor( [[0.0133, 0.0865, 0.0095], [0.3093, -0.2576, -0.7418], [-0.1720, -0.2117, -0.2869]] ) else: # base lowercase__ : Optional[int] = torch.Size((1, 42, 768) ) lowercase__ : int = torch.tensor([[0.0037, 0.1368, -0.0091], [0.1099, 0.3329, -0.1095], [0.0765, 0.5335, 0.1179]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCamelCase__ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": lowercase__ : Optional[int] = torch.Size((1, 1, 1_024) ) lowercase__ : Optional[Any] = torch.tensor([[0.0466, -0.0106, -0.0179]] ) else: # base lowercase__ : int = torch.Size((1, 1, 768) ) lowercase__ : int = torch.tensor([[0.1457, 0.1044, 0.0174]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowerCamelCase__ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print("Saving PyTorch model to {}".format(lowerCamelCase__ ) ) model.save_pretrained(lowerCamelCase__ ) def __lowerCamelCase ( lowerCamelCase__ ): """simple docstring""" lowercase__ : Tuple = {} with open(lowerCamelCase__ , "r" , encoding="utf-8" ) as f: for index, line in enumerate(lowerCamelCase__ ): lowercase__ , lowercase__ : Any = line.rstrip().split("\t" ) lowercase__ : Any = index return entity_vocab if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) lowerCAmelCase__ = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
121
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''BridgeTower/bridgetower-base''': '''https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json''', '''BridgeTower/bridgetower-base-itm-mlm''': ( '''https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json''' ), } class snake_case__(_UpperCamelCase ): """simple docstring""" lowercase_ = """bridgetower_vision_model""" def __init__( self : str , SCREAMING_SNAKE_CASE : Dict=768 , SCREAMING_SNAKE_CASE : Union[str, Any]=12 , SCREAMING_SNAKE_CASE : Dict=3 , SCREAMING_SNAKE_CASE : Dict=16 , SCREAMING_SNAKE_CASE : List[Any]=288 , SCREAMING_SNAKE_CASE : List[Any]=1 , SCREAMING_SNAKE_CASE : List[str]=1E-0_5 , SCREAMING_SNAKE_CASE : str=False , SCREAMING_SNAKE_CASE : Any=True , SCREAMING_SNAKE_CASE : Tuple=False , **SCREAMING_SNAKE_CASE : List[str] , ): super().__init__(**SCREAMING_SNAKE_CASE ) lowercase__ : Union[str, Any] = hidden_size lowercase__ : str = num_hidden_layers lowercase__ : str = num_channels lowercase__ : Optional[int] = patch_size lowercase__ : Dict = image_size lowercase__ : List[Any] = initializer_factor lowercase__ : int = layer_norm_eps lowercase__ : List[str] = stop_gradient lowercase__ : Optional[int] = share_layernorm lowercase__ : Optional[int] = remove_last_layer @classmethod def snake_case ( cls : Tuple , SCREAMING_SNAKE_CASE : Union[str, os.PathLike] , **SCREAMING_SNAKE_CASE : str ): lowercase__ , lowercase__ : Optional[Any] = cls.get_config_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) if config_dict.get("model_type" ) == "bridgetower": lowercase__ : Optional[int] = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) class snake_case__(_UpperCamelCase ): """simple docstring""" lowercase_ = """bridgetower_text_model""" def __init__( self : Tuple , SCREAMING_SNAKE_CASE : Union[str, Any]=50_265 , SCREAMING_SNAKE_CASE : Optional[int]=768 , SCREAMING_SNAKE_CASE : List[str]=12 , SCREAMING_SNAKE_CASE : Union[str, Any]=12 , SCREAMING_SNAKE_CASE : List[Any]=1 , SCREAMING_SNAKE_CASE : Any=3_072 , SCREAMING_SNAKE_CASE : Union[str, Any]="gelu" , SCREAMING_SNAKE_CASE : Optional[int]=0.1 , SCREAMING_SNAKE_CASE : List[str]=0.1 , SCREAMING_SNAKE_CASE : Tuple=514 , SCREAMING_SNAKE_CASE : List[str]=1 , SCREAMING_SNAKE_CASE : Dict=1E-0_5 , SCREAMING_SNAKE_CASE : List[str]=1 , SCREAMING_SNAKE_CASE : str=0 , SCREAMING_SNAKE_CASE : List[Any]=2 , SCREAMING_SNAKE_CASE : Optional[Any]="absolute" , SCREAMING_SNAKE_CASE : int=True , **SCREAMING_SNAKE_CASE : int , ): super().__init__(**SCREAMING_SNAKE_CASE ) lowercase__ : Any = vocab_size lowercase__ : Optional[Any] = hidden_size lowercase__ : Optional[int] = num_hidden_layers lowercase__ : Optional[int] = num_attention_heads lowercase__ : int = hidden_act lowercase__ : Optional[Any] = initializer_factor lowercase__ : Dict = intermediate_size lowercase__ : Tuple = hidden_dropout_prob lowercase__ : List[Any] = attention_probs_dropout_prob lowercase__ : Optional[int] = max_position_embeddings lowercase__ : str = type_vocab_size lowercase__ : str = layer_norm_eps lowercase__ : Dict = position_embedding_type lowercase__ : Optional[Any] = use_cache lowercase__ : List[str] = pad_token_id lowercase__ : Optional[Any] = bos_token_id lowercase__ : Optional[Any] = eos_token_id @classmethod def snake_case ( cls : Optional[int] , SCREAMING_SNAKE_CASE : Union[str, os.PathLike] , **SCREAMING_SNAKE_CASE : str ): lowercase__ , lowercase__ : Optional[Any] = cls.get_config_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) if config_dict.get("model_type" ) == "bridgetower": lowercase__ : List[Any] = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) class snake_case__(_UpperCamelCase ): """simple docstring""" lowercase_ = """bridgetower""" def __init__( self : Tuple , SCREAMING_SNAKE_CASE : Optional[int]=True , SCREAMING_SNAKE_CASE : Dict="gelu" , SCREAMING_SNAKE_CASE : Any=768 , SCREAMING_SNAKE_CASE : Dict=1 , SCREAMING_SNAKE_CASE : Union[str, Any]=1E-0_5 , SCREAMING_SNAKE_CASE : Tuple=False , SCREAMING_SNAKE_CASE : List[Any]="add" , SCREAMING_SNAKE_CASE : Any=12 , SCREAMING_SNAKE_CASE : Any=6 , SCREAMING_SNAKE_CASE : Any=False , SCREAMING_SNAKE_CASE : List[Any]=False , SCREAMING_SNAKE_CASE : str=None , SCREAMING_SNAKE_CASE : Dict=None , **SCREAMING_SNAKE_CASE : List[Any] , ): # TODO: remove this once the Hub files are updated. lowercase__ : int = kwargs.pop("text_config_dict" , SCREAMING_SNAKE_CASE ) lowercase__ : int = kwargs.pop("vision_config_dict" , SCREAMING_SNAKE_CASE ) super().__init__(**SCREAMING_SNAKE_CASE ) lowercase__ : Dict = share_cross_modal_transformer_layers lowercase__ : int = hidden_act lowercase__ : int = hidden_size lowercase__ : Optional[int] = initializer_factor lowercase__ : Optional[int] = layer_norm_eps lowercase__ : str = share_link_tower_layers lowercase__ : Optional[int] = link_tower_type lowercase__ : List[Any] = num_attention_heads lowercase__ : Dict = num_hidden_layers lowercase__ : Dict = tie_word_embeddings lowercase__ : Optional[Any] = init_layernorm_from_vision_encoder if text_config is None: lowercase__ : Optional[Any] = {} logger.info("`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values." ) if vision_config is None: lowercase__ : List[Any] = {} logger.info("`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values." ) lowercase__ : Any = BridgeTowerTextConfig(**SCREAMING_SNAKE_CASE ) lowercase__ : Dict = BridgeTowerVisionConfig(**SCREAMING_SNAKE_CASE ) @classmethod def snake_case ( cls : Optional[int] , SCREAMING_SNAKE_CASE : BridgeTowerTextConfig , SCREAMING_SNAKE_CASE : BridgeTowerVisionConfig , **SCREAMING_SNAKE_CASE : List[str] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **SCREAMING_SNAKE_CASE ) def snake_case ( self : Optional[int] ): lowercase__ : List[str] = copy.deepcopy(self.__dict__ ) lowercase__ : Dict = self.text_config.to_dict() lowercase__ : Tuple = self.vision_config.to_dict() lowercase__ : Optional[Any] = self.__class__.model_type return output
121
1
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __snake_case ( unittest.TestCase ): @property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase : Any =UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : Tuple =self.dummy_uncond_unet UpperCAmelCase : Optional[int] =KarrasVeScheduler() UpperCAmelCase : List[Any] =KarrasVePipeline(unet=snake_case__ , scheduler=snake_case__ ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase : List[str] =torch.manual_seed(0 ) UpperCAmelCase : List[str] =pipe(num_inference_steps=2 , generator=snake_case__ , output_type='''numpy''' ).images UpperCAmelCase : str =torch.manual_seed(0 ) UpperCAmelCase : str =pipe(num_inference_steps=2 , generator=snake_case__ , output_type='''numpy''' , return_dict=snake_case__ )[0] UpperCAmelCase : Any =image[0, -3:, -3:, -1] UpperCAmelCase : List[str] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) UpperCAmelCase : int =np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase : Tuple ='''google/ncsnpp-celebahq-256''' UpperCAmelCase : int =UNetaDModel.from_pretrained(snake_case__ ) UpperCAmelCase : Dict =KarrasVeScheduler() UpperCAmelCase : Union[str, Any] =KarrasVePipeline(unet=snake_case__ , scheduler=snake_case__ ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase : Any =torch.manual_seed(0 ) UpperCAmelCase : Tuple =pipe(num_inference_steps=20 , generator=snake_case__ , output_type='''numpy''' ).images UpperCAmelCase : Optional[int] =image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) UpperCAmelCase : Tuple =np.array([0.578, 0.5811, 0.5924, 0.5809, 0.587, 0.5886, 0.5861, 0.5802, 0.586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
348
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { '''asapp/sew-d-tiny-100k''': '''https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json''', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class __snake_case ( lowerCamelCase__ ): __lowerCamelCase : Optional[Any] = """sew-d""" def __init__( self , snake_case__=32 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3072 , snake_case__=2 , snake_case__=512 , snake_case__=256 , snake_case__=True , snake_case__=True , snake_case__=("p2c", "c2p") , snake_case__="layer_norm" , snake_case__="gelu_python" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.1 , snake_case__=0.0 , snake_case__=0.1 , snake_case__=0.02 , snake_case__=1e-7 , snake_case__=1e-5 , snake_case__="group" , snake_case__="gelu" , snake_case__=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , snake_case__=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , snake_case__=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , snake_case__=False , snake_case__=128 , snake_case__=16 , snake_case__=True , snake_case__=0.05 , snake_case__=10 , snake_case__=2 , snake_case__=0.0 , snake_case__=10 , snake_case__=0 , snake_case__="mean" , snake_case__=False , snake_case__=False , snake_case__=256 , snake_case__=0 , snake_case__=1 , snake_case__=2 , **snake_case__ , ) -> int: '''simple docstring''' super().__init__(**snake_case__ , pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ ) UpperCAmelCase : Union[str, Any] =hidden_size UpperCAmelCase : Union[str, Any] =feat_extract_norm UpperCAmelCase : Optional[Any] =feat_extract_activation UpperCAmelCase : List[str] =list(snake_case__ ) UpperCAmelCase : int =list(snake_case__ ) UpperCAmelCase : List[str] =list(snake_case__ ) UpperCAmelCase : str =conv_bias UpperCAmelCase : Tuple =num_conv_pos_embeddings UpperCAmelCase : Dict =num_conv_pos_embedding_groups UpperCAmelCase : str =len(self.conv_dim ) UpperCAmelCase : Dict =num_hidden_layers UpperCAmelCase : Optional[int] =intermediate_size UpperCAmelCase : List[Any] =squeeze_factor UpperCAmelCase : str =max_position_embeddings UpperCAmelCase : int =position_buckets UpperCAmelCase : Optional[int] =share_att_key UpperCAmelCase : Optional[int] =relative_attention UpperCAmelCase : Tuple =norm_rel_ebd UpperCAmelCase : List[Any] =list(snake_case__ ) UpperCAmelCase : Dict =hidden_act UpperCAmelCase : Optional[int] =num_attention_heads UpperCAmelCase : Any =hidden_dropout UpperCAmelCase : str =attention_dropout UpperCAmelCase : Union[str, Any] =activation_dropout UpperCAmelCase : str =feat_proj_dropout UpperCAmelCase : Union[str, Any] =final_dropout UpperCAmelCase : Optional[int] =layer_norm_eps UpperCAmelCase : str =feature_layer_norm_eps UpperCAmelCase : str =initializer_range UpperCAmelCase : Any =vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect.''' '''It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,''' f'''but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)''' f'''= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase : Union[str, Any] =apply_spec_augment UpperCAmelCase : Optional[Any] =mask_time_prob UpperCAmelCase : Tuple =mask_time_length UpperCAmelCase : str =mask_time_min_masks UpperCAmelCase : Optional[int] =mask_feature_prob UpperCAmelCase : Optional[Any] =mask_feature_length UpperCAmelCase : List[Any] =mask_feature_min_masks # ctc loss UpperCAmelCase : str =ctc_loss_reduction UpperCAmelCase : Optional[int] =ctc_zero_infinity # sequence classification UpperCAmelCase : Union[str, Any] =use_weighted_layer_sum UpperCAmelCase : int =classifier_proj_size @property def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
348
1
"""simple docstring""" from __future__ import annotations import time import numpy as np __magic_name__ = [8, 5, 9, 7] __magic_name__ = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __magic_name__ = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class SCREAMING_SNAKE_CASE_ : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): __SCREAMING_SNAKE_CASE = claim_vector __SCREAMING_SNAKE_CASE = allocated_resources_table __SCREAMING_SNAKE_CASE = maximum_claim_table def snake_case_ ( self): return [ sum(p_item[i] for p_item in self.__allocated_resources_table) for i in range(len(self.__allocated_resources_table[0])) ] def snake_case_ ( self): return np.array(self.__claim_vector) - np.array( self.__processes_resource_summation()) def snake_case_ ( self): return [ list(np.array(self.__maximum_claim_table[i]) - np.array(_lowerCamelCase)) for i, allocated_resource in enumerate(self.__allocated_resources_table) ] def snake_case_ ( self): return {self.__need().index(_lowerCamelCase): i for i in self.__need()} def snake_case_ ( self , **lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.__need() __SCREAMING_SNAKE_CASE = self.__allocated_resources_table __SCREAMING_SNAKE_CASE = self.__available_resources() __SCREAMING_SNAKE_CASE = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("""_""" * 5_0 + """\n""") while need_list: __SCREAMING_SNAKE_CASE = False for each_need in need_list: __SCREAMING_SNAKE_CASE = True for index, need in enumerate(_lowerCamelCase): if need > available_resources[index]: __SCREAMING_SNAKE_CASE = False break if execution: __SCREAMING_SNAKE_CASE = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: __SCREAMING_SNAKE_CASE = original_need_index print(f"Process {process_number + 1} is executing.") # remove the process run from stack need_list.remove(_lowerCamelCase) # update available/freed resources stack __SCREAMING_SNAKE_CASE = np.array(_lowerCamelCase) + np.array( alloc_resources_table[process_number]) print( """Updated available resource stack for processes: """ + """ """.join([str(_lowerCamelCase) for x in available_resources])) break if safe: print("""The process is in a safe state.\n""") else: print("""System in unsafe state. Aborting...\n""") break def snake_case_ ( self): print(""" """ * 9 + """Allocated Resource Table""") for item in self.__allocated_resources_table: print( f"P{self.__allocated_resources_table.index(_lowerCamelCase) + 1}" + """ """.join(f"{it:>8}" for it in item) + """\n""") print(""" """ * 9 + """System Resource Table""") for item in self.__maximum_claim_table: print( f"P{self.__maximum_claim_table.index(_lowerCamelCase) + 1}" + """ """.join(f"{it:>8}" for it in item) + """\n""") print( """Current Usage by Active Processes: """ + """ """.join(str(_lowerCamelCase) for x in self.__claim_vector)) print( """Initial Available Resources: """ + """ """.join(str(_lowerCamelCase) for x in self.__available_resources())) time.sleep(1) if __name__ == "__main__": import doctest doctest.testmod()
369
"""simple docstring""" import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow __magic_name__ = False class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" def snake_case_ ( self , lowerCAmelCase__=3_2): set_seed(0) __SCREAMING_SNAKE_CASE = UNetaDModel(sample_size=lowerCAmelCase__ , in_channels=3 , out_channels=3) __SCREAMING_SNAKE_CASE = torch.optim.SGD(model.parameters() , lr=0.00_01) return model, optimizer @slow def snake_case_ ( self): __SCREAMING_SNAKE_CASE = """cpu""" # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable __SCREAMING_SNAKE_CASE = DDPMScheduler( num_train_timesteps=1_0_0_0 , beta_start=0.00_01 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=lowerCAmelCase__ , ) __SCREAMING_SNAKE_CASE = DDIMScheduler( num_train_timesteps=1_0_0_0 , beta_start=0.00_01 , beta_end=0.02 , beta_schedule="""linear""" , clip_sample=lowerCAmelCase__ , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0) __SCREAMING_SNAKE_CASE = [torch.randn((4, 3, 3_2, 3_2)).clip(-1 , 1).to(lowerCAmelCase__) for _ in range(4)] __SCREAMING_SNAKE_CASE = [torch.randn((4, 3, 3_2, 3_2)).to(lowerCAmelCase__) for _ in range(4)] __SCREAMING_SNAKE_CASE = [torch.randint(0 , 1_0_0_0 , (4,)).long().to(lowerCAmelCase__) for _ in range(4)] # train with a DDPM scheduler __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = self.get_model_optimizer(resolution=3_2) model.train().to(lowerCAmelCase__) for i in range(4): optimizer.zero_grad() __SCREAMING_SNAKE_CASE = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i]) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , timesteps[i]).sample __SCREAMING_SNAKE_CASE = torch.nn.functional.mse_loss(lowerCAmelCase__ , noise[i]) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = self.get_model_optimizer(resolution=3_2) model.train().to(lowerCAmelCase__) for i in range(4): optimizer.zero_grad() __SCREAMING_SNAKE_CASE = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i]) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , timesteps[i]).sample __SCREAMING_SNAKE_CASE = torch.nn.functional.mse_loss(lowerCAmelCase__ , noise[i]) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-5)) self.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-5))
255
0
import mpmath # for roots of unity import numpy as np class __magic_name__ : def __init__( self , __snake_case=None , __snake_case=None ) -> Tuple: '''simple docstring''' # Input as list __a =list(poly_a or [0] )[:] __a =list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() __a =len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() __a =len(self.polyB ) # Add 0 to make lengths equal a power of 2 __a =int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform __a =complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product __a =self.__multiply() def __magic_name__ ( self , __snake_case ) -> List[str]: '''simple docstring''' __a =[[x] for x in self.polyA] if which == 'A' else [[x] for x in self.polyB] # Corner case if len(__snake_case ) <= 1: return dft[0] # __a =self.c_max_length // 2 while next_ncol > 0: __a =[[] for i in range(__snake_case )] __a =self.root**next_ncol # First half of next step __a =1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(__snake_case ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step __a =1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(__snake_case ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update __a =new_dft __a =next_ncol // 2 return dft[0] def __magic_name__ ( self ) -> Tuple: '''simple docstring''' __a =self.__dft('A' ) __a =self.__dft('B' ) __a =[[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT __a =2 while next_ncol <= self.c_max_length: __a =[[] for i in range(__snake_case )] __a =self.root ** (next_ncol // 2) __a =1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update __a =new_inverse_c next_ncol *= 2 # Unpack __a =[round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1J for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self ) -> Optional[Any]: '''simple docstring''' __a ='A = ' + ' + '.join( f'{coef}*x^{i}' for coef, i in enumerate(self.polyA[: self.len_A] ) ) __a ='B = ' + ' + '.join( f'{coef}*x^{i}' for coef, i in enumerate(self.polyB[: self.len_B] ) ) __a ='A*B = ' + ' + '.join( f'{coef}*x^{i}' for coef, i in enumerate(self.product ) ) return f'{a}\n{b}\n{c}' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
218
"""simple docstring""" def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) if len(lowercase__ ) == 1: return True _lowerCamelCase : List[Any] = series[1] - series[0] for index in range(len(lowercase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) _lowerCamelCase : Optional[int] = 0 for val in series: answer += val return answer / len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
96
0
'''simple docstring''' from collections import defaultdict from graphs.minimum_spanning_tree_prims import prisms_algorithm as mst def SCREAMING_SNAKE_CASE__( ) -> Any: '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ = 9, 14 # noqa: F841 UpperCamelCase__ = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] UpperCamelCase__ = defaultdict(_UpperCamelCase ) for nodea, nodea, cost in edges: adjancency[nodea].append([nodea, cost] ) adjancency[nodea].append([nodea, cost] ) UpperCamelCase__ = mst(_UpperCamelCase ) UpperCamelCase__ = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] for answer in expected: UpperCamelCase__ = tuple(answer[:2] ) UpperCamelCase__ = tuple(edge[::-1] ) assert edge in result or reverse in result
31
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE__( _UpperCamelCase : list[int | str] ) -> None: '''simple docstring''' create_state_space_tree(_UpperCamelCase , [] , 0 , [0 for i in range(len(_UpperCamelCase ) )] ) def SCREAMING_SNAKE_CASE__( _UpperCamelCase : list[int | str] , _UpperCamelCase : list[int | str] , _UpperCamelCase : int , _UpperCamelCase : list[int] , ) -> None: '''simple docstring''' if index == len(_UpperCamelCase ): print(_UpperCamelCase ) return for i in range(len(_UpperCamelCase ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCamelCase__ = True create_state_space_tree(_UpperCamelCase , _UpperCamelCase , index + 1 , _UpperCamelCase ) current_sequence.pop() UpperCamelCase__ = False __lowercase: list[int | str] = [3, 1, 2, 4] generate_all_permutations(sequence) __lowercase: list[int | str] = ["A", "B", "C"] generate_all_permutations(sequence_a)
31
1
import math import flax.linen as nn import jax.numpy as jnp def _a ( SCREAMING_SNAKE_CASE_ : jnp.ndarray , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : float = 1 , SCREAMING_SNAKE_CASE_ : float = 1 , SCREAMING_SNAKE_CASE_ : float = 1.0E4 , SCREAMING_SNAKE_CASE_ : bool = False , SCREAMING_SNAKE_CASE_ : float = 1.0 , ): assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, F"""Embedding dimension {embedding_dim} should be even""" __lowerCAmelCase = float(embedding_dim // 2 ) __lowerCAmelCase = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) __lowerCAmelCase = min_timescale * jnp.exp(jnp.arange(SCREAMING_SNAKE_CASE_ , dtype=jnp.floataa ) * -log_timescale_increment ) __lowerCAmelCase = jnp.expand_dims(SCREAMING_SNAKE_CASE_ , 1 ) * jnp.expand_dims(SCREAMING_SNAKE_CASE_ , 0 ) # scale embeddings __lowerCAmelCase = scale * emb if flip_sin_to_cos: __lowerCAmelCase = jnp.concatenate([jnp.cos(SCREAMING_SNAKE_CASE_ ), jnp.sin(SCREAMING_SNAKE_CASE_ )] , axis=1 ) else: __lowerCAmelCase = jnp.concatenate([jnp.sin(SCREAMING_SNAKE_CASE_ ), jnp.cos(SCREAMING_SNAKE_CASE_ )] , axis=1 ) __lowerCAmelCase = jnp.reshape(SCREAMING_SNAKE_CASE_ , [jnp.shape(SCREAMING_SNAKE_CASE_ )[0], embedding_dim] ) return signal class a__ ( nn.Module ): _a : int = 3_2 _a : jnp.dtype = jnp.floataa @nn.compact def __call__( self , _A ): """simple docstring""" __lowerCAmelCase = nn.Dense(self.time_embed_dim , dtype=self.dtype , name="linear_1" )(_A ) __lowerCAmelCase = nn.silu(_A ) __lowerCAmelCase = nn.Dense(self.time_embed_dim , dtype=self.dtype , name="linear_2" )(_A ) return temb class a__ ( nn.Module ): _a : int = 3_2 _a : bool = False _a : float = 1 @nn.compact def __call__( self , _A ): """simple docstring""" return get_sinusoidal_embeddings( _A , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
92
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE_ : def __init__( self : Optional[int] , lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : str=3 , lowerCamelCase_ : Tuple=32 , lowerCamelCase_ : List[str]=3 , lowerCamelCase_ : Optional[int]=10 , lowerCamelCase_ : List[str]=[10, 20, 30, 40] , lowerCamelCase_ : Tuple=[1, 1, 2, 1] , lowerCamelCase_ : Dict=True , lowerCamelCase_ : str=True , lowerCamelCase_ : Tuple="relu" , lowerCamelCase_ : List[str]=3 , lowerCamelCase_ : Dict=None , ): """simple docstring""" UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = image_size UpperCamelCase = num_channels UpperCamelCase = embeddings_size UpperCamelCase = hidden_sizes UpperCamelCase = depths UpperCamelCase = is_training UpperCamelCase = use_labels UpperCamelCase = hidden_act UpperCamelCase = num_labels UpperCamelCase = scope UpperCamelCase = len(lowerCamelCase_ ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase = None if self.use_labels: UpperCamelCase = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def lowerCamelCase_ ( self : Any , lowerCamelCase_ : List[Any] , lowerCamelCase_ : Optional[int] , lowerCamelCase_ : Tuple ): """simple docstring""" UpperCamelCase = TFResNetModel(config=lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowerCamelCase_ ( self : List[Any] , lowerCamelCase_ : List[str] , lowerCamelCase_ : str , lowerCamelCase_ : Optional[int] ): """simple docstring""" UpperCamelCase = self.num_labels UpperCamelCase = TFResNetForImageClassification(lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ , labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase = config_and_inputs UpperCamelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): __lowerCAmelCase = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () __lowerCAmelCase = ( {"""feature-extraction""": TFResNetModel, """image-classification""": TFResNetForImageClassification} if is_tf_available() else {} ) __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False def lowerCamelCase_ ( self : Any ): """simple docstring""" UpperCamelCase = TFResNetModelTester(self ) UpperCamelCase = ConfigTester(self , config_class=lowerCamelCase_ , has_text_modality=lowerCamelCase_ ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" return @unittest.skip(reason="""ResNet does not use inputs_embeds""" ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" pass @unittest.skip(reason="""ResNet does not support input and output embeddings""" ) def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" pass def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase = [*signature.parameters.keys()] UpperCamelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCamelCase_ ) def lowerCamelCase_ ( self : List[str] ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" def check_hidden_states_output(lowerCamelCase_ : Tuple , lowerCamelCase_ : int , lowerCamelCase_ : str ): UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model(**self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) ) UpperCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCamelCase = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase_ ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: UpperCamelCase = layer_type UpperCamelCase = True check_hidden_states_output(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True check_hidden_states_output(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase_ ) @slow def lowerCamelCase_ ( self : Any ): """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase = TFResNetModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) def lowercase( ) -> Any: '''simple docstring''' UpperCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self : Dict ): """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self : Dict ): """simple docstring""" UpperCamelCase = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCamelCase = self.default_image_processor UpperCamelCase = prepare_img() UpperCamelCase = image_processor(images=lowerCamelCase_ , return_tensors="""tf""" ) # forward pass UpperCamelCase = model(**lowerCamelCase_ ) # verify the logits UpperCamelCase = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase_ ) UpperCamelCase = tf.constant([-1_1.1_0_6_9, -9.7_8_7_7, -8.3_7_7_7] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , lowerCamelCase_ , atol=1E-4 ) )
343
0
"""simple docstring""" from abc import ABC, abstractmethod from argparse import ArgumentParser class _snake_case ( a__ ): @staticmethod @abstractmethod def lowerCamelCase__ ( UpperCAmelCase : ArgumentParser ): raise NotImplementedError() @abstractmethod def lowerCamelCase__ ( self : Dict ): raise NotImplementedError()
64
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __A = logging.get_logger(__name__) __A = { '''ut/deta''': '''https://huggingface.co/ut/deta/resolve/main/config.json''', } class _snake_case ( a__ ): snake_case__ = "deta" snake_case__ = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self : Any , UpperCAmelCase : Optional[int]=None , UpperCAmelCase : Dict=900 , UpperCAmelCase : str=2048 , UpperCAmelCase : Optional[Any]=6 , UpperCAmelCase : Optional[Any]=2048 , UpperCAmelCase : Optional[Any]=8 , UpperCAmelCase : Union[str, Any]=6 , UpperCAmelCase : int=1024 , UpperCAmelCase : str=8 , UpperCAmelCase : Dict=0.0 , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Optional[Any]="relu" , UpperCAmelCase : List[str]=256 , UpperCAmelCase : int=0.1 , UpperCAmelCase : int=0.0 , UpperCAmelCase : Any=0.0 , UpperCAmelCase : List[Any]=0.0_2 , UpperCAmelCase : Optional[Any]=1.0 , UpperCAmelCase : Union[str, Any]=True , UpperCAmelCase : Dict=False , UpperCAmelCase : Optional[int]="sine" , UpperCAmelCase : Tuple=5 , UpperCAmelCase : Any=4 , UpperCAmelCase : Dict=4 , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Dict=300 , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Tuple=True , UpperCAmelCase : Union[str, Any]=1 , UpperCAmelCase : str=5 , UpperCAmelCase : Tuple=2 , UpperCAmelCase : str=1 , UpperCAmelCase : Tuple=1 , UpperCAmelCase : Tuple=5 , UpperCAmelCase : List[Any]=2 , UpperCAmelCase : List[Any]=0.1 , UpperCAmelCase : Tuple=0.2_5 , **UpperCAmelCase : List[str] , ): if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) __lowerCamelCase : Tuple = CONFIG_MAPPING["resnet"](out_features=["stage2", "stage3", "stage4"] ) else: if isinstance(UpperCAmelCase , UpperCAmelCase ): __lowerCamelCase : List[Any] = backbone_config.pop("model_type" ) __lowerCamelCase : Any = CONFIG_MAPPING[backbone_model_type] __lowerCamelCase : List[Any] = config_class.from_dict(UpperCAmelCase ) __lowerCamelCase : List[str] = backbone_config __lowerCamelCase : Any = num_queries __lowerCamelCase : Tuple = max_position_embeddings __lowerCamelCase : Dict = d_model __lowerCamelCase : List[str] = encoder_ffn_dim __lowerCamelCase : int = encoder_layers __lowerCamelCase : str = encoder_attention_heads __lowerCamelCase : Dict = decoder_ffn_dim __lowerCamelCase : Tuple = decoder_layers __lowerCamelCase : str = decoder_attention_heads __lowerCamelCase : Dict = dropout __lowerCamelCase : List[Any] = attention_dropout __lowerCamelCase : int = activation_dropout __lowerCamelCase : int = activation_function __lowerCamelCase : Any = init_std __lowerCamelCase : Optional[Any] = init_xavier_std __lowerCamelCase : int = encoder_layerdrop __lowerCamelCase : Dict = auxiliary_loss __lowerCamelCase : Optional[int] = position_embedding_type # deformable attributes __lowerCamelCase : Tuple = num_feature_levels __lowerCamelCase : str = encoder_n_points __lowerCamelCase : List[str] = decoder_n_points __lowerCamelCase : List[str] = two_stage __lowerCamelCase : Dict = two_stage_num_proposals __lowerCamelCase : int = with_box_refine __lowerCamelCase : Union[str, Any] = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError("If two_stage is True, with_box_refine must be True." ) # Hungarian matcher __lowerCamelCase : str = class_cost __lowerCamelCase : Optional[Any] = bbox_cost __lowerCamelCase : Tuple = giou_cost # Loss coefficients __lowerCamelCase : List[Any] = mask_loss_coefficient __lowerCamelCase : Dict = dice_loss_coefficient __lowerCamelCase : Any = bbox_loss_coefficient __lowerCamelCase : Dict = giou_loss_coefficient __lowerCamelCase : Optional[Any] = eos_coefficient __lowerCamelCase : Optional[int] = focal_alpha super().__init__(is_encoder_decoder=UpperCAmelCase , **UpperCAmelCase ) @property def lowerCamelCase__ ( self : Dict ): return self.encoder_attention_heads @property def lowerCamelCase__ ( self : Union[str, Any] ): return self.d_model def lowerCamelCase__ ( self : List[str] ): __lowerCamelCase : Dict = copy.deepcopy(self.__dict__ ) __lowerCamelCase : Dict = self.backbone_config.to_dict() __lowerCamelCase : str = self.__class__.model_type return output
64
1
import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowercase__( UpperCAmelCase , UpperCAmelCase ): """simple docstring""" @register_to_config def __init__( self : Tuple , *, SCREAMING_SNAKE_CASE_ : int = 4 , SCREAMING_SNAKE_CASE_ : int = 7_6_8 , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Optional[Any] , ) -> Optional[int]: super().__init__() lowercase_ = nn.Parameter(torch.zeros(SCREAMING_SNAKE_CASE_ ) ) # parameters for additional clip time embeddings lowercase_ = nn.Linear(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase_ = nn.Linear(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # parameters for encoder hidden states lowercase_ = clip_extra_context_tokens lowercase_ = nn.Linear( SCREAMING_SNAKE_CASE_ , self.clip_extra_context_tokens * cross_attention_dim ) lowercase_ = nn.Linear(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase_ = nn.LayerNorm(SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : Tuple , *, SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Any ) -> List[str]: if do_classifier_free_guidance: # Add the classifier free guidance embeddings to the image embeddings lowercase_ = image_embeddings.shape[0] lowercase_ = self.learned_classifier_free_guidance_embeddings.unsqueeze(0 ) lowercase_ = classifier_free_guidance_embeddings.expand( SCREAMING_SNAKE_CASE_ , -1 ) lowercase_ = torch.cat([classifier_free_guidance_embeddings, image_embeddings] , dim=0 ) # The image embeddings batch size and the text embeddings batch size are equal assert image_embeddings.shape[0] == prompt_embeds.shape[0] lowercase_ = prompt_embeds.shape[0] # "Specifically, we modify the architecture described in Nichol et al. (2021) by projecting and # adding CLIP embeddings to the existing timestep embedding, ... lowercase_ = self.embedding_proj(SCREAMING_SNAKE_CASE_ ) lowercase_ = self.clip_image_embeddings_project_to_time_embeddings(SCREAMING_SNAKE_CASE_ ) lowercase_ = time_projected_image_embeddings + time_projected_prompt_embeds # ... and by projecting CLIP embeddings into four # extra tokens of context that are concatenated to the sequence of outputs from the GLIDE text encoder" lowercase_ = self.clip_extra_context_tokens_proj(SCREAMING_SNAKE_CASE_ ) lowercase_ = clip_extra_context_tokens.reshape(SCREAMING_SNAKE_CASE_ , -1 , self.clip_extra_context_tokens ) lowercase_ = clip_extra_context_tokens.permute(0 , 2 , 1 ) lowercase_ = self.encoder_hidden_states_proj(SCREAMING_SNAKE_CASE_ ) lowercase_ = self.text_encoder_hidden_states_norm(SCREAMING_SNAKE_CASE_ ) lowercase_ = torch.cat([clip_extra_context_tokens, text_encoder_hidden_states] , dim=1 ) return text_encoder_hidden_states, additive_clip_time_embeddings
30
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class snake_case__ : """simple docstring""" def __init__( self : Tuple, _snake_case : Any, _snake_case : int=1_3, _snake_case : Optional[int]=3_2, _snake_case : Tuple=2, _snake_case : Any=3, _snake_case : Tuple=1_6, _snake_case : Tuple=[1, 2, 1], _snake_case : Dict=[2, 2, 4], _snake_case : str=2, _snake_case : Union[str, Any]=2.0, _snake_case : Dict=True, _snake_case : Dict=0.0, _snake_case : str=0.0, _snake_case : str=0.1, _snake_case : List[str]="gelu", _snake_case : int=False, _snake_case : Optional[Any]=True, _snake_case : List[Any]=0.0_2, _snake_case : Union[str, Any]=1e-5, _snake_case : Union[str, Any]=True, _snake_case : List[Any]=None, _snake_case : Any=True, _snake_case : List[Any]=1_0, _snake_case : str=8, ) ->Union[str, Any]: snake_case__ : Any = parent snake_case__ : Tuple = batch_size snake_case__ : Tuple = image_size snake_case__ : Any = patch_size snake_case__ : Optional[int] = num_channels snake_case__ : Tuple = embed_dim snake_case__ : Any = depths snake_case__ : Any = num_heads snake_case__ : List[str] = window_size snake_case__ : Dict = mlp_ratio snake_case__ : Optional[int] = qkv_bias snake_case__ : Optional[Any] = hidden_dropout_prob snake_case__ : List[str] = attention_probs_dropout_prob snake_case__ : Union[str, Any] = drop_path_rate snake_case__ : str = hidden_act snake_case__ : Union[str, Any] = use_absolute_embeddings snake_case__ : Union[str, Any] = patch_norm snake_case__ : Any = layer_norm_eps snake_case__ : Tuple = initializer_range snake_case__ : Dict = is_training snake_case__ : Any = scope snake_case__ : Optional[Any] = use_labels snake_case__ : str = type_sequence_label_size snake_case__ : List[Any] = encoder_stride def lowercase_ ( self : Tuple ) ->str: snake_case__ : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case__ : List[Any] = None if self.use_labels: snake_case__ : Optional[Any] = ids_tensor([self.batch_size], self.type_sequence_label_size ) snake_case__ : Any = self.get_config() return config, pixel_values, labels def lowercase_ ( self : Optional[int] ) ->Optional[int]: return SwinvaConfig( image_size=self.image_size, patch_size=self.patch_size, num_channels=self.num_channels, embed_dim=self.embed_dim, depths=self.depths, num_heads=self.num_heads, window_size=self.window_size, mlp_ratio=self.mlp_ratio, qkv_bias=self.qkv_bias, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, drop_path_rate=self.drop_path_rate, hidden_act=self.hidden_act, use_absolute_embeddings=self.use_absolute_embeddings, path_norm=self.patch_norm, layer_norm_eps=self.layer_norm_eps, initializer_range=self.initializer_range, encoder_stride=self.encoder_stride, ) def lowercase_ ( self : Optional[int], _snake_case : str, _snake_case : List[str], _snake_case : int ) ->Dict: snake_case__ : List[Any] = SwinvaModel(config=_snake_case ) model.to(_snake_case ) model.eval() snake_case__ : Optional[int] = model(_snake_case ) snake_case__ : List[Any] = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case__ : List[Any] = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, expected_seq_len, expected_dim) ) def lowercase_ ( self : Optional[Any], _snake_case : Any, _snake_case : List[str], _snake_case : Dict ) ->List[Any]: snake_case__ : List[str] = SwinvaForMaskedImageModeling(config=_snake_case ) model.to(_snake_case ) model.eval() snake_case__ : Union[str, Any] = model(_snake_case ) self.parent.assertEqual( result.logits.shape, (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case__ : Optional[Any] = 1 snake_case__ : Optional[int] = SwinvaForMaskedImageModeling(_snake_case ) model.to(_snake_case ) model.eval() snake_case__ : Tuple = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case__ : Any = model(_snake_case ) self.parent.assertEqual(result.logits.shape, (self.batch_size, 1, self.image_size, self.image_size) ) def lowercase_ ( self : List[str], _snake_case : int, _snake_case : List[Any], _snake_case : Optional[int] ) ->Any: snake_case__ : Tuple = self.type_sequence_label_size snake_case__ : int = SwinvaForImageClassification(_snake_case ) model.to(_snake_case ) model.eval() snake_case__ : Tuple = model(_snake_case, labels=_snake_case ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.type_sequence_label_size) ) def lowercase_ ( self : Any ) ->Dict: snake_case__ : str = self.prepare_config_and_inputs() snake_case__ , snake_case__ , snake_case__ : List[str] = config_and_inputs snake_case__ : Union[str, Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class snake_case__ ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) _SCREAMING_SNAKE_CASE = ( {"""feature-extraction""": SwinvaModel, """image-classification""": SwinvaForImageClassification} if is_torch_available() else {} ) _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False def lowercase_ ( self : Union[str, Any] ) ->Dict: snake_case__ : Optional[int] = SwinvaModelTester(self ) snake_case__ : int = ConfigTester(self, config_class=_snake_case, embed_dim=3_7 ) def lowercase_ ( self : Tuple ) ->int: self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase_ ( self : Any ) ->str: snake_case__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def lowercase_ ( self : Any ) ->Union[str, Any]: pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def lowercase_ ( self : str ) ->Union[str, Any]: pass def lowercase_ ( self : Optional[Any] ) ->Union[str, Any]: snake_case__ , snake_case__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ : Union[str, Any] = model_class(_snake_case ) self.assertIsInstance(model.get_input_embeddings(), (nn.Module) ) snake_case__ : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_snake_case, nn.Linear ) ) def lowercase_ ( self : List[str] ) ->Optional[int]: snake_case__ , snake_case__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ : Any = model_class(_snake_case ) snake_case__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case__ : Optional[Any] = [*signature.parameters.keys()] snake_case__ : List[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1], _snake_case ) def lowercase_ ( self : str ) ->Union[str, Any]: snake_case__ , snake_case__ : Any = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ : int = True for model_class in self.all_model_classes: snake_case__ : str = True snake_case__ : Union[str, Any] = False snake_case__ : Tuple = True snake_case__ : int = model_class(_snake_case ) model.to(_snake_case ) model.eval() with torch.no_grad(): snake_case__ : Optional[int] = model(**self._prepare_for_class(_snake_case, _snake_case ) ) snake_case__ : List[str] = outputs.attentions snake_case__ : List[Any] = len(self.model_tester.depths ) self.assertEqual(len(_snake_case ), _snake_case ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case__ : str = True snake_case__ : Tuple = config.window_size**2 snake_case__ : Optional[int] = model_class(_snake_case ) model.to(_snake_case ) model.eval() with torch.no_grad(): snake_case__ : str = model(**self._prepare_for_class(_snake_case, _snake_case ) ) snake_case__ : Tuple = outputs.attentions self.assertEqual(len(_snake_case ), _snake_case ) self.assertListEqual( list(attentions[0].shape[-3:] ), [self.model_tester.num_heads[0], window_size_squared, window_size_squared], ) snake_case__ : Optional[Any] = len(_snake_case ) # Check attention is always last and order is fine snake_case__ : Optional[int] = True snake_case__ : Dict = True snake_case__ : List[Any] = model_class(_snake_case ) model.to(_snake_case ) model.eval() with torch.no_grad(): snake_case__ : Optional[int] = model(**self._prepare_for_class(_snake_case, _snake_case ) ) if hasattr(self.model_tester, 'num_hidden_states_types' ): snake_case__ : str = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states snake_case__ : Dict = 2 self.assertEqual(out_len + added_hidden_states, len(_snake_case ) ) snake_case__ : Any = outputs.attentions self.assertEqual(len(_snake_case ), _snake_case ) self.assertListEqual( list(self_attentions[0].shape[-3:] ), [self.model_tester.num_heads[0], window_size_squared, window_size_squared], ) def lowercase_ ( self : Dict, _snake_case : Tuple, _snake_case : Any, _snake_case : int, _snake_case : Optional[int] ) ->str: snake_case__ : Dict = model_class(_snake_case ) model.to(_snake_case ) model.eval() with torch.no_grad(): snake_case__ : List[Any] = model(**self._prepare_for_class(_snake_case, _snake_case ) ) snake_case__ : Dict = outputs.hidden_states snake_case__ : int = getattr( self.model_tester, 'expected_num_hidden_layers', len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_snake_case ), _snake_case ) # Swinv2 has a different seq_length snake_case__ : int = ( config.patch_size if isinstance(config.patch_size, collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case__ : Optional[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ), [num_patches, self.model_tester.embed_dim], ) snake_case__ : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(_snake_case ), _snake_case ) snake_case__ , snake_case__ , snake_case__ , snake_case__ : str = reshaped_hidden_states[0].shape snake_case__ : Any = ( reshaped_hidden_states[0].view(_snake_case, _snake_case, height * width ).permute(0, 2, 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ), [num_patches, self.model_tester.embed_dim], ) def lowercase_ ( self : str ) ->List[Any]: snake_case__ , snake_case__ : Any = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ : List[Any] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size, collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: snake_case__ : Optional[int] = True self.check_hidden_states_output(_snake_case, _snake_case, _snake_case, _snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case__ : Dict = True self.check_hidden_states_output(_snake_case, _snake_case, _snake_case, _snake_case ) def lowercase_ ( self : List[str] ) ->str: snake_case__ , snake_case__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ : List[str] = 3 snake_case__ : Union[str, Any] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size, collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case__ : str = ( config.patch_size if isinstance(config.patch_size, collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case__ : Tuple = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case__ : Optional[Any] = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: snake_case__ : int = True self.check_hidden_states_output(_snake_case, _snake_case, _snake_case, (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case__ : List[str] = True self.check_hidden_states_output(_snake_case, _snake_case, _snake_case, (padded_height, padded_width) ) def lowercase_ ( self : List[str] ) ->Optional[int]: snake_case__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_snake_case ) def lowercase_ ( self : List[Any] ) ->str: snake_case__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_snake_case ) @slow def lowercase_ ( self : str ) ->Union[str, Any]: for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case__ : Dict = SwinvaModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) def lowercase_ ( self : Optional[int] ) ->List[str]: snake_case__ , snake_case__ : Any = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ : List[Any] = _config_zero_init(_snake_case ) for model_class in self.all_model_classes: snake_case__ : List[str] = model_class(config=_snake_case ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item(), [0.0, 1.0], msg=F'''Parameter {name} of model {model_class} seems not properly initialized''', ) @require_vision @require_torch class snake_case__ ( unittest.TestCase ): """simple docstring""" @cached_property def lowercase_ ( self : Union[str, Any] ) ->List[str]: return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def lowercase_ ( self : int ) ->List[Any]: snake_case__ : Any = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( _snake_case ) snake_case__ : int = self.default_image_processor snake_case__ : Union[str, Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) snake_case__ : Optional[Any] = image_processor(images=_snake_case, return_tensors='pt' ).to(_snake_case ) # forward pass with torch.no_grad(): snake_case__ : List[str] = model(**_snake_case ) # verify the logits snake_case__ : int = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape, _snake_case ) snake_case__ : Optional[int] = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(_snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3], _snake_case, atol=1e-4 ) )
277
0
'''simple docstring''' def UpperCamelCase_ ( A__ : str ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = hex_num.strip() if not hex_num: raise ValueError("""No value was passed to the function""" ) lowerCAmelCase_ : Any = hex_num[0] == """-""" if is_negative: lowerCAmelCase_ : Any = hex_num[1:] try: lowerCAmelCase_ : int = int(A__ , 16 ) except ValueError: raise ValueError("""Invalid value was passed to the function""" ) lowerCAmelCase_ : List[Any] = """""" while int_num > 0: lowerCAmelCase_ : List[Any] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(("""-""" + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
89
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __A : Optional[Any] = logging.get_logger(__name__) def UpperCamelCase_ ( A__ : List[Any] , A__ : str=False ): '''simple docstring''' lowerCAmelCase_ : List[str] = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith("""head""" ): lowerCAmelCase_ : str = """segformer.encoder.""" + key if key.startswith("""backbone""" ): lowerCAmelCase_ : str = key.replace("""backbone""" , """segformer.encoder""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCAmelCase_ : List[str] = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] lowerCAmelCase_ : List[Any] = key.replace(f'patch_embed{idx}' , f'patch_embeddings.{int(A__ )-1}' ) if "norm" in key: lowerCAmelCase_ : Any = key.replace("""norm""" , """layer_norm""" ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCAmelCase_ : Tuple = key[key.find("""segformer.encoder.layer_norm""" ) + len("""segformer.encoder.layer_norm""" )] lowerCAmelCase_ : int = key.replace(f'layer_norm{idx}' , f'layer_norm.{int(A__ )-1}' ) if "layer_norm1" in key: lowerCAmelCase_ : str = key.replace("""layer_norm1""" , """layer_norm_1""" ) if "layer_norm2" in key: lowerCAmelCase_ : Union[str, Any] = key.replace("""layer_norm2""" , """layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 lowerCAmelCase_ : Any = key[key.find("""block""" ) + len("""block""" )] lowerCAmelCase_ : str = key.replace(f'block{idx}' , f'block.{int(A__ )-1}' ) if "attn.q" in key: lowerCAmelCase_ : List[Any] = key.replace("""attn.q""" , """attention.self.query""" ) if "attn.proj" in key: lowerCAmelCase_ : Optional[int] = key.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in key: lowerCAmelCase_ : str = key.replace("""attn""" , """attention.self""" ) if "fc1" in key: lowerCAmelCase_ : Optional[Any] = key.replace("""fc1""" , """dense1""" ) if "fc2" in key: lowerCAmelCase_ : Optional[Any] = key.replace("""fc2""" , """dense2""" ) if "linear_pred" in key: lowerCAmelCase_ : List[Any] = key.replace("""linear_pred""" , """classifier""" ) if "linear_fuse" in key: lowerCAmelCase_ : Optional[Any] = key.replace("""linear_fuse.conv""" , """linear_fuse""" ) lowerCAmelCase_ : Any = key.replace("""linear_fuse.bn""" , """batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCAmelCase_ : str = key[key.find("""linear_c""" ) + len("""linear_c""" )] lowerCAmelCase_ : Dict = key.replace(f'linear_c{idx}' , f'linear_c.{int(A__ )-1}' ) if key.startswith("""head""" ): lowerCAmelCase_ : int = key.replace("""head""" , """classifier""" ) lowerCAmelCase_ : int = value return new_state_dict def UpperCamelCase_ ( A__ : int , A__ : Union[str, Any] ): '''simple docstring''' for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCAmelCase_ : int = state_dict.pop(f'segformer.encoder.block.{i}.{j}.attention.self.kv.weight' ) lowerCAmelCase_ : Optional[int] = state_dict.pop(f'segformer.encoder.block.{i}.{j}.attention.self.kv.bias' ) # next, add keys and values (in that order) to the state dict lowerCAmelCase_ : List[str] = kv_weight[ : config.hidden_sizes[i], : ] lowerCAmelCase_ : Optional[Any] = kv_bias[: config.hidden_sizes[i]] lowerCAmelCase_ : Union[str, Any] = kv_weight[ config.hidden_sizes[i] :, : ] lowerCAmelCase_ : str = kv_bias[ config.hidden_sizes[i] : ] def UpperCamelCase_ ( ): '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCAmelCase_ : Optional[Any] = Image.open(requests.get(A__ , stream=A__ ).raw ) return image @torch.no_grad() def UpperCamelCase_ ( A__ : Optional[Any] , A__ : List[Any] , A__ : Tuple ): '''simple docstring''' lowerCAmelCase_ : str = SegformerConfig() lowerCAmelCase_ : Optional[Any] = False # set attributes based on model_name lowerCAmelCase_ : int = """huggingface/label-files""" if "segformer" in model_name: lowerCAmelCase_ : Optional[int] = model_name[len("""segformer.""" ) : len("""segformer.""" ) + 2] if "ade" in model_name: lowerCAmelCase_ : List[Any] = 1_50 lowerCAmelCase_ : int = """ade20k-id2label.json""" lowerCAmelCase_ : Tuple = (1, 1_50, 1_28, 1_28) elif "city" in model_name: lowerCAmelCase_ : List[str] = 19 lowerCAmelCase_ : Dict = """cityscapes-id2label.json""" lowerCAmelCase_ : List[str] = (1, 19, 1_28, 1_28) else: raise ValueError(f'Model {model_name} not supported' ) elif "mit" in model_name: lowerCAmelCase_ : Dict = True lowerCAmelCase_ : Optional[int] = model_name[4:6] lowerCAmelCase_ : Union[str, Any] = 10_00 lowerCAmelCase_ : int = """imagenet-1k-id2label.json""" lowerCAmelCase_ : Optional[Any] = (1, 10_00) else: raise ValueError(f'Model {model_name} not supported' ) # set config attributes lowerCAmelCase_ : Optional[Any] = json.load(open(hf_hub_download(A__ , A__ , repo_type="""dataset""" ) , """r""" ) ) lowerCAmelCase_ : List[Any] = {int(A__ ): v for k, v in idalabel.items()} lowerCAmelCase_ : List[str] = idalabel lowerCAmelCase_ : List[Any] = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": lowerCAmelCase_ : Any = [64, 1_28, 3_20, 5_12] lowerCAmelCase_ : int = 2_56 elif size == "b2": lowerCAmelCase_ : Any = [64, 1_28, 3_20, 5_12] lowerCAmelCase_ : List[str] = 7_68 lowerCAmelCase_ : Any = [3, 4, 6, 3] elif size == "b3": lowerCAmelCase_ : List[str] = [64, 1_28, 3_20, 5_12] lowerCAmelCase_ : Union[str, Any] = 7_68 lowerCAmelCase_ : Union[str, Any] = [3, 4, 18, 3] elif size == "b4": lowerCAmelCase_ : Tuple = [64, 1_28, 3_20, 5_12] lowerCAmelCase_ : Tuple = 7_68 lowerCAmelCase_ : Tuple = [3, 8, 27, 3] elif size == "b5": lowerCAmelCase_ : Union[str, Any] = [64, 1_28, 3_20, 5_12] lowerCAmelCase_ : str = 7_68 lowerCAmelCase_ : Any = [3, 6, 40, 3] else: raise ValueError(f'Size {size} not supported' ) # load image processor (only resize + normalize) lowerCAmelCase_ : List[Any] = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=A__ , align=A__ , do_random_crop=A__ ) # prepare image lowerCAmelCase_ : Optional[Any] = prepare_img() lowerCAmelCase_ : Union[str, Any] = image_processor(images=A__ , return_tensors="""pt""" ).pixel_values logger.info(f'Converting model {model_name}...' ) # load original state dict if encoder_only: lowerCAmelCase_ : str = torch.load(A__ , map_location=torch.device("""cpu""" ) ) else: lowerCAmelCase_ : List[str] = torch.load(A__ , map_location=torch.device("""cpu""" ) )["""state_dict"""] # rename keys lowerCAmelCase_ : Dict = rename_keys(A__ , encoder_only=A__ ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(A__ , A__ ) # create HuggingFace model and load state dict if encoder_only: lowerCAmelCase_ : Dict = False lowerCAmelCase_ : List[Any] = SegformerForImageClassification(A__ ) else: lowerCAmelCase_ : str = SegformerForSemanticSegmentation(A__ ) model.load_state_dict(A__ ) model.eval() # forward pass lowerCAmelCase_ : Tuple = model(A__ ) lowerCAmelCase_ : Union[str, Any] = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": lowerCAmelCase_ : Tuple = torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": lowerCAmelCase_ : List[Any] = torch.tensor( [ [[-7.5820, -8.7231, -8.3215], [-8.0600, -10.3529, -10.0304], [-7.5208, -9.4103, -9.6239]], [[-12.6918, -13.8994, -13.7137], [-13.3196, -15.7523, -15.4789], [-12.9343, -14.8757, -14.9689]], [[-11.1911, -11.9421, -11.3243], [-11.3342, -13.6839, -13.3581], [-10.3909, -12.1832, -12.4858]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": lowerCAmelCase_ : List[Any] = torch.tensor( [ [[-11.8173, -14.3850, -16.3128], [-14.5648, -16.5804, -18.6568], [-14.7223, -15.7387, -18.4218]], [[-15.7290, -17.9171, -19.4423], [-18.3105, -19.9448, -21.4661], [-17.9296, -18.6497, -20.7910]], [[-15.0783, -17.0336, -18.2789], [-16.8771, -18.6870, -20.1612], [-16.2454, -17.1426, -19.5055]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": lowerCAmelCase_ : List[Any] = torch.tensor( [ [[-9.0878, -10.2081, -10.1891], [-9.3144, -10.7941, -10.9843], [-9.2294, -10.3855, -10.5704]], [[-12.2316, -13.9068, -13.6102], [-12.9161, -14.3702, -14.3235], [-12.5233, -13.7174, -13.7932]], [[-14.6275, -15.2490, -14.9727], [-14.3400, -15.9687, -16.2827], [-14.1484, -15.4033, -15.8937]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": lowerCAmelCase_ : List[str] = torch.tensor( [ [[-12.3144, -13.2447, -14.0802], [-13.3614, -14.5816, -15.6117], [-13.3340, -14.4433, -16.2219]], [[-19.2781, -20.4128, -20.7506], [-20.6153, -21.6566, -22.0998], [-19.9800, -21.0430, -22.1494]], [[-18.8739, -19.7804, -21.1834], [-20.1233, -21.6765, -23.2944], [-20.0315, -21.2641, -23.6944]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": lowerCAmelCase_ : List[str] = torch.tensor( [ [[-9.5524, -12.0835, -11.7348], [-10.5229, -13.6446, -14.5662], [-9.5842, -12.8851, -13.9414]], [[-15.3432, -17.5323, -17.0818], [-16.3330, -18.9255, -19.2101], [-15.1340, -17.7848, -18.3971]], [[-12.6072, -14.9486, -14.6631], [-13.7629, -17.0907, -17.7745], [-12.7899, -16.1695, -17.1671]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": lowerCAmelCase_ : Dict = torch.tensor( [ [[-11.9295, -13.4057, -14.8106], [-13.3431, -14.8179, -15.3781], [-14.2836, -15.5942, -16.1588]], [[-11.4906, -12.8067, -13.6564], [-13.1189, -14.0500, -14.1543], [-13.8748, -14.5136, -14.8789]], [[0.5374, 0.1067, -0.4742], [0.1141, -0.2255, -0.7099], [-0.3000, -0.5924, -1.3105]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": lowerCAmelCase_ : List[Any] = torch.tensor( [ [[-7.8217, -9.8767, -10.1717], [-9.4438, -10.9058, -11.4047], [-9.7939, -12.3495, -12.1079]], [[-7.1514, -9.5336, -10.0860], [-9.7776, -11.6822, -11.8439], [-10.1411, -12.7655, -12.8972]], [[0.3021, 0.0805, -0.2310], [-0.0328, -0.1605, -0.2714], [-0.1408, -0.5477, -0.6976]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": lowerCAmelCase_ : Dict = torch.tensor( [ [ [-1.13_72E01, -1.27_87E01, -1.34_77E01], [-1.25_36E01, -1.41_94E01, -1.44_09E01], [-1.32_17E01, -1.48_88E01, -1.53_27E01], ], [ [-1.47_91E01, -1.71_22E01, -1.82_77E01], [-1.71_63E01, -1.91_92E01, -1.95_33E01], [-1.78_97E01, -1.99_91E01, -2.03_15E01], ], [ [7.67_23E-01, 4.19_21E-01, -7.78_78E-02], [4.77_72E-01, 9.55_57E-03, -2.80_82E-01], [3.60_32E-01, -2.48_26E-01, -5.11_68E-01], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": lowerCAmelCase_ : str = torch.tensor( [ [[-9.4959, -11.3087, -11.7479], [-11.0025, -12.6540, -12.3319], [-11.4064, -13.0487, -12.9905]], [[-9.8905, -11.3084, -12.0854], [-11.1726, -12.7698, -12.9583], [-11.5985, -13.3278, -14.1774]], [[0.2213, 0.0192, -0.2466], [-0.1731, -0.4213, -0.4874], [-0.3126, -0.6541, -1.1389]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": lowerCAmelCase_ : Optional[int] = torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": lowerCAmelCase_ : List[Any] = torch.tensor( [ [[-16.0976, -16.4856, -17.3962], [-16.6234, -19.0342, -19.7685], [-16.0900, -18.0661, -19.1180]], [[-18.4750, -18.8488, -19.5074], [-19.4030, -22.1570, -22.5977], [-19.1191, -20.8486, -22.3783]], [[-4.5178, -5.5037, -6.5109], [-5.0884, -7.2174, -8.0334], [-4.4156, -5.8117, -7.2970]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": lowerCAmelCase_ : int = torch.tensor( [ [[-14.2081, -14.4732, -14.1977], [-14.5867, -16.4423, -16.6356], [-13.4441, -14.9685, -16.8696]], [[-14.4576, -14.7073, -15.0451], [-15.0816, -17.6237, -17.9873], [-14.4213, -16.0199, -18.5992]], [[-4.7349, -4.9588, -5.0966], [-4.3210, -6.9325, -7.2591], [-3.4312, -4.7484, -7.1917]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": lowerCAmelCase_ : Union[str, Any] = torch.tensor( [ [[-11.7737, -11.9526, -11.3273], [-13.6692, -14.4574, -13.8878], [-13.8937, -14.6924, -15.9345]], [[-14.6706, -14.5330, -14.1306], [-16.1502, -16.8180, -16.4269], [-16.8338, -17.8939, -20.1746]], [[1.0491, 0.8289, 1.0310], [1.1044, 0.5219, 0.8055], [1.0899, 0.6926, 0.5590]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": lowerCAmelCase_ : List[Any] = torch.tensor( [ [[-12.5641, -13.4777, -13.0684], [-13.9587, -15.8983, -16.6557], [-13.3109, -15.7350, -16.3141]], [[-14.7074, -15.4352, -14.5944], [-16.6353, -18.1663, -18.6120], [-15.1702, -18.0329, -18.1547]], [[-1.7990, -2.0951, -1.7784], [-2.6397, -3.8245, -3.9686], [-1.5264, -2.8126, -2.9316]], ] ) else: lowerCAmelCase_ : Optional[Any] = logits.argmax(-1 ).item() print("""Predicted class:""" , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , A__ , atol=1E-2 ) # finally, save model and image processor logger.info(f'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(A__ ).mkdir(exist_ok=A__ ) model.save_pretrained(A__ ) image_processor.save_pretrained(A__ ) if __name__ == "__main__": __A : Tuple = argparse.ArgumentParser() parser.add_argument( "--model_name", default="segformer.b0.512x512.ade.160k", type=str, help="Name of the model you'd like to convert.", ) parser.add_argument( "--checkpoint_path", default=None, type=str, help="Path to the original PyTorch checkpoint (.pth file)." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) __A : Tuple = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
89
1
'''simple docstring''' import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class lowercase : """simple docstring""" def __init__( self ,a_ ,a_=13 ,a_=64 ,a_=2 ,a_=3 ,a_=True ,a_=True ,a_=32 ,a_=5 ,a_=4 ,a_=37 ,a_="gelu" ,a_=0.1 ,a_=0.1 ,a_=10 ,a_=0.02 ,a_=[1, 16, 4, 4] ,a_=None ,) -> Tuple: _UpperCAmelCase : Dict = parent _UpperCAmelCase : List[str] = batch_size _UpperCAmelCase : Union[str, Any] = image_size _UpperCAmelCase : int = patch_size _UpperCAmelCase : Tuple = num_channels _UpperCAmelCase : int = is_training _UpperCAmelCase : Optional[Any] = use_labels _UpperCAmelCase : Optional[Any] = hidden_size _UpperCAmelCase : List[Any] = num_hidden_layers _UpperCAmelCase : Tuple = num_attention_heads _UpperCAmelCase : Union[str, Any] = intermediate_size _UpperCAmelCase : Any = hidden_act _UpperCAmelCase : Optional[Any] = hidden_dropout_prob _UpperCAmelCase : Optional[Any] = attention_probs_dropout_prob _UpperCAmelCase : Tuple = type_sequence_label_size _UpperCAmelCase : List[Any] = initializer_range _UpperCAmelCase : str = scope _UpperCAmelCase : str = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size _UpperCAmelCase : Dict = (self.image_size // 32) ** 2 _UpperCAmelCase : Tuple = num_patches + 1 def _snake_case ( self ) -> Optional[int]: _UpperCAmelCase : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase : str = None if self.use_labels: _UpperCAmelCase : int = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) _UpperCAmelCase : Any = self.get_config() return config, pixel_values, labels def _snake_case ( self ) -> Any: _UpperCAmelCase : Optional[int] = { """global_padding""": """same""", """layer_type""": """bottleneck""", """depths""": [3, 4, 9], """out_features""": ["""stage1""", """stage2""", """stage3"""], """embedding_dynamic_padding""": True, """hidden_sizes""": [4, 8, 16, 32], """num_groups""": 2, } return ViTHybridConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=a_ ,initializer_range=self.initializer_range ,backbone_featmap_shape=self.backbone_featmap_shape ,backbone_config=a_ ,) def _snake_case ( self ,a_ ,a_ ,a_ ) -> Optional[int]: _UpperCAmelCase : str = ViTHybridModel(config=a_ ) model.to(a_ ) model.eval() _UpperCAmelCase : Union[str, Any] = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _snake_case ( self ,a_ ,a_ ,a_ ) -> Tuple: _UpperCAmelCase : Tuple = self.type_sequence_label_size _UpperCAmelCase : int = ViTHybridForImageClassification(a_ ) model.to(a_ ) model.eval() _UpperCAmelCase : Dict = model(a_ ,labels=a_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def _snake_case ( self ) -> Dict: _UpperCAmelCase : Tuple = self.prepare_config_and_inputs() _UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase : Union[str, Any] = config_and_inputs _UpperCAmelCase : Any = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowercase ( _lowerCamelCase , _lowerCamelCase , unittest.TestCase ): """simple docstring""" UpperCAmelCase = (ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () UpperCAmelCase = ( {"""feature-extraction""": ViTHybridModel, """image-classification""": ViTHybridForImageClassification} if is_torch_available() else {} ) UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = False def _snake_case ( self ) -> Dict: _UpperCAmelCase : Optional[int] = ViTHybridModelTester(self ) _UpperCAmelCase : Optional[int] = ConfigTester(self ,config_class=a_ ,has_text_modality=a_ ,hidden_size=37 ) def _snake_case ( self ) -> List[Any]: self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def _snake_case ( self ) -> Union[str, Any]: pass def _snake_case ( self ) -> Any: _UpperCAmelCase ,_UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase : Union[str, Any] = model_class(a_ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) _UpperCAmelCase : Tuple = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a_ ,nn.Linear ) ) def _snake_case ( self ) -> Optional[int]: _UpperCAmelCase ,_UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase : Any = model_class(a_ ) _UpperCAmelCase : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase : Union[str, Any] = [*signature.parameters.keys()] _UpperCAmelCase : List[Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,a_ ) def _snake_case ( self ) -> List[Any]: _UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def _snake_case ( self ) -> List[str]: _UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a_ ) def _snake_case ( self ) -> Tuple: _UpperCAmelCase ,_UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : str = _config_zero_init(a_ ) for model_class in self.all_model_classes: _UpperCAmelCase : Union[str, Any] = model_class(config=a_ ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": _UpperCAmelCase : Dict = [f'''{name}.{key}''' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() ,[0.0, 1.0] ,msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' ,) @slow def _snake_case ( self ) -> Union[str, Any]: for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase : Any = ViTHybridModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def snake_case_ ( )-> Any: '''simple docstring''' _UpperCAmelCase : List[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class lowercase ( unittest.TestCase ): """simple docstring""" @cached_property def _snake_case ( self ) -> Optional[int]: return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def _snake_case ( self ) -> List[str]: _UpperCAmelCase : int = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( a_ ) _UpperCAmelCase : List[Any] = self.default_image_processor _UpperCAmelCase : Dict = prepare_img() _UpperCAmelCase : List[Any] = image_processor(images=a_ ,return_tensors="""pt""" ).to(a_ ) # forward pass with torch.no_grad(): _UpperCAmelCase : Dict = model(**a_ ) # verify the logits _UpperCAmelCase : Tuple = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape ,a_ ) _UpperCAmelCase : Dict = torch.tensor([-1.9090, -0.4993, -0.2389] ).to(a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,a_ ,atol=1E-4 ) ) @slow @require_accelerate def _snake_case ( self ) -> Dict: _UpperCAmelCase : Optional[Any] = ViTHybridImageProcessor.from_pretrained("""google/vit-hybrid-base-bit-384""" ) _UpperCAmelCase : List[str] = ViTHybridForImageClassification.from_pretrained("""google/vit-hybrid-base-bit-384""" ,device_map="""auto""" ) _UpperCAmelCase : List[str] = prepare_img() _UpperCAmelCase : str = image_processor(images=a_ ,return_tensors="""pt""" ) _UpperCAmelCase : Any = model(**a_ ) _UpperCAmelCase : Dict = outputs.logits # model predicts one of the 1000 ImageNet classes _UpperCAmelCase : List[Any] = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] ,"""tabby, tabby cat""" )
215
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging A_ : Any = logging.get_logger(__name__) A_ : Any = { """microsoft/markuplm-base""": """https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json""", """microsoft/markuplm-large""": """https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json""", } class lowercase ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = """markuplm""" def __init__( self ,a_=30_522 ,a_=768 ,a_=12 ,a_=12 ,a_=3_072 ,a_="gelu" ,a_=0.1 ,a_=0.1 ,a_=512 ,a_=2 ,a_=0.02 ,a_=1E-1_2 ,a_=0 ,a_=0 ,a_=2 ,a_=256 ,a_=1_024 ,a_=216 ,a_=1_001 ,a_=32 ,a_=50 ,a_="absolute" ,a_=True ,a_=None ,**a_ ,) -> Union[str, Any]: super().__init__( pad_token_id=a_ ,bos_token_id=a_ ,eos_token_id=a_ ,**a_ ,) _UpperCAmelCase : Optional[int] = vocab_size _UpperCAmelCase : Tuple = hidden_size _UpperCAmelCase : str = num_hidden_layers _UpperCAmelCase : Dict = num_attention_heads _UpperCAmelCase : int = hidden_act _UpperCAmelCase : Optional[Any] = intermediate_size _UpperCAmelCase : Tuple = hidden_dropout_prob _UpperCAmelCase : List[str] = attention_probs_dropout_prob _UpperCAmelCase : Optional[int] = max_position_embeddings _UpperCAmelCase : Tuple = type_vocab_size _UpperCAmelCase : Dict = initializer_range _UpperCAmelCase : List[Any] = layer_norm_eps _UpperCAmelCase : Optional[Any] = position_embedding_type _UpperCAmelCase : Any = use_cache _UpperCAmelCase : List[Any] = classifier_dropout # additional properties _UpperCAmelCase : Dict = max_depth _UpperCAmelCase : Union[str, Any] = max_xpath_tag_unit_embeddings _UpperCAmelCase : Optional[int] = max_xpath_subs_unit_embeddings _UpperCAmelCase : List[Any] = tag_pad_id _UpperCAmelCase : Tuple = subs_pad_id _UpperCAmelCase : List[str] = xpath_unit_hidden_size
215
1
"""simple docstring""" import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE ( a_ , unittest.TestCase ): """simple docstring""" lowercase__ = None lowercase__ = BloomTokenizerFast lowercase__ = BloomTokenizerFast lowercase__ = True lowercase__ = False lowercase__ = "tokenizer_file" lowercase__ = {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"} def __lowerCAmelCase ( self : Optional[int] ): super().setUp() lowerCAmelCase__ : List[str] = BloomTokenizerFast.from_pretrained('''bigscience/tokenizer''' ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ,**lowercase_ : List[Any] ): kwargs.update(self.special_tokens_map ) return BloomTokenizerFast.from_pretrained(self.tmpdirname ,**lowercase_ ) def __lowerCAmelCase ( self : List[str] ): lowerCAmelCase__ : str = self.get_rust_tokenizer() lowerCAmelCase__ : int = ['''The quick brown fox</s>''', '''jumps over the lazy dog</s>'''] lowerCAmelCase__ : List[str] = [[2_1_7_5, 2_3_7_1_4, 7_3_1_7_3, 1_4_4_2_5_2, 2], [7_7, 1_3_2_6_1_9, 3_4_7_8, 3_6_8, 1_0_9_5_8_6, 3_5_4_3_3, 2]] lowerCAmelCase__ : str = tokenizer.batch_encode_plus(lowercase_ )['''input_ids'''] self.assertListEqual(lowercase_ ,lowercase_ ) lowerCAmelCase__ : int = tokenizer.batch_decode(lowercase_ ) self.assertListEqual(lowercase_ ,lowercase_ ) def __lowerCAmelCase ( self : int ,lowercase_ : Any=6 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): lowerCAmelCase__ : Any = self.rust_tokenizer_class.from_pretrained(lowercase_ ,**lowercase_ ) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input lowerCAmelCase__ : Optional[Any] = '''This is a simple input''' lowerCAmelCase__ : List[Any] = ['''This is a simple input 1''', '''This is a simple input 2'''] lowerCAmelCase__ : Optional[int] = ('''This is a simple input''', '''This is a pair''') lowerCAmelCase__ : List[str] = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests try: tokenizer_r.encode(lowercase_ ,max_length=lowercase_ ) tokenizer_r.encode_plus(lowercase_ ,max_length=lowercase_ ) tokenizer_r.batch_encode_plus(lowercase_ ,max_length=lowercase_ ) tokenizer_r.encode(lowercase_ ,max_length=lowercase_ ) tokenizer_r.batch_encode_plus(lowercase_ ,max_length=lowercase_ ) except ValueError: self.fail('''Bloom Tokenizer should be able to deal with padding''' ) lowerCAmelCase__ : Union[str, Any] = None # Hotfixing padding = None self.assertRaises(lowercase_ ,tokenizer_r.encode ,lowercase_ ,max_length=lowercase_ ,padding='''max_length''' ) # Simple input self.assertRaises(lowercase_ ,tokenizer_r.encode_plus ,lowercase_ ,max_length=lowercase_ ,padding='''max_length''' ) # Simple input self.assertRaises( lowercase_ ,tokenizer_r.batch_encode_plus ,lowercase_ ,max_length=lowercase_ ,padding='''max_length''' ,) # Pair input self.assertRaises(lowercase_ ,tokenizer_r.encode ,lowercase_ ,max_length=lowercase_ ,padding='''max_length''' ) # Pair input self.assertRaises(lowercase_ ,tokenizer_r.encode_plus ,lowercase_ ,max_length=lowercase_ ,padding='''max_length''' ) # Pair input self.assertRaises( lowercase_ ,tokenizer_r.batch_encode_plus ,lowercase_ ,max_length=lowercase_ ,padding='''max_length''' ,) def __lowerCAmelCase ( self : Any ): lowerCAmelCase__ : Any = self.get_rust_tokenizer() lowerCAmelCase__ : Union[str, Any] = load_dataset('''xnli''' ,'''all_languages''' ,split='''test''' ,streaming=lowercase_ ) lowerCAmelCase__ : Any = next(iter(lowercase_ ) )['''premise'''] # pick up one data lowerCAmelCase__ : Optional[Any] = list(sample_data.values() ) lowerCAmelCase__ : str = list(map(tokenizer.encode ,lowercase_ ) ) lowerCAmelCase__ : Optional[Any] = [tokenizer.decode(lowercase_ ,clean_up_tokenization_spaces=lowercase_ ) for x in output_tokens] self.assertListEqual(lowercase_ ,lowercase_ ) def __lowerCAmelCase ( self : Optional[Any] ): # The test has to be overriden because BLOOM uses ALiBi positional embeddings that does not have # any sequence length constraints. This test of the parent class will fail since it relies on the # maximum sequence length of the positoonal embeddings. self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map ) ,1 ) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values() )[0] ) ,1 )
74
"""simple docstring""" from manim import * class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" def __lowerCAmelCase ( self : List[Any] ): lowerCAmelCase__ : List[str] = Rectangle(height=0.5 ,width=0.5 ) lowerCAmelCase__ : List[Any] = Rectangle(height=0.46 ,width=0.46 ).set_stroke(width=0 ) lowerCAmelCase__ : Tuple = [mem.copy() for i in range(6 )] lowerCAmelCase__ : str = [mem.copy() for i in range(6 )] lowerCAmelCase__ : Optional[Any] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Optional[Any] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Optional[int] = VGroup(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Dict = Text('''CPU''' ,font_size=2_4 ) lowerCAmelCase__ : Optional[int] = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0.5 ,aligned_edge=lowercase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowercase_ ) lowerCAmelCase__ : Tuple = [mem.copy() for i in range(4 )] lowerCAmelCase__ : List[Any] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Any = Text('''GPU''' ,font_size=2_4 ) lowerCAmelCase__ : str = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0.5 ,aligned_edge=lowercase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowercase_ ) lowerCAmelCase__ : str = [mem.copy() for i in range(6 )] lowerCAmelCase__ : List[str] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Tuple = Text('''Model''' ,font_size=2_4 ) lowerCAmelCase__ : List[str] = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,buff=0.5 ,aligned_edge=lowercase_ ) model.move_to([3, -1.0, 0] ) self.add(lowercase_ ) lowerCAmelCase__ : Tuple = [] for i, rect in enumerate(lowercase_ ): rect.set_stroke(lowercase_ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) lowerCAmelCase__ : List[Any] = Rectangle(height=0.46 / 4 ,width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(lowercase_ ,opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) ,buff=0.02 ,direction=lowercase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] ,direction=lowercase_ ,buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] ,direction=lowercase_ ,buff=0.0 ) self.add(lowercase_ ) cpu_targs.append(lowercase_ ) lowerCAmelCase__ : List[str] = [mem.copy() for i in range(6 )] lowerCAmelCase__ : Optional[int] = VGroup(*lowercase_ ).arrange(lowercase_ ,buff=0 ) lowerCAmelCase__ : Dict = Text('''Loaded Checkpoint''' ,font_size=2_4 ) lowerCAmelCase__ : List[str] = Group(lowercase_ ,lowercase_ ).arrange(lowercase_ ,aligned_edge=lowercase_ ,buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) lowerCAmelCase__ : List[Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowerCAmelCase__ : Optional[Any] = MarkupText( F'<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model' ,font_size=1_8 ,) key_text.move_to([-5, 2.4, 0] ) self.add(lowercase_ ,lowercase_ ) lowerCAmelCase__ : Tuple = MarkupText( F'<span fgcolor=\'{BLUE}\'>●</span> Checkpoint' ,font_size=1_8 ,) blue_text.next_to(lowercase_ ,DOWN * 2.4 ,aligned_edge=key_text.get_left() ) lowerCAmelCase__ : Tuple = MarkupText( F'Next, a <i><span fgcolor="{BLUE}">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor="{BLUE}">single shard</span>.' ,font_size=2_4 ,) step_a.move_to([2, 2, 0] ) self.play(Write(lowercase_ ) ,Write(lowercase_ ) ) self.play(Write(lowercase_ ,run_time=1 ) ,Create(lowercase_ ,run_time=1 ) ) lowerCAmelCase__ : int = [] lowerCAmelCase__ : Optional[Any] = [] for i, rect in enumerate(lowercase_ ): lowerCAmelCase__ : List[str] = fill.copy().set_fill(lowercase_ ,opacity=0.7 ) target.move_to(lowercase_ ) first_animations.append(GrowFromCenter(lowercase_ ,run_time=1 ) ) lowerCAmelCase__ : List[Any] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowercase_ ,run_time=1.5 ) ) self.play(*lowercase_ ) self.play(*lowercase_ ) self.wait()
74
1
'''simple docstring''' from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=__snake_case ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCAmelCase_ = Features({"text": Value("string" )} ) lowerCAmelCase_ = Features({} ) lowerCAmelCase_ = "text" @property def lowercase (self ) -> Dict[str, str]: return {self.text_column: "text"}
341
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'microsoft/deberta-v2-xlarge': 'https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json', 'microsoft/deberta-v2-xxlarge': 'https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json', 'microsoft/deberta-v2-xlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json' ), 'microsoft/deberta-v2-xxlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json' ), } class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = "deberta-v2" def __init__(self , UpperCAmelCase=128100 , UpperCAmelCase=1536 , UpperCAmelCase=24 , UpperCAmelCase=24 , UpperCAmelCase=6144 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=512 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-7 , UpperCAmelCase=False , UpperCAmelCase=-1 , UpperCAmelCase=0 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=0 , UpperCAmelCase="gelu" , **UpperCAmelCase , ) -> List[str]: super().__init__(**UpperCAmelCase ) _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = relative_attention _snake_case = max_relative_positions _snake_case = pad_token_id _snake_case = position_biased_input # Backwards compatibility if type(UpperCAmelCase ) == str: _snake_case = [x.strip() for x in pos_att_type.lower().split("""|""" )] _snake_case = pos_att_type _snake_case = vocab_size _snake_case = layer_norm_eps _snake_case = kwargs.get("""pooler_hidden_size""" , UpperCAmelCase ) _snake_case = pooler_dropout _snake_case = pooler_hidden_act class _lowerCAmelCase ( __snake_case ): '''simple docstring''' @property def lowercase (self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _snake_case = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _snake_case = {0: """batch""", 1: """sequence"""} if self._config.type_vocab_size > 0: return OrderedDict( [("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis)] ) else: return OrderedDict([("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis)] ) @property def lowercase (self ) -> int: return 12 def lowercase (self , UpperCAmelCase , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = 3 , UpperCAmelCase = 40 , UpperCAmelCase = 40 , UpperCAmelCase = None , ) -> Mapping[str, Any]: _snake_case = super().generate_dummy_inputs(preprocessor=UpperCAmelCase , framework=UpperCAmelCase ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
341
1
'''simple docstring''' import argparse import os import torch from transformers.utils import WEIGHTS_NAME SCREAMING_SNAKE_CASE_: Dict =['small', 'medium', 'large'] SCREAMING_SNAKE_CASE_: Dict ='lm_head.decoder.weight' SCREAMING_SNAKE_CASE_: Dict ='lm_head.weight' def lowerCAmelCase_ ( snake_case_ : str , snake_case_ : str ) -> Dict: '''simple docstring''' UpperCAmelCase_ = torch.load(snake_case_ ) UpperCAmelCase_ = d.pop(snake_case_ ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) torch.save(snake_case_ , os.path.join(snake_case_ , snake_case_ ) ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_: Union[str, Any] =argparse.ArgumentParser() parser.add_argument('--dialogpt_path', default='.', type=str) SCREAMING_SNAKE_CASE_: Optional[Any] =parser.parse_args() for MODEL in DIALOGPT_MODELS: SCREAMING_SNAKE_CASE_: Optional[Any] =os.path.join(args.dialogpt_path, f"{MODEL}_ft.pkl") SCREAMING_SNAKE_CASE_: Dict =f"./DialoGPT-{MODEL}" convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
106
'''simple docstring''' import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def lowerCAmelCase_ ( snake_case_ : Union[str, Any] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = FileLock(str(tmpdir / "foo.lock" ) ) UpperCAmelCase_ = FileLock(str(tmpdir / "foo.lock" ) ) UpperCAmelCase_ = 0.01 with locka.acquire(): with pytest.raises(snake_case_ ): UpperCAmelCase_ = time.time() locka.acquire(snake_case_ ) assert time.time() - _start > timeout def lowerCAmelCase_ ( snake_case_ : int ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = "a" * 10_00 + ".lock" UpperCAmelCase_ = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith(".lock" ) assert not locka._lock_file.endswith(snake_case_ ) assert len(os.path.basename(locka._lock_file ) ) <= 2_55 UpperCAmelCase_ = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(snake_case_ ): locka.acquire(0 )
106
1
"""simple docstring""" import logging from transformers.configuration_utils import PretrainedConfig UpperCAmelCase__ = logging.getLogger(__name__) class lowerCAmelCase__ ( A_ ): __a = """masked_bert""" def __init__( self : Union[str, Any] , _lowerCamelCase : Any=30522 , _lowerCamelCase : Union[str, Any]=768 , _lowerCamelCase : Tuple=12 , _lowerCamelCase : Any=12 , _lowerCamelCase : str=3072 , _lowerCamelCase : str="gelu" , _lowerCamelCase : int=0.1 , _lowerCamelCase : Optional[int]=0.1 , _lowerCamelCase : Dict=512 , _lowerCamelCase : List[Any]=2 , _lowerCamelCase : int=0.0_2 , _lowerCamelCase : Union[str, Any]=1e-12 , _lowerCamelCase : Union[str, Any]=0 , _lowerCamelCase : List[str]="topK" , _lowerCamelCase : Optional[Any]="constant" , _lowerCamelCase : Optional[Any]=0.0 , **_lowerCamelCase : str , ): super().__init__(pad_token_id=_lowerCamelCase , **_lowerCamelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = pruning_method _snake_case = mask_init _snake_case = mask_scale
288
"""simple docstring""" def _UpperCAmelCase ( __lowerCamelCase : Tuple ) -> Any: stooge(__lowerCamelCase , 0 , len(__lowerCamelCase ) - 1 ) return arr def _UpperCAmelCase ( __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : str ) -> int: if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _snake_case , _snake_case = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _snake_case = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(__lowerCamelCase , __lowerCamelCase , (h - t) ) # Recursively sort last 2/3 elements stooge(__lowerCamelCase , i + t , (__lowerCamelCase) ) # Recursively sort first 2/3 elements stooge(__lowerCamelCase , __lowerCamelCase , (h - t) ) if __name__ == "__main__": UpperCAmelCase__ = input('Enter numbers separated by a comma:\n').strip() UpperCAmelCase__ = [int(item) for item in user_input.split(',')] print(stooge_sort(unsorted))
288
1
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( ): return [list(range(1000 - i , -1000 - i , -1 ) ) for i in range(1000 )] UpperCAmelCase: Any = generate_large_matrix() UpperCAmelCase: Dict = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): assert all(row == sorted(__UpperCAmelCase , reverse=__UpperCAmelCase ) for row in grid ) assert all(list(__UpperCAmelCase ) == sorted(__UpperCAmelCase , reverse=__UpperCAmelCase ) for col in zip(*__UpperCAmelCase ) ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : Tuple = 0 _lowercase : List[Any] = len(__UpperCAmelCase ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _lowercase : Tuple = (left + right) // 2 _lowercase : List[Any] = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _lowercase : Dict = mid + 1 else: _lowercase : Dict = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : Any = 0 _lowercase : Optional[int] = len(grid[0] ) for i in range(len(__UpperCAmelCase ) ): _lowercase : Union[str, Any] = find_negative_index(grid[i][:bound] ) total += bound return (len(__UpperCAmelCase ) * len(grid[0] )) - total def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): return len([number for row in grid for number in row if number < 0] ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : Tuple = 0 for row in grid: for i, number in enumerate(__UpperCAmelCase ): if number < 0: total += len(__UpperCAmelCase ) - i break return total def __SCREAMING_SNAKE_CASE ( ): from timeit import timeit print("""Running benchmarks""" ) _lowercase : Tuple = ( """from __main__ import count_negatives_binary_search, """ """count_negatives_brute_force, count_negatives_brute_force_with_break, grid""" ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _lowercase : Dict = timeit(F"""{func}(grid=grid)""" , setup=__UpperCAmelCase , number=500 ) print(F"""{func}() took {time:0.4f} seconds""" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
336
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class UpperCamelCase ( snake_case ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = ["image_processor", "tokenizer"] SCREAMING_SNAKE_CASE_ : Optional[int] = "BlipImageProcessor" SCREAMING_SNAKE_CASE_ : Tuple = "AutoTokenizer" def __init__( self ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ): super().__init__(UpperCAmelCase_ ,UpperCAmelCase_ ) # add QFormer tokenizer _lowercase : Optional[int] = qformer_tokenizer def __call__( self ,UpperCAmelCase_ = None ,UpperCAmelCase_ = None ,UpperCAmelCase_ = True ,UpperCAmelCase_ = False ,UpperCAmelCase_ = None ,UpperCAmelCase_ = None ,UpperCAmelCase_ = 0 ,UpperCAmelCase_ = None ,UpperCAmelCase_ = None ,UpperCAmelCase_ = False ,UpperCAmelCase_ = False ,UpperCAmelCase_ = False ,UpperCAmelCase_ = False ,UpperCAmelCase_ = False ,UpperCAmelCase_ = True ,UpperCAmelCase_ = None ,**UpperCAmelCase_ ,): if images is None and text is None: raise ValueError("""You have to specify at least images or text.""" ) _lowercase : List[Any] = BatchFeature() if text is not None: _lowercase : List[str] = self.tokenizer( text=UpperCAmelCase_ ,add_special_tokens=UpperCAmelCase_ ,padding=UpperCAmelCase_ ,truncation=UpperCAmelCase_ ,max_length=UpperCAmelCase_ ,stride=UpperCAmelCase_ ,pad_to_multiple_of=UpperCAmelCase_ ,return_attention_mask=UpperCAmelCase_ ,return_overflowing_tokens=UpperCAmelCase_ ,return_special_tokens_mask=UpperCAmelCase_ ,return_offsets_mapping=UpperCAmelCase_ ,return_token_type_ids=UpperCAmelCase_ ,return_length=UpperCAmelCase_ ,verbose=UpperCAmelCase_ ,return_tensors=UpperCAmelCase_ ,**UpperCAmelCase_ ,) encoding.update(UpperCAmelCase_ ) _lowercase : Dict = self.qformer_tokenizer( text=UpperCAmelCase_ ,add_special_tokens=UpperCAmelCase_ ,padding=UpperCAmelCase_ ,truncation=UpperCAmelCase_ ,max_length=UpperCAmelCase_ ,stride=UpperCAmelCase_ ,pad_to_multiple_of=UpperCAmelCase_ ,return_attention_mask=UpperCAmelCase_ ,return_overflowing_tokens=UpperCAmelCase_ ,return_special_tokens_mask=UpperCAmelCase_ ,return_offsets_mapping=UpperCAmelCase_ ,return_token_type_ids=UpperCAmelCase_ ,return_length=UpperCAmelCase_ ,verbose=UpperCAmelCase_ ,return_tensors=UpperCAmelCase_ ,**UpperCAmelCase_ ,) _lowercase : str = qformer_text_encoding.pop("""input_ids""" ) _lowercase : int = qformer_text_encoding.pop("""attention_mask""" ) if images is not None: _lowercase : Optional[Any] = self.image_processor(UpperCAmelCase_ ,return_tensors=UpperCAmelCase_ ) encoding.update(UpperCAmelCase_ ) return encoding def lowerCamelCase__ ( self ,*UpperCAmelCase_ ,**UpperCAmelCase_ ): return self.tokenizer.batch_decode(*UpperCAmelCase_ ,**UpperCAmelCase_ ) def lowerCamelCase__ ( self ,*UpperCAmelCase_ ,**UpperCAmelCase_ ): return self.tokenizer.decode(*UpperCAmelCase_ ,**UpperCAmelCase_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def lowerCamelCase__ ( self ): _lowercase : Union[str, Any] = self.tokenizer.model_input_names _lowercase : int = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def lowerCamelCase__ ( self ,UpperCAmelCase_ ,**UpperCAmelCase_ ): if os.path.isfile(UpperCAmelCase_ ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(UpperCAmelCase_ ,exist_ok=UpperCAmelCase_ ) _lowercase : Any = os.path.join(UpperCAmelCase_ ,"""qformer_tokenizer""" ) self.qformer_tokenizer.save_pretrained(UpperCAmelCase_ ) return super().save_pretrained(UpperCAmelCase_ ,**UpperCAmelCase_ ) @classmethod def lowerCamelCase__ ( cls ,UpperCAmelCase_ ,**UpperCAmelCase_ ): _lowercase : Optional[Any] = AutoTokenizer.from_pretrained(UpperCAmelCase_ ,subfolder="""qformer_tokenizer""" ) _lowercase : Any = cls._get_arguments_from_pretrained(UpperCAmelCase_ ,**UpperCAmelCase_ ) args.append(UpperCAmelCase_ ) return cls(*UpperCAmelCase_ )
336
1
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase : Optional[int] = LayoutLMTokenizer __UpperCamelCase : List[Any] = LayoutLMTokenizerFast __UpperCamelCase : int = True __UpperCamelCase : Any = True def __magic_name__ ( self : Tuple ): """simple docstring""" super().setUp() _A: Union[str, Any] = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] _A: Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __magic_name__ ( self : int , **lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase_ ) def __magic_name__ ( self : int , lowerCAmelCase_ : List[str] ): """simple docstring""" _A: Dict = '''UNwant\u00E9d,running''' _A: Union[str, Any] = '''unwanted, running''' return input_text, output_text def __magic_name__ ( self : Any ): """simple docstring""" _A: List[str] = self.tokenizer_class(self.vocab_file ) _A: Tuple = tokenizer.tokenize('''UNwant\u00E9d,running''' ) self.assertListEqual(lowerCAmelCase_ , ['''un''', '''##want''', '''##ed''', ''',''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [7, 4, 5, 1_0, 8, 9] ) def __magic_name__ ( self : Optional[Any] ): """simple docstring""" pass
121
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Sequence, Value from .base import TaskTemplate @dataclass(frozen=SCREAMING_SNAKE_CASE__ ) class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization __UpperCamelCase : str = field(default='''question-answering-extractive''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) __UpperCamelCase : ClassVar[Features] = Features({'''question''': Value('''string''' ), '''context''': Value('''string''' )} ) __UpperCamelCase : ClassVar[Features] = Features( { '''answers''': Sequence( { '''text''': Value('''string''' ), '''answer_start''': Value('''int32''' ), } ) } ) __UpperCamelCase : str = "question" __UpperCamelCase : str = "context" __UpperCamelCase : str = "answers" @property def __magic_name__ ( self : List[str] ): """simple docstring""" return {self.question_column: "question", self.context_column: "context", self.answers_column: "answers"}
121
1
"""simple docstring""" from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class lowercase__ : pass
351
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging __UpperCamelCase : List[str] = logging.get_logger(__name__) __UpperCamelCase : List[Any] = '▁' __UpperCamelCase : str = {'vocab_file': 'sentencepiece.bpe.model'} __UpperCamelCase : Tuple = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), } } __UpperCamelCase : List[str] = { 'facebook/mbart-large-en-ro': 1024, 'facebook/mbart-large-cc25': 1024, } # fmt: off __UpperCamelCase : Dict = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class lowercase__ ( UpperCamelCase_): UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = ["""input_ids""", """attention_mask"""] UpperCamelCase_ = [] UpperCamelCase_ = [] def __init__( self : Dict , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[int]="<s>" , UpperCamelCase__ : Union[str, Any]="</s>" , UpperCamelCase__ : str="</s>" , UpperCamelCase__ : List[Any]="<s>" , UpperCamelCase__ : List[str]="<unk>" , UpperCamelCase__ : Tuple="<pad>" , UpperCamelCase__ : str="<mask>" , UpperCamelCase__ : Any=None , UpperCamelCase__ : Dict=None , UpperCamelCase__ : str=None , UpperCamelCase__ : Optional[Dict[str, Any]] = None , UpperCamelCase__ : List[Any]=None , **UpperCamelCase__ : Dict , ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token SCREAMING_SNAKE_CASE : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , src_lang=UpperCamelCase__ , tgt_lang=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE : Dict = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token SCREAMING_SNAKE_CASE : str = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab SCREAMING_SNAKE_CASE : Tuple = 1 SCREAMING_SNAKE_CASE : Dict = len(self.sp_model ) SCREAMING_SNAKE_CASE : Union[str, Any] = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(UpperCamelCase__ ) } SCREAMING_SNAKE_CASE : Any = {v: k for k, v in self.lang_code_to_id.items()} SCREAMING_SNAKE_CASE : Optional[Any] = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) SCREAMING_SNAKE_CASE : Any = {v: k for k, v in self.fairseq_tokens_to_ids.items()} SCREAMING_SNAKE_CASE : int = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) SCREAMING_SNAKE_CASE : List[str] = src_lang if src_lang is not None else '''en_XX''' SCREAMING_SNAKE_CASE : Any = self.lang_code_to_id[self._src_lang] SCREAMING_SNAKE_CASE : int = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self : Union[str, Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = self.__dict__.copy() SCREAMING_SNAKE_CASE : int = None SCREAMING_SNAKE_CASE : Any = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple , UpperCamelCase__ : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): SCREAMING_SNAKE_CASE : Tuple = {} SCREAMING_SNAKE_CASE : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def __A ( self : int ): '''simple docstring''' return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def __A ( self : Union[str, Any] ): '''simple docstring''' return self._src_lang @src_lang.setter def __A ( self : Tuple , UpperCamelCase__ : str ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __A ( self : Union[str, Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Union[str, Any] = [1] * len(self.prefix_tokens ) SCREAMING_SNAKE_CASE : Dict = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(UpperCamelCase__ )) + suffix_ones return prefix_ones + ([0] * len(UpperCamelCase__ )) + ([0] * len(UpperCamelCase__ )) + suffix_ones def __A ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __A ( self : Optional[int] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = [self.sep_token_id] SCREAMING_SNAKE_CASE : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __A ( self : Tuple , UpperCamelCase__ : Any , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] , UpperCamelCase__ : Optional[str] , **UpperCamelCase__ : Tuple ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) SCREAMING_SNAKE_CASE : Dict = src_lang SCREAMING_SNAKE_CASE : Union[str, Any] = self(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , return_tensors=UpperCamelCase__ , **UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[int] = self.convert_tokens_to_ids(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Any = tgt_lang_id return inputs def __A ( self : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = {self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __A ( self : Any , UpperCamelCase__ : str ): '''simple docstring''' return self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) def __A ( self : str , UpperCamelCase__ : int ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] SCREAMING_SNAKE_CASE : Optional[Any] = self.sp_model.PieceToId(UpperCamelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __A ( self : Dict , UpperCamelCase__ : str ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def __A ( self : int , UpperCamelCase__ : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = ''''''.join(UpperCamelCase__ ).replace(UpperCamelCase__ , ''' ''' ).strip() return out_string def __A ( self : Dict , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(UpperCamelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return SCREAMING_SNAKE_CASE : Union[str, Any] = os.path.join( UpperCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , '''wb''' ) as fi: SCREAMING_SNAKE_CASE : List[str] = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,) def __A ( self : Any , UpperCamelCase__ : List[str] , UpperCamelCase__ : str = "en_XX" , UpperCamelCase__ : Optional[List[str]] = None , UpperCamelCase__ : str = "ro_RO" , **UpperCamelCase__ : List[str] , ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = src_lang SCREAMING_SNAKE_CASE : List[str] = tgt_lang return super().prepare_seqaseq_batch(UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) def __A ( self : List[Any] ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def __A ( self : List[str] ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __A ( self : str , UpperCamelCase__ : str ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = self.lang_code_to_id[src_lang] SCREAMING_SNAKE_CASE : Tuple = [] SCREAMING_SNAKE_CASE : Optional[Any] = [self.eos_token_id, self.cur_lang_code] def __A ( self : List[Any] , UpperCamelCase__ : str ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = self.lang_code_to_id[lang] SCREAMING_SNAKE_CASE : Optional[int] = [] SCREAMING_SNAKE_CASE : List[str] = [self.eos_token_id, self.cur_lang_code]
258
0
# Algorithm for the pigeonhole sorting def A_ ( _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Optional[Any] = min(_UpperCAmelCase ) # min() finds the minimum value SCREAMING_SNAKE_CASE_: Tuple = max(_UpperCAmelCase ) # max() finds the maximum value SCREAMING_SNAKE_CASE_: str = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size SCREAMING_SNAKE_CASE_: List[Any] = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(_UpperCAmelCase , _UpperCAmelCase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. SCREAMING_SNAKE_CASE_: int = 0 for count in range(_UpperCAmelCase ): while holes[count] > 0: holes[count] -= 1 SCREAMING_SNAKE_CASE_: str = count + min_val i += 1 def A_ ( ): SCREAMING_SNAKE_CASE_: List[str] = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(_UpperCAmelCase ) print("Sorted order is:" , " ".join(_UpperCAmelCase ) ) if __name__ == "__main__": main()
13
"""simple docstring""" from __future__ import annotations def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase = None ) -> list[list[str]]: '''simple docstring''' lowercase : str = word_bank or [] # create a table lowercase : int = len(_UpperCAmelCase ) + 1 lowercase : list[list[list[str]]] = [] for _ in range(_UpperCAmelCase ): table.append([] ) # seed value lowercase : int = [[]] # because empty string has empty combination # iterate through the indices for i in range(_UpperCAmelCase ): # condition if table[i] != []: for word in word_bank: # slice condition if target[i : i + len(_UpperCAmelCase )] == word: lowercase : list[list[str]] = [ [word, *way] for way in table[i] ] # adds the word to every combination the current position holds # now,push that combination to the table[i+len(word)] table[i + len(_UpperCAmelCase )] += new_combinations # combinations are in reverse order so reverse for better output for combination in table[len(_UpperCAmelCase )]: combination.reverse() return table[len(_UpperCAmelCase )] if __name__ == "__main__": print(all_construct('jwajalapa', ['jwa', 'j', 'w', 'a', 'la', 'lapa'])) print(all_construct('rajamati', ['s', 'raj', 'amat', 'raja', 'ma', 'i', 't'])) print( all_construct( 'hexagonosaurus', ['h', 'ex', 'hex', 'ag', 'ago', 'ru', 'auru', 'rus', 'go', 'no', 'o', 's'], ) )
255
0
'''simple docstring''' from __future__ import annotations def lowerCAmelCase_ ( snake_case__ , snake_case__ ): '''simple docstring''' A : Dict = sorted(numsa + numsa ) A : Any = divmod(len(snake_case__ ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() lowercase : List[Any] = [float(x) for x in input('Enter the elements of first array: ').split()] lowercase : Any = [float(x) for x in input('Enter the elements of second array: ').split()] print(f'''The median of two arrays is: {median_of_two_arrays(array_a, array_a)}''')
357
'''simple docstring''' import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class A : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=13 , SCREAMING_SNAKE_CASE=7 , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=99 , SCREAMING_SNAKE_CASE=32 , SCREAMING_SNAKE_CASE=5 , SCREAMING_SNAKE_CASE=4 , SCREAMING_SNAKE_CASE=37 , SCREAMING_SNAKE_CASE="gelu" , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=50 , SCREAMING_SNAKE_CASE=0.02 , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=None , ) -> str: """simple docstring""" A : Any = parent A : List[Any] = batch_size A : Union[str, Any] = seq_length A : Any = is_training A : int = use_input_mask A : Union[str, Any] = vocab_size A : List[Any] = hidden_size A : List[Any] = num_hidden_layers A : Optional[int] = num_attention_heads A : str = intermediate_size A : Tuple = hidden_act A : Union[str, Any] = hidden_dropout_prob A : Union[str, Any] = attention_probs_dropout_prob A : int = max_position_embeddings A : Optional[int] = initializer_range A : Any = use_labels A : Optional[int] = scope def __lowerCAmelCase ( self ) -> List[Any]: """simple docstring""" A : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A : Optional[int] = None if self.use_input_mask: A : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: A : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A : Dict = self.get_config() return config, input_ids, input_mask, token_labels def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" ( ( A ), ( A ), ( A ), ( A ), ) : Any = self.prepare_config_and_inputs() A : Tuple = True A : int = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE , ) -> Any: """simple docstring""" A : List[str] = BertGenerationEncoder(config=SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : List[Any] = model(SCREAMING_SNAKE_CASE , attention_mask=SCREAMING_SNAKE_CASE ) A : int = model(SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE , ) -> Union[str, Any]: """simple docstring""" A : List[str] = True A : Union[str, Any] = BertGenerationEncoder(config=SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : str = model( SCREAMING_SNAKE_CASE , attention_mask=SCREAMING_SNAKE_CASE , encoder_hidden_states=SCREAMING_SNAKE_CASE , encoder_attention_mask=SCREAMING_SNAKE_CASE , ) A : List[Any] = model( SCREAMING_SNAKE_CASE , attention_mask=SCREAMING_SNAKE_CASE , encoder_hidden_states=SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE , ) -> List[str]: """simple docstring""" A : Optional[Any] = True A : Tuple = True A : Optional[int] = BertGenerationDecoder(config=SCREAMING_SNAKE_CASE ).to(SCREAMING_SNAKE_CASE ).eval() # first forward pass A : str = model( SCREAMING_SNAKE_CASE , attention_mask=SCREAMING_SNAKE_CASE , encoder_hidden_states=SCREAMING_SNAKE_CASE , encoder_attention_mask=SCREAMING_SNAKE_CASE , use_cache=SCREAMING_SNAKE_CASE , ) A : Optional[int] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids A : List[str] = ids_tensor((self.batch_size, 3) , config.vocab_size ) A : Tuple = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and A : Dict = torch.cat([input_ids, next_tokens] , dim=-1 ) A : List[str] = torch.cat([input_mask, next_mask] , dim=-1 ) A : str = model( SCREAMING_SNAKE_CASE , attention_mask=SCREAMING_SNAKE_CASE , encoder_hidden_states=SCREAMING_SNAKE_CASE , encoder_attention_mask=SCREAMING_SNAKE_CASE , output_hidden_states=SCREAMING_SNAKE_CASE , )['''hidden_states'''][0] A : Any = model( SCREAMING_SNAKE_CASE , attention_mask=SCREAMING_SNAKE_CASE , encoder_hidden_states=SCREAMING_SNAKE_CASE , encoder_attention_mask=SCREAMING_SNAKE_CASE , past_key_values=SCREAMING_SNAKE_CASE , output_hidden_states=SCREAMING_SNAKE_CASE , )['''hidden_states'''][0] # select random slice A : int = ids_tensor((1,) , output_from_past.shape[-1] ).item() A : List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() A : str = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , atol=1e-3 ) ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , *SCREAMING_SNAKE_CASE , ) -> Any: """simple docstring""" A : Optional[Any] = BertGenerationDecoder(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : Optional[Any] = model(SCREAMING_SNAKE_CASE , attention_mask=SCREAMING_SNAKE_CASE , labels=SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self ) -> str: """simple docstring""" A, A, A, A : Optional[int] = self.prepare_config_and_inputs() A : str = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class A ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): __magic_name__ = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () __magic_name__ = (BertGenerationDecoder,) if is_torch_available() else () __magic_name__ = ( {'''feature-extraction''': BertGenerationEncoder, '''text-generation''': BertGenerationDecoder} if is_torch_available() else {} ) def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" A : List[str] = BertGenerationEncoderTester(self ) A : Union[str, Any] = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE , hidden_size=37 ) def __lowerCAmelCase ( self ) -> List[Any]: """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" A : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A, A, A, A : Tuple = self.model_tester.prepare_config_and_inputs() A : str = '''bert''' self.model_tester.create_and_check_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> int: """simple docstring""" A : int = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> List[Any]: """simple docstring""" A : List[str] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" ( ( A ), ( A ), ( A ), ( A ), ( A ), ( A ), ) : Tuple = self.model_tester.prepare_config_and_inputs_for_decoder() A : Union[str, Any] = None self.model_tester.create_and_check_model_as_decoder( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) def __lowerCAmelCase ( self ) -> List[Any]: """simple docstring""" A : Dict = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*SCREAMING_SNAKE_CASE ) @slow def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A : Optional[Any] = BertGenerationEncoder.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''' ) self.assertIsNotNone(SCREAMING_SNAKE_CASE ) @require_torch class A ( unittest.TestCase ): @slow def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" A : Tuple = BertGenerationEncoder.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''' ) A : Optional[Any] = torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 10140, 102]] ) with torch.no_grad(): A : Dict = model(SCREAMING_SNAKE_CASE )[0] A : Optional[Any] = torch.Size([1, 8, 1024] ) self.assertEqual(output.shape , SCREAMING_SNAKE_CASE ) A : Dict = torch.tensor( [[[0.1_775, 0.0_083, -0.0_321], [1.6_002, 0.1_287, 0.3_912], [2.1_473, 0.5_791, 0.6_066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) ) @require_torch class A ( unittest.TestCase ): @slow def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" A : Optional[Any] = BertGenerationDecoder.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''' ) A : Dict = torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 10140, 102]] ) with torch.no_grad(): A : Optional[Any] = model(SCREAMING_SNAKE_CASE )[0] A : Optional[Any] = torch.Size([1, 8, 50358] ) self.assertEqual(output.shape , SCREAMING_SNAKE_CASE ) A : Any = torch.tensor( [[[-0.5_788, -2.5_994, -3.7_054], [0.0_438, 4.7_997, 1.8_795], [1.5_862, 6.6_409, 4.4_638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) )
311
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: __SCREAMING_SNAKE_CASE : Optional[Any] = None __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Union[str, Any] = """▁""" __SCREAMING_SNAKE_CASE : str = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} __SCREAMING_SNAKE_CASE : int = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } __SCREAMING_SNAKE_CASE : str = { """google/pegasus-xsum""": 512, } class lowerCamelCase_ (snake_case__ ): '''simple docstring''' __UpperCamelCase: Optional[int] = VOCAB_FILES_NAMES __UpperCamelCase: Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase: List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase: Optional[int] = PegasusTokenizer __UpperCamelCase: Optional[Any] = ["input_ids", "attention_mask"] def __init__( self : Dict , A : List[str]=None , A : Union[str, Any]=None , A : Optional[int]="<pad>" , A : Tuple="</s>" , A : Union[str, Any]="<unk>" , A : Union[str, Any]="<mask_2>" , A : Dict="<mask_1>" , A : Union[str, Any]=None , A : int=103 , **A : Optional[Any] , ): _UpperCAmelCase : Dict = offset if additional_special_tokens is not None: if not isinstance(A , A ): raise TypeError( F"""additional_special_tokens should be of type {type(A )}, but is""" F""" {type(A )}""" ) _UpperCAmelCase : Optional[int] = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"""<unk_{i}>""" for i in range(len(A ) , self.offset - 1 ) ] if len(set(A ) ) != len(A ): raise ValueError( "Please make sure that the provided additional_special_tokens do not contain an incorrectly" F""" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.""" ) _UpperCAmelCase : Any = additional_special_tokens_extended else: _UpperCAmelCase : Dict = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"""<unk_{i}>""" for i in range(2 , self.offset )] super().__init__( A , tokenizer_file=A , pad_token=A , eos_token=A , unk_token=A , mask_token=A , mask_token_sent=A , offset=A , additional_special_tokens=A , **A , ) _UpperCAmelCase : Optional[Any] = vocab_file _UpperCAmelCase : Optional[Any] = False if not self.vocab_file else True def _A ( self : List[str] , A : Optional[Any] ): _UpperCAmelCase : Any = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( "There should be 3 special tokens: mask_token, pad_token, and eos_token +" F""" {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}""" ) return [1 if x in all_special_ids else 0 for x in seq] def _A ( self : str , A : List , A : Optional[List] = None , A : bool = False ): if already_has_special_tokens: return self._special_token_mask(A ) elif token_ids_a is None: return self._special_token_mask(A ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A ( self : Optional[int] , A : Union[str, Any] , A : int=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A ( self : Union[str, Any] , A : str , A : Optional[str] = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(A ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _UpperCAmelCase : List[Any] = os.path.join( A , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A ): copyfile(self.vocab_file , A ) return (out_vocab_file,)
31
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer __SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : str = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} __SCREAMING_SNAKE_CASE : Dict = { """vocab_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-german-cased""": ( """https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json""" ), """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json""" ), }, } __SCREAMING_SNAKE_CASE : Optional[Any] = { """distilbert-base-uncased""": 512, """distilbert-base-uncased-distilled-squad""": 512, """distilbert-base-cased""": 512, """distilbert-base-cased-distilled-squad""": 512, """distilbert-base-german-cased""": 512, """distilbert-base-multilingual-cased""": 512, } __SCREAMING_SNAKE_CASE : List[Any] = { """distilbert-base-uncased""": {"""do_lower_case""": True}, """distilbert-base-uncased-distilled-squad""": {"""do_lower_case""": True}, """distilbert-base-cased""": {"""do_lower_case""": False}, """distilbert-base-cased-distilled-squad""": {"""do_lower_case""": False}, """distilbert-base-german-cased""": {"""do_lower_case""": False}, """distilbert-base-multilingual-cased""": {"""do_lower_case""": False}, } class lowerCamelCase_ (snake_case__ ): '''simple docstring''' __UpperCamelCase: Union[str, Any] = VOCAB_FILES_NAMES __UpperCamelCase: str = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase: Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase: Any = PRETRAINED_INIT_CONFIGURATION __UpperCamelCase: str = ["input_ids", "attention_mask"] __UpperCamelCase: List[str] = DistilBertTokenizer def __init__( self : str , A : int=None , A : Tuple=None , A : Tuple=True , A : Dict="[UNK]" , A : List[Any]="[SEP]" , A : Optional[Any]="[PAD]" , A : Dict="[CLS]" , A : Tuple="[MASK]" , A : str=True , A : Dict=None , **A : List[Any] , ): super().__init__( A , tokenizer_file=A , do_lower_case=A , unk_token=A , sep_token=A , pad_token=A , cls_token=A , mask_token=A , tokenize_chinese_chars=A , strip_accents=A , **A , ) _UpperCAmelCase : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , A ) != do_lower_case or normalizer_state.get("strip_accents" , A ) != strip_accents or normalizer_state.get("handle_chinese_chars" , A ) != tokenize_chinese_chars ): _UpperCAmelCase : Dict = getattr(A , normalizer_state.pop("type" ) ) _UpperCAmelCase : int = do_lower_case _UpperCAmelCase : Optional[int] = strip_accents _UpperCAmelCase : str = tokenize_chinese_chars _UpperCAmelCase : List[Any] = normalizer_class(**A ) _UpperCAmelCase : Dict = do_lower_case def _A ( self : List[Any] , A : Tuple , A : Any=None ): _UpperCAmelCase : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _A ( self : int , A : List[int] , A : Optional[List[int]] = None ): _UpperCAmelCase : Any = [self.sep_token_id] _UpperCAmelCase : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _A ( self : Dict , A : str , A : Optional[str] = None ): _UpperCAmelCase : Any = self._tokenizer.model.save(A , name=A ) return tuple(A )
31
1
import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowerCamelCase_ : Any = False class a__ ( unittest.TestCase ): pass @slow @require_torch_gpu class a__ ( unittest.TestCase ): def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: __a = VersatileDiffusionImageVariationPipeline.from_pretrained('shi-labs/versatile-diffusion' ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __a = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg' ) __a = torch.manual_seed(0 ) __a = pipe( image=SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=7.5 , num_inference_steps=5_0 , output_type='numpy' , ).images __a = image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __a = np.array([0.0_441, 0.0_469, 0.0_507, 0.0_575, 0.0_632, 0.0_650, 0.0_865, 0.0_909, 0.0_945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
365
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCamelCase_ : Dict = {"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : int = [ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Any = [ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys lowerCamelCase_ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
197
0
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType A_ = logging.get_logger(__name__) A_ = { '''microsoft/deberta-v2-xlarge''': '''https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json''', '''microsoft/deberta-v2-xxlarge''': '''https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json''', '''microsoft/deberta-v2-xlarge-mnli''': ( '''https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json''' ), '''microsoft/deberta-v2-xxlarge-mnli''': ( '''https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json''' ), } class lowercase( __a ): '''simple docstring''' lowercase__ = "deberta-v2" def __init__( self: List[str], a_: Tuple=128_100, a_: Union[str, Any]=1_536, a_: Optional[Any]=24, a_: Union[str, Any]=24, a_: Dict=6_144, a_: Optional[int]="gelu", a_: Dict=0.1, a_: Dict=0.1, a_: Optional[int]=512, a_: Dict=0, a_: List[str]=0.02, a_: Optional[int]=1E-7, a_: Dict=False, a_: Optional[int]=-1, a_: Optional[int]=0, a_: str=True, a_: Dict=None, a_: Union[str, Any]=0, a_: str="gelu", **a_: List[str], ): '''simple docstring''' super().__init__(**a_ ) _snake_case : Union[str, Any] = hidden_size _snake_case : Tuple = num_hidden_layers _snake_case : Any = num_attention_heads _snake_case : Any = intermediate_size _snake_case : Optional[int] = hidden_act _snake_case : List[Any] = hidden_dropout_prob _snake_case : int = attention_probs_dropout_prob _snake_case : Optional[int] = max_position_embeddings _snake_case : List[Any] = type_vocab_size _snake_case : Optional[Any] = initializer_range _snake_case : List[Any] = relative_attention _snake_case : Any = max_relative_positions _snake_case : Union[str, Any] = pad_token_id _snake_case : Any = position_biased_input # Backwards compatibility if type(a_ ) == str: _snake_case : Optional[int] = [x.strip() for x in pos_att_type.lower().split("""|""" )] _snake_case : Optional[int] = pos_att_type _snake_case : str = vocab_size _snake_case : Optional[Any] = layer_norm_eps _snake_case : Optional[Any] = kwargs.get("""pooler_hidden_size""", a_ ) _snake_case : Tuple = pooler_dropout _snake_case : List[str] = pooler_hidden_act class lowercase( __a ): '''simple docstring''' @property def UpperCamelCase_ ( self: str ): '''simple docstring''' if self.task == "multiple-choice": _snake_case : Any = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _snake_case : Dict = {0: """batch""", 1: """sequence"""} if self._config.type_vocab_size > 0: return OrderedDict( [("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis)] ) else: return OrderedDict([("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis)] ) @property def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' return 12 def UpperCamelCase_ ( self: str, a_: Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"], a_: int = -1, a_: int = -1, a_: int = -1, a_: bool = False, a_: Optional["TensorType"] = None, a_: int = 3, a_: int = 40, a_: int = 40, a_: "PreTrainedTokenizerBase" = None, ): '''simple docstring''' _snake_case : Optional[int] = super().generate_dummy_inputs(preprocessor=a_, framework=a_ ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
64
"""simple docstring""" from __future__ import annotations from collections.abc import Sequence from typing import Literal def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str ): """simple docstring""" _snake_case : Optional[Any] = list(snake_case__ ) _snake_case : List[Any] = list(snake_case__ ) _snake_case : List[Any] = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count += 1 _snake_case : Any = """_""" if count > 1: return False else: return "".join(snake_case__ ) def UpperCAmelCase__ (snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [] while True: _snake_case : Union[str, Any] = ["""$"""] * len(snake_case__ ) _snake_case : int = [] for i in range(len(snake_case__ ) ): for j in range(i + 1 , len(snake_case__ ) ): _snake_case : List[Any] = compare_string(binary[i] , binary[j] ) if k is False: _snake_case : Dict = """*""" _snake_case : List[Any] = """*""" temp.append("""X""" ) for i in range(len(snake_case__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(snake_case__ ) == 0: return pi _snake_case : Optional[int] = list(set(snake_case__ ) ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Sequence[float] ): """simple docstring""" _snake_case : Optional[int] = [] for minterm in minterms: _snake_case : Any = """""" for _ in range(snake_case__ ): _snake_case : Optional[Any] = str(minterm % 2 ) + string minterm //= 2 temp.append(snake_case__ ) return temp def UpperCAmelCase__ (snake_case__ : str , snake_case__ : str , snake_case__ : int ): """simple docstring""" _snake_case : Dict = list(snake_case__ ) _snake_case : List[str] = list(snake_case__ ) _snake_case : Tuple = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def UpperCAmelCase__ (snake_case__ : list[list[int]] , snake_case__ : list[str] ): """simple docstring""" _snake_case : Any = [] _snake_case : Union[str, Any] = [0] * len(snake_case__ ) for i in range(len(chart[0] ) ): _snake_case : Tuple = 0 _snake_case : str = -1 for j in range(len(snake_case__ ) ): if chart[j][i] == 1: count += 1 _snake_case : Union[str, Any] = j if count == 1: _snake_case : Union[str, Any] = 1 for i in range(len(snake_case__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(snake_case__ ) ): _snake_case : List[Any] = 0 temp.append(prime_implicants[i] ) while True: _snake_case : Optional[int] = 0 _snake_case : str = -1 _snake_case : Any = 0 for i in range(len(snake_case__ ) ): _snake_case : Union[str, Any] = chart[i].count(1 ) if count_n > max_n: _snake_case : Dict = count_n _snake_case : Dict = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(snake_case__ ) ): _snake_case : Optional[Any] = 0 def UpperCAmelCase__ (snake_case__ : list[str] , snake_case__ : list[str] ): """simple docstring""" _snake_case : int = [[0 for x in range(len(snake_case__ ) )] for x in range(len(snake_case__ ) )] for i in range(len(snake_case__ ) ): _snake_case : Any = prime_implicants[i].count("""_""" ) for j in range(len(snake_case__ ) ): if is_for_table(prime_implicants[i] , binary[j] , snake_case__ ): _snake_case : Tuple = 1 return chart def UpperCAmelCase__ (): """simple docstring""" _snake_case : int = int(input("""Enter the no. of variables\n""" ) ) _snake_case : List[str] = [ float(snake_case__ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] _snake_case : List[str] = decimal_to_binary(snake_case__ , snake_case__ ) _snake_case : str = check(snake_case__ ) print("""Prime Implicants are:""" ) print(snake_case__ ) _snake_case : int = prime_implicant_chart(snake_case__ , snake_case__ ) _snake_case : str = selection(snake_case__ , snake_case__ ) print("""Essential Prime Implicants are:""" ) print(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
64
1
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case__ : Any = logging.get_logger(__name__) snake_case__ : Any = """▁""" snake_case__ : str = {"""vocab_file""": """spiece.model"""} snake_case__ : Dict = { """vocab_file""": { """google/reformer-crime-and-punishment""": ( """https://huggingface.co/google/reformer-crime-and-punishment/resolve/main/spiece.model""" ) } } snake_case__ : List[Any] = { """google/reformer-crime-and-punishment""": 52_4288, } class __SCREAMING_SNAKE_CASE ( lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ :List[Any] = VOCAB_FILES_NAMES lowerCamelCase_ :List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ :List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ :str = ["input_ids", "attention_mask"] def __init__( self , snake_case_ , snake_case_="</s>" , snake_case_="<unk>" , snake_case_=[] , snake_case_ = None , **snake_case_ , ): '''simple docstring''' UpperCAmelCase_ : Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=a__ , unk_token=a__ , additional_special_tokens=a__ , sp_model_kwargs=self.sp_model_kwargs , **a__ , ) UpperCAmelCase_ : List[str] = vocab_file UpperCAmelCase_ : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(a__ ) @property def _UpperCamelCase ( self ): '''simple docstring''' return self.sp_model.get_piece_size() def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : int = {self.convert_ids_to_tokens(a__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = self.__dict__.copy() UpperCAmelCase_ : List[Any] = None return state def __setstate__( self , snake_case_ ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): UpperCAmelCase_ : int = {} UpperCAmelCase_ : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCamelCase ( self , snake_case_ ): '''simple docstring''' return self.sp_model.encode(a__ , out_type=a__ ) def _UpperCamelCase ( self , snake_case_ ): '''simple docstring''' return self.sp_model.piece_to_id(a__ ) def _UpperCamelCase ( self , snake_case_ ): '''simple docstring''' if index < self.sp_model.get_piece_size(): UpperCAmelCase_ : str = self.sp_model.IdToPiece(a__ ) return token def _UpperCamelCase ( self , snake_case_ ): '''simple docstring''' UpperCAmelCase_ : List[str] = [] UpperCAmelCase_ : Optional[int] = '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(a__ ) + token UpperCAmelCase_ : Tuple = [] else: current_sub_tokens.append(a__ ) out_string += self.sp_model.decode(a__ ) return out_string.strip() def _UpperCamelCase ( self , snake_case_ , snake_case_ = None ): '''simple docstring''' if not os.path.isdir(a__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase_ : str = os.path.join( a__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , a__ ) elif not os.path.isfile(self.vocab_file ): with open(a__ , 'wb' ) as fi: UpperCAmelCase_ : List[str] = self.sp_model.serialized_model_proto() fi.write(a__ ) return (out_vocab_file,)
364
'''simple docstring''' import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence from datasets.features import ArrayaD, ClassLabel, Features, Image, Value from datasets.features.features import ArrayaDExtensionType, cast_to_python_objects from datasets.keyhash import DuplicatedKeysError, InvalidKeyError from .utils import require_pil class __SCREAMING_SNAKE_CASE ( lowerCamelCase_ ): '''simple docstring''' def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = pa.array(TypedSequence([1, 2, 3] ) ) self.assertEqual(arr.type , pa.intaa() ) def _UpperCamelCase ( self ): '''simple docstring''' with self.assertRaises(snake_case_ ): UpperCAmelCase_ : Optional[int] = pa.array(TypedSequence([1, 2, 3] ) , type=pa.intaa() ) def _UpperCamelCase ( self ): '''simple docstring''' with self.assertRaises(snake_case_ ): UpperCAmelCase_ : List[Any] = pa.array(TypedSequence([1, 2, 3] , try_type=Value('bool' ) , type=Value('int64' ) ) ) def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Any = pa.array(TypedSequence([1, 2, 3] , type=Value('int32' ) ) ) self.assertEqual(arr.type , pa.intaa() ) def _UpperCamelCase ( self ): '''simple docstring''' with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): UpperCAmelCase_ : List[str] = pa.array(TypedSequence(['foo', 'bar'] , type=Value('int64' ) ) ) def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : str = pa.array(TypedSequence([1, 2, 3] , try_type=Value('int32' ) ) ) self.assertEqual(arr.type , pa.intaa() ) def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : List[str] = pa.array(TypedSequence(['foo', 'bar'] , try_type=Value('int64' ) ) ) self.assertEqual(arr.type , pa.string() ) def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Tuple = pa.array(TypedSequence([[[1, 2, 3]]] , type=ArrayaD((1, 3) , 'int64' ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , 'int64' ) ) def _UpperCamelCase ( self ): '''simple docstring''' with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): UpperCAmelCase_ : Any = pa.array(TypedSequence(['foo', 'bar'] , type=ArrayaD((1, 3) , 'int64' ) ) ) def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : int = pa.array(TypedSequence([[[1, 2, 3]]] , try_type=ArrayaD((1, 3) , 'int64' ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , 'int64' ) ) def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = pa.array(TypedSequence(['foo', 'bar'] , try_type=ArrayaD((1, 3) , 'int64' ) ) ) self.assertEqual(arr.type , pa.string() ) @require_pil def _UpperCamelCase ( self ): '''simple docstring''' import PIL.Image UpperCAmelCase_ : Any = PIL.Image.fromarray(np.arange(1_0 , dtype=np.uinta ).reshape(2 , 5 ) ) with patch( 'datasets.arrow_writer.cast_to_python_objects' , side_effect=snake_case_ ) as mock_cast_to_python_objects: UpperCAmelCase_ : int = pa.array(TypedSequence([{'path': None, 'bytes': b'image_bytes'}, pil_image] , type=Image() ) ) UpperCAmelCase_ , UpperCAmelCase_ : Any = mock_cast_to_python_objects.call_args_list[-1] self.assertIn('optimize_list_casting' , snake_case_ ) self.assertFalse(kwargs['optimize_list_casting'] ) def _lowerCamelCase ( lowerCamelCase_ : Any , lowerCamelCase_ : int ): """simple docstring""" UpperCAmelCase_ : str = pa.BufferReader(lowerCamelCase_ ) if isinstance(lowerCamelCase_ , pa.Buffer ) else pa.memory_map(lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = pa.ipc.open_stream(lowerCamelCase_ ) UpperCAmelCase_ : pa.Table = f.read_all() assert len(pa_table.to_batches() ) == expected_num_chunks assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} del pa_table @pytest.mark.parametrize('writer_batch_size' , [None, 1, 10] ) @pytest.mark.parametrize( 'fields' , [None, {'col_1': pa.string(), 'col_2': pa.intaa()}, {'col_1': pa.string(), 'col_2': pa.intaa()}] ) def _lowerCamelCase ( lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : int ): """simple docstring""" UpperCAmelCase_ : List[str] = pa.BufferOutputStream() UpperCAmelCase_ : Dict = pa.schema(lowerCamelCase_ ) if fields else None with ArrowWriter(stream=lowerCamelCase_ , schema=lowerCamelCase_ , writer_batch_size=lowerCamelCase_ ) as writer: writer.write({'col_1': 'foo', 'col_2': 1} ) writer.write({'col_1': 'bar', 'col_2': 2} ) UpperCAmelCase_ , UpperCAmelCase_ : Dict = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: UpperCAmelCase_ : Optional[int] = {'col_1': pa.string(), 'col_2': pa.intaa()} assert writer._schema == pa.schema(lowerCamelCase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def _lowerCamelCase ( ): """simple docstring""" UpperCAmelCase_ : Tuple = pa.BufferOutputStream() UpperCAmelCase_ : Optional[int] = Features({'labels': ClassLabel(names=['neg', 'pos'] )} ) with ArrowWriter(stream=lowerCamelCase_ , features=lowerCamelCase_ ) as writer: writer.write({'labels': 0} ) writer.write({'labels': 1} ) UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == features.arrow_schema assert writer._schema.metadata == features.arrow_schema.metadata UpperCAmelCase_ : Any = pa.BufferReader(output.getvalue() ) UpperCAmelCase_ : List[str] = pa.ipc.open_stream(lowerCamelCase_ ) UpperCAmelCase_ : pa.Table = f.read_all() UpperCAmelCase_ : Optional[Any] = pa_table.schema assert pa_table.num_rows == 2 assert schema == features.arrow_schema assert schema.metadata == features.arrow_schema.metadata assert features == Features.from_arrow_schema(lowerCamelCase_ ) @pytest.mark.parametrize('writer_batch_size' , [None, 1, 10] ) def _lowerCamelCase ( lowerCamelCase_ : List[Any] ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = pa.BufferOutputStream() with ArrowWriter( stream=lowerCamelCase_ , writer_batch_size=lowerCamelCase_ , hash_salt='split_name' , check_duplicates=lowerCamelCase_ , ) as writer: with pytest.raises(lowerCamelCase_ ): writer.write({'col_1': 'foo', 'col_2': 1} , key=[1, 2] ) UpperCAmelCase_ , UpperCAmelCase_ : Any = writer.finalize() @pytest.mark.parametrize('writer_batch_size' , [None, 2, 10] ) def _lowerCamelCase ( lowerCamelCase_ : Union[str, Any] ): """simple docstring""" UpperCAmelCase_ : int = pa.BufferOutputStream() with ArrowWriter( stream=lowerCamelCase_ , writer_batch_size=lowerCamelCase_ , hash_salt='split_name' , check_duplicates=lowerCamelCase_ , ) as writer: with pytest.raises(lowerCamelCase_ ): writer.write({'col_1': 'foo', 'col_2': 1} , key=10 ) writer.write({'col_1': 'bar', 'col_2': 2} , key=10 ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = writer.finalize() @pytest.mark.parametrize('writer_batch_size' , [None, 2, 10] ) def _lowerCamelCase ( lowerCamelCase_ : Optional[int] ): """simple docstring""" UpperCAmelCase_ : str = pa.BufferOutputStream() with ArrowWriter( stream=lowerCamelCase_ , writer_batch_size=lowerCamelCase_ , hash_salt='split_name' , check_duplicates=lowerCamelCase_ , ) as writer: writer.write({'col_1': 'foo', 'col_2': 1} , key=1 ) writer.write({'col_1': 'bar', 'col_2': 2} , key=2 ) UpperCAmelCase_ , UpperCAmelCase_ : Dict = writer.finalize() assert num_examples == 2 assert num_bytes > 0 _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('writer_batch_size' , [None, 1, 10] ) @pytest.mark.parametrize( 'fields' , [None, {'col_1': pa.string(), 'col_2': pa.intaa()}, {'col_1': pa.string(), 'col_2': pa.intaa()}] ) def _lowerCamelCase ( lowerCamelCase_ : Dict , lowerCamelCase_ : str ): """simple docstring""" UpperCAmelCase_ : List[Any] = pa.BufferOutputStream() UpperCAmelCase_ : Optional[Any] = pa.schema(lowerCamelCase_ ) if fields else None with ArrowWriter(stream=lowerCamelCase_ , schema=lowerCamelCase_ , writer_batch_size=lowerCamelCase_ ) as writer: writer.write_batch({'col_1': ['foo', 'bar'], 'col_2': [1, 2]} ) writer.write_batch({'col_1': [], 'col_2': []} ) UpperCAmelCase_ , UpperCAmelCase_ : str = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: UpperCAmelCase_ : int = {'col_1': pa.string(), 'col_2': pa.intaa()} assert writer._schema == pa.schema(lowerCamelCase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('writer_batch_size' , [None, 1, 10] ) @pytest.mark.parametrize( 'fields' , [None, {'col_1': pa.string(), 'col_2': pa.intaa()}, {'col_1': pa.string(), 'col_2': pa.intaa()}] ) def _lowerCamelCase ( lowerCamelCase_ : int , lowerCamelCase_ : Union[str, Any] ): """simple docstring""" UpperCAmelCase_ : List[str] = pa.BufferOutputStream() UpperCAmelCase_ : Optional[int] = pa.schema(lowerCamelCase_ ) if fields else None with ArrowWriter(stream=lowerCamelCase_ , schema=lowerCamelCase_ , writer_batch_size=lowerCamelCase_ ) as writer: writer.write_table(pa.Table.from_pydict({'col_1': ['foo', 'bar'], 'col_2': [1, 2]} ) ) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: UpperCAmelCase_ : Any = {'col_1': pa.string(), 'col_2': pa.intaa()} assert writer._schema == pa.schema(lowerCamelCase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize('writer_batch_size' , [None, 1, 10] ) @pytest.mark.parametrize( 'fields' , [None, {'col_1': pa.string(), 'col_2': pa.intaa()}, {'col_1': pa.string(), 'col_2': pa.intaa()}] ) def _lowerCamelCase ( lowerCamelCase_ : Dict , lowerCamelCase_ : List[str] ): """simple docstring""" UpperCAmelCase_ : Tuple = pa.BufferOutputStream() UpperCAmelCase_ : Tuple = pa.schema(lowerCamelCase_ ) if fields else None with ArrowWriter(stream=lowerCamelCase_ , schema=lowerCamelCase_ , writer_batch_size=lowerCamelCase_ ) as writer: writer.write_row(pa.Table.from_pydict({'col_1': ['foo'], 'col_2': [1]} ) ) writer.write_row(pa.Table.from_pydict({'col_1': ['bar'], 'col_2': [2]} ) ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: UpperCAmelCase_ : Union[str, Any] = {'col_1': pa.string(), 'col_2': pa.intaa()} assert writer._schema == pa.schema(lowerCamelCase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def _lowerCamelCase ( ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase_ : Union[str, Any] = {'col_1': pa.string(), 'col_2': pa.intaa()} UpperCAmelCase_ : Tuple = os.path.join(lowerCamelCase_ , 'test.arrow' ) with ArrowWriter(path=lowerCamelCase_ , schema=pa.schema(lowerCamelCase_ ) ) as writer: writer.write_batch({'col_1': ['foo', 'bar'], 'col_2': [1, 2]} ) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == pa.schema(lowerCamelCase_ , metadata=writer._schema.metadata ) _check_output(lowerCamelCase_ , 1 ) def _lowerCamelCase ( lowerCamelCase_ : Optional[Any] ): """simple docstring""" if pa.types.is_list(lowerCamelCase_ ): return get_base_dtype(arr_type.value_type ) else: return arr_type def _lowerCamelCase ( lowerCamelCase_ : Optional[int] , lowerCamelCase_ : Optional[Any] ): """simple docstring""" if isinstance(lst[0] , lowerCamelCase_ ): change_first_primitive_element_in_list(lst[0] , lowerCamelCase_ ) else: UpperCAmelCase_ : str = value @pytest.mark.parametrize('optimized_int_type, expected_dtype' , [(None, pa.intaa()), (Value('int32' ), pa.intaa())] ) @pytest.mark.parametrize('sequence' , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def _lowerCamelCase ( lowerCamelCase_ : int , lowerCamelCase_ : Dict , lowerCamelCase_ : int ): """simple docstring""" UpperCAmelCase_ : List[Any] = pa.array(TypedSequence(lowerCamelCase_ , optimized_int_type=lowerCamelCase_ ) ) assert get_base_dtype(arr.type ) == expected_dtype @pytest.mark.parametrize( 'col, expected_dtype' , [ ('attention_mask', pa.inta()), ('special_tokens_mask', pa.inta()), ('token_type_ids', pa.inta()), ('input_ids', pa.intaa()), ('other', pa.intaa()), ] , ) @pytest.mark.parametrize('sequence' , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def _lowerCamelCase ( lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : Tuple , lowerCamelCase_ : List[Any] ): """simple docstring""" UpperCAmelCase_ : Tuple = pa.array(OptimizedTypedSequence(lowerCamelCase_ , col=lowerCamelCase_ ) ) assert get_base_dtype(arr.type ) == expected_dtype # not in range if col != "other": # avoids errors due to in-place modifications UpperCAmelCase_ : int = copy.deepcopy(lowerCamelCase_ ) UpperCAmelCase_ : str = np.iinfo(expected_dtype.to_pandas_dtype() ).max + 1 change_first_primitive_element_in_list(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase_ : Dict = pa.array(OptimizedTypedSequence(lowerCamelCase_ , col=lowerCamelCase_ ) ) assert get_base_dtype(arr.type ) == pa.intaa() @pytest.mark.parametrize('raise_exception' , [False, True] ) def _lowerCamelCase ( lowerCamelCase_ : List[str] , lowerCamelCase_ : List[Any] ): """simple docstring""" UpperCAmelCase_ : int = str(tmp_path / 'dataset-train.arrow' ) try: with ArrowWriter(path=lowerCamelCase_ ) as writer: if raise_exception: raise pa.lib.ArrowInvalid() else: writer.stream.close() except pa.lib.ArrowInvalid: pass finally: assert writer.stream.closed def _lowerCamelCase ( lowerCamelCase_ : Dict ): """simple docstring""" UpperCAmelCase_ : Dict = 'mock://dataset-train.arrow' with ArrowWriter(path=lowerCamelCase_ , storage_options=mockfs.storage_options ) as writer: assert isinstance(writer._fs , type(lowerCamelCase_ ) ) assert writer._fs.storage_options == mockfs.storage_options writer.write({'col_1': 'foo', 'col_2': 1} ) writer.write({'col_1': 'bar', 'col_2': 2} ) UpperCAmelCase_ , UpperCAmelCase_ : Tuple = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert mockfs.exists(lowerCamelCase_ ) def _lowerCamelCase ( ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = pa.BufferOutputStream() with ParquetWriter(stream=lowerCamelCase_ ) as writer: writer.write({'col_1': 'foo', 'col_2': 1} ) writer.write({'col_1': 'bar', 'col_2': 2} ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 UpperCAmelCase_ : List[str] = pa.BufferReader(output.getvalue() ) UpperCAmelCase_ : pa.Table = pq.read_table(lowerCamelCase_ ) assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} @require_pil @pytest.mark.parametrize('embed_local_files' , [False, True] ) def _lowerCamelCase ( lowerCamelCase_ : List[Any] , lowerCamelCase_ : Union[str, Any] ): """simple docstring""" import PIL.Image UpperCAmelCase_ : str = str(tmp_path / 'test_image_rgb.jpg' ) PIL.Image.fromarray(np.zeros((5, 5) , dtype=np.uinta ) ).save(lowerCamelCase_ , format='png' ) UpperCAmelCase_ : int = pa.BufferOutputStream() with ParquetWriter( stream=lowerCamelCase_ , features=Features({'image': Image()} ) , embed_local_files=lowerCamelCase_ ) as writer: writer.write({'image': image_path} ) writer.finalize() UpperCAmelCase_ : Optional[int] = pa.BufferReader(output.getvalue() ) UpperCAmelCase_ : pa.Table = pq.read_table(lowerCamelCase_ ) UpperCAmelCase_ : List[str] = pa_table.to_pydict() if embed_local_files: assert isinstance(out['image'][0]['path'] , lowerCamelCase_ ) with open(lowerCamelCase_ , 'rb' ) as f: assert out["image"][0]["bytes"] == f.read() else: assert out["image"][0]["path"] == image_path assert out["image"][0]["bytes"] is None def _lowerCamelCase ( ): """simple docstring""" UpperCAmelCase_ : List[str] = pa.schema([pa.field('col_1' , pa.string() , nullable=lowerCamelCase_ )] ) UpperCAmelCase_ : Dict = pa.BufferOutputStream() with ArrowWriter(stream=lowerCamelCase_ ) as writer: writer._build_writer(inferred_schema=lowerCamelCase_ ) assert writer._schema == pa.schema([pa.field('col_1' , pa.string() )] )
274
0
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging __lowerCAmelCase = logging.get_logger(__name__) class __magic_name__ ( _UpperCamelCase ): lowerCAmelCase : str = ['input_values', 'padding_mask'] def __init__( self : Optional[int] ,_UpperCAmelCase : int = 1 ,_UpperCAmelCase : int = 24000 ,_UpperCAmelCase : float = 0.0 ,_UpperCAmelCase : float = None ,_UpperCAmelCase : float = None ,**_UpperCAmelCase : str ,): super().__init__(feature_size=_UpperCAmelCase ,sampling_rate=_UpperCAmelCase ,padding_value=_UpperCAmelCase ,**_UpperCAmelCase ) _a : Optional[Any] = chunk_length_s _a : str = overlap @property def __lowercase ( self : Union[str, Any] ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def __lowercase ( self : Optional[int] ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 ,int((1.0 - self.overlap) * self.chunk_length ) ) def __call__( self : Tuple ,_UpperCAmelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,_UpperCAmelCase : Optional[Union[bool, str, PaddingStrategy]] = None ,_UpperCAmelCase : Optional[bool] = False ,_UpperCAmelCase : Optional[int] = None ,_UpperCAmelCase : Optional[Union[str, TensorType]] = None ,_UpperCAmelCase : Optional[int] = None ,): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"""The model corresponding to this feature extractor: {self} was trained using a sampling rate of""" F""" {self.sampling_rate}. Please make sure that the provided audio input was sampled with""" F""" {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) if padding and truncation: raise ValueError('Both padding and truncation were set. Make sure you only set one.' ) elif padding is None: # by default let's pad the inputs _a : Union[str, Any] = True _a : Optional[int] = bool( isinstance(_UpperCAmelCase ,(list, tuple) ) and (isinstance(raw_audio[0] ,(np.ndarray, tuple, list) )) ) if is_batched: _a : Tuple = [np.asarray(_UpperCAmelCase ,dtype=np.floataa ).T for audio in raw_audio] elif not is_batched and not isinstance(_UpperCAmelCase ,np.ndarray ): _a : List[str] = np.asarray(_UpperCAmelCase ,dtype=np.floataa ) elif isinstance(_UpperCAmelCase ,np.ndarray ) and raw_audio.dtype is np.dtype(np.floataa ): _a : Optional[int] = raw_audio.astype(np.floataa ) # always return batch if not is_batched: _a : int = [np.asarray(_UpperCAmelCase ).T] # verify inputs are valid for idx, example in enumerate(_UpperCAmelCase ): if example.ndim > 2: raise ValueError(F"""Expected input shape (channels, length) but got shape {example.shape}""" ) if self.feature_size == 1 and example.ndim != 1: raise ValueError(F"""Expected mono audio but example has {example.shape[-1]} channels""" ) if self.feature_size == 2 and example.shape[-1] != 2: raise ValueError(F"""Expected stereo audio but example has {example.shape[-1]} channels""" ) _a : Tuple = None _a : str = BatchFeature({'input_values': raw_audio} ) if self.chunk_stride is not None and self.chunk_length is not None and max_length is None: if truncation: _a : Optional[Any] = min(array.shape[0] for array in raw_audio ) _a : Optional[int] = int(np.floor(max_length / self.chunk_stride ) ) _a : int = (nb_step - 1) * self.chunk_stride + self.chunk_length elif padding: _a : str = max(array.shape[0] for array in raw_audio ) _a : Optional[int] = int(np.ceil(max_length / self.chunk_stride ) ) _a : Optional[Any] = (nb_step - 1) * self.chunk_stride + self.chunk_length _a : str = 'max_length' else: _a : Dict = input_values # normal padding on batch if padded_inputs is None: _a : Tuple = self.pad( _UpperCAmelCase ,max_length=_UpperCAmelCase ,truncation=_UpperCAmelCase ,padding=_UpperCAmelCase ,return_attention_mask=_UpperCAmelCase ,) if padding: _a : Optional[Any] = padded_inputs.pop('attention_mask' ) _a : Any = [] for example in padded_inputs.pop('input_values' ): if self.feature_size == 1: _a : Optional[int] = example[..., None] input_values.append(example.T ) _a : Optional[Any] = input_values if return_tensors is not None: _a : str = padded_inputs.convert_to_tensors(_UpperCAmelCase ) return padded_inputs
89
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( _UpperCamelCase , unittest.TestCase ): lowerCAmelCase : Optional[int] = BarthezTokenizer lowerCAmelCase : int = BarthezTokenizerFast lowerCAmelCase : Dict = True lowerCAmelCase : str = True def __lowercase ( self : List[Any] ): super().setUp() _a : List[Any] = BarthezTokenizerFast.from_pretrained('moussaKam/mbarthez' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ,legacy_format=_UpperCAmelCase ) _a : Union[str, Any] = tokenizer def __lowercase ( self : Tuple ): _a : Optional[Any] = '<pad>' _a : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCAmelCase ) ,_UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCAmelCase ) ,_UpperCAmelCase ) def __lowercase ( self : str ): _a : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'<s>' ) self.assertEqual(vocab_keys[1] ,'<pad>' ) self.assertEqual(vocab_keys[-1] ,'<mask>' ) self.assertEqual(len(_UpperCAmelCase ) ,101122 ) def __lowercase ( self : Dict ): self.assertEqual(self.get_tokenizer().vocab_size ,101122 ) @require_torch def __lowercase ( self : Dict ): _a : Any = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] _a : Dict = [0, 57, 3018, 70307, 91, 2] _a : Dict = self.tokenizer( _UpperCAmelCase ,max_length=len(_UpperCAmelCase ) ,padding=_UpperCAmelCase ,truncation=_UpperCAmelCase ,return_tensors='pt' ) self.assertIsInstance(_UpperCAmelCase ,_UpperCAmelCase ) self.assertEqual((2, 6) ,batch.input_ids.shape ) self.assertEqual((2, 6) ,batch.attention_mask.shape ) _a : Tuple = batch.input_ids.tolist()[0] self.assertListEqual(_UpperCAmelCase ,_UpperCAmelCase ) def __lowercase ( self : Optional[Any] ): if not self.test_rust_tokenizer: return _a : str = self.get_tokenizer() _a : List[str] = self.get_rust_tokenizer() _a : Dict = 'I was born in 92000, and this is falsé.' _a : List[Any] = tokenizer.tokenize(_UpperCAmelCase ) _a : Tuple = rust_tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase ,_UpperCAmelCase ) _a : Optional[Any] = tokenizer.encode(_UpperCAmelCase ,add_special_tokens=_UpperCAmelCase ) _a : Optional[int] = rust_tokenizer.encode(_UpperCAmelCase ,add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase ,_UpperCAmelCase ) _a : Union[str, Any] = self.get_rust_tokenizer() _a : Any = tokenizer.encode(_UpperCAmelCase ) _a : Optional[int] = rust_tokenizer.encode(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase ,_UpperCAmelCase ) @slow def __lowercase ( self : Optional[int] ): # fmt: off _a : Optional[int] = {'input_ids': [[0, 490, 14328, 4507, 354, 47, 43669, 95, 25, 78117, 20215, 19779, 190, 22, 400, 4, 35343, 80310, 603, 86, 24937, 105, 33438, 94762, 196, 39642, 7, 15, 15933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 10534, 87, 25, 66, 3358, 196, 55289, 8, 82961, 81, 2204, 75203, 7, 15, 763, 12956, 216, 178, 14328, 9595, 1377, 69693, 7, 448, 71021, 196, 18106, 1437, 13974, 108, 9083, 4, 49315, 7, 39, 86, 1326, 2793, 46333, 4, 448, 196, 74588, 7, 49315, 7, 39, 21, 822, 38470, 74, 21, 66723, 62480, 8, 22050, 5, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. _a : Optional[Any] = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=_UpperCAmelCase ,model_name='moussaKam/mbarthez' ,revision='c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6' ,sequences=_UpperCAmelCase ,)
89
1
import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __snake_case : Any =get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowerCamelCase__ ( lowerCamelCase__ , unittest.TestCase): '''simple docstring''' snake_case_ =XLNetTokenizer snake_case_ =XLNetTokenizerFast snake_case_ =True snake_case_ =True def lowerCAmelCase__ (self ) -> Optional[Any]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase__ : int = XLNetTokenizer(__lowerCamelCase ,keep_accents=__lowerCamelCase ) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase__ (self ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : str = '''<s>''' lowerCAmelCase__ : Tuple = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCamelCase ) ,__lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCamelCase ) ,__lowerCamelCase ) def lowerCAmelCase__ (self ) -> Dict: """simple docstring""" lowerCAmelCase__ : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'''<unk>''' ) self.assertEqual(vocab_keys[1] ,'''<s>''' ) self.assertEqual(vocab_keys[-1] ,'''<eod>''' ) self.assertEqual(len(__lowerCamelCase ) ,10_06 ) def lowerCAmelCase__ (self ) -> Optional[Any]: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size ,10_00 ) def lowerCAmelCase__ (self ) -> Any: """simple docstring""" lowerCAmelCase__ : Dict = XLNetTokenizer(__lowerCamelCase ,keep_accents=__lowerCamelCase ) lowerCAmelCase__ : Tuple = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__lowerCamelCase ,['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) ,[2_85, 46, 10, 1_70, 3_82] ) lowerCAmelCase__ : int = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowerCamelCase ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] ,) lowerCAmelCase__ : str = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase ,[8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] ) lowerCAmelCase__ : Tuple = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] ,) def lowerCAmelCase__ (self ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = XLNetTokenizer(__lowerCamelCase ,do_lower_case=__lowerCamelCase ) lowerCAmelCase__ : Dict = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowerCamelCase ,[ SPIECE_UNDERLINE + '''''', '''i''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''se''', '''.''', ] ,) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) ,['''▁he''', '''ll''', '''o'''] ) def lowerCAmelCase__ (self ) -> Dict: """simple docstring""" lowerCAmelCase__ : Optional[Any] = XLNetTokenizer(__lowerCamelCase ,do_lower_case=__lowerCamelCase ) lowerCAmelCase__ : int = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowerCamelCase ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''se''', '''.''', ] ,) @slow def lowerCAmelCase__ (self ) -> Any: """simple docstring""" lowerCAmelCase__ : Optional[Any] = XLNetTokenizer.from_pretrained('''xlnet-base-cased''' ) lowerCAmelCase__ : Any = tokenizer.encode('''sequence builders''' ,add_special_tokens=__lowerCamelCase ) lowerCAmelCase__ : Any = tokenizer.encode('''multi-sequence build''' ,add_special_tokens=__lowerCamelCase ) lowerCAmelCase__ : Optional[int] = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase ) lowerCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase ,__lowerCamelCase ) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def lowerCAmelCase__ (self ) -> Union[str, Any]: """simple docstring""" lowerCAmelCase__ : Dict = {'''input_ids''': [[17, 2_14_42, 2_70, 17, 10, 1_46_45, 3_18, 34, 17, 45_46, 31_45, 7_87, 13, 77_52, 2_20_18, 23, 21, 17, 45_46, 31_45, 7_87, 13, 33_52, 1_44_31, 13, 55_00, 11, 11_76, 5_80, 13, 1_68_19, 47_97, 23, 17, 10, 1_71_35, 6_58, 19, 4_57, 79_32, 13, 1_84, 19, 31_54, 1_71_35, 64_68, 19, 14_04, 1_22_69, 19, 42_29, 53_56, 1_62_64, 46, 19, 17, 2_05_45, 1_03_95, 9, 9, 9, 11, 28, 64_21, 95_31, 2_07_29, 17, 10, 3_53, 1_70_22, 11, 21, 64_21, 95_31, 1_69_49, 17, 10, 1_15_09, 7_53, 11, 33, 95, 24_21, 73_85, 9_56, 1_44_31, 26_26, 25, 8_42, 73_85, 48_36, 21, 14_29, 22_72, 98_55, 31_20, 1_61, 2_47_38, 19, 1_32_03, 6_58, 2_18, 7_87, 21, 4_30, 1_84_82, 8_47, 26_37, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_22, 2_21_78, 27, 10_64, 22, 9_56, 13, 1_11_01, 14_29, 58_54, 2_43_13, 1_89_53, 40, 4_22, 2_43_66, 68, 17_58, 37, 1_04_83, 1_42_57, 31, 2_07, 2_63, 21, 2_03, 37_73, 25, 71, 97_35, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 32, 20_49, 34_42, 17, 1_38_94, 33_80, 23, 95, 18, 1_76_34, 22_88, 9, 4, 3]], '''token_type_ids''': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCamelCase ,model_name='''xlnet-base-cased''' ,revision='''c841166438c31ec7ca9a106dee7bb312b73ae511''' ,)
94
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __snake_case : Optional[int] ={ 'configuration_vision_text_dual_encoder': ['VisionTextDualEncoderConfig'], 'processing_vision_text_dual_encoder': ['VisionTextDualEncoderProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : int =['VisionTextDualEncoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : List[str] =['FlaxVisionTextDualEncoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Tuple =['TFVisionTextDualEncoderModel'] if TYPE_CHECKING: from .configuration_vision_text_dual_encoder import VisionTextDualEncoderConfig from .processing_vision_text_dual_encoder import VisionTextDualEncoderProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_text_dual_encoder import VisionTextDualEncoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_text_dual_encoder import FlaxVisionTextDualEncoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_text_dual_encoder import TFVisionTextDualEncoderModel else: import sys __snake_case : str =_LazyModule(__name__, globals()['__file__'], _import_structure)
94
1
"""simple docstring""" import re def _snake_case ( snake_case__ : str ): A = re.compile( r'^(?:0|94|\+94|0{2}94)' r'7(0|1|2|4|5|6|7|8)' r'(-| |)' r'\d{7}$' ) return bool(re.search(snake_case__ , snake_case__ ) ) if __name__ == "__main__": _lowercase = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
74
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''bigcode/gpt_bigcode-santacoder''': '''https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json''', } class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' _lowerCamelCase: List[str] = '''gpt_bigcode''' _lowerCamelCase: List[Any] = ['''past_key_values'''] _lowerCamelCase: int = { '''hidden_size''': '''n_embd''', '''max_position_embeddings''': '''n_positions''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : Optional[int] ,A_ : Dict=5_0257 ,A_ : Union[str, Any]=1024 ,A_ : str=768 ,A_ : Any=12 ,A_ : Any=12 ,A_ : Optional[int]=None ,A_ : Any="gelu_pytorch_tanh" ,A_ : List[str]=0.1 ,A_ : Optional[int]=0.1 ,A_ : List[str]=0.1 ,A_ : Tuple=1e-5 ,A_ : Optional[int]=0.02 ,A_ : List[str]=True ,A_ : Optional[Any]=True ,A_ : List[Any]=5_0256 ,A_ : Union[str, Any]=5_0256 ,A_ : int=True ,A_ : Optional[Any]=True ,A_ : Dict=True ,**A_ : Union[str, Any] ,) -> Union[str, Any]: A = vocab_size A = n_positions A = n_embd A = n_layer A = n_head A = n_inner A = activation_function A = resid_pdrop A = embd_pdrop A = attn_pdrop A = layer_norm_epsilon A = initializer_range A = scale_attn_weights A = use_cache A = attention_softmax_in_fpaa A = scale_attention_softmax_in_fpaa A = multi_query A = bos_token_id A = eos_token_id super().__init__(bos_token_id=A_ ,eos_token_id=A_ ,**A_ )
74
1
"""simple docstring""" import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class a_ ( snake_case_ ): '''simple docstring''' lowerCamelCase__ : Dict = '' lowerCamelCase__ : str = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) lowerCamelCase__ : str = None # compression type in fsspec. ex: "gzip" lowerCamelCase__ : str = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__(self, lowerCamelCase_ = "", lowerCamelCase_ = None, lowerCamelCase_ = None, **lowerCamelCase_ ): '''simple docstring''' super().__init__(self, **lowerCamelCase_ ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode lowerCamelCase__ : Optional[Any] = fsspec.open( lowerCamelCase_, mode='rb', protocol=lowerCamelCase_, compression=self.compression, client_kwargs={ 'requote_redirect_url': False, # see https://github.com/huggingface/datasets/pull/5459 'trust_env': True, # Enable reading proxy env variables. **(target_options or {}).pop('client_kwargs', {} ), # To avoid issues if it was already passed. }, **(target_options or {}), ) lowerCamelCase__ : int = os.path.basename(self.file.path.split('::' )[0] ) lowerCamelCase__ : str = ( self.compressed_name[: self.compressed_name.rindex('.' )] if '.' in self.compressed_name else self.compressed_name ) lowerCamelCase__ : List[Any] = None @classmethod def a__ (cls, lowerCamelCase_ ): '''simple docstring''' return super()._strip_protocol(lowerCamelCase_ ).lstrip('/' ) def a__ (self ): '''simple docstring''' if self.dir_cache is None: lowerCamelCase__ : Any = {**self.file.fs.info(self.file.path ), 'name': self.uncompressed_name} lowerCamelCase__ : int = {f['name']: f} def a__ (self, lowerCamelCase_ ): '''simple docstring''' return self.file.open().read() def a__ (self, lowerCamelCase_, lowerCamelCase_ = "rb", lowerCamelCase_=None, lowerCamelCase_=True, lowerCamelCase_=None, **lowerCamelCase_, ): '''simple docstring''' lowerCamelCase__ : Dict = self._strip_protocol(lowerCamelCase_ ) if mode != "rb": raise ValueError(f'''Tried to read with mode {mode} on file {self.file.path} opened with mode \'rb\'''' ) return self.file.open() class a_ ( snake_case_ ): '''simple docstring''' lowerCamelCase__ : Optional[int] = 'bz2' lowerCamelCase__ : List[str] = 'bz2' lowerCamelCase__ : int = '.bz2' class a_ ( snake_case_ ): '''simple docstring''' lowerCamelCase__ : Optional[Any] = 'gzip' lowerCamelCase__ : List[str] = 'gzip' lowerCamelCase__ : Union[str, Any] = '.gz' class a_ ( snake_case_ ): '''simple docstring''' lowerCamelCase__ : Tuple = 'lz4' lowerCamelCase__ : int = 'lz4' lowerCamelCase__ : Dict = '.lz4' class a_ ( snake_case_ ): '''simple docstring''' lowerCamelCase__ : Tuple = 'xz' lowerCamelCase__ : Optional[int] = 'xz' lowerCamelCase__ : Dict = '.xz' class a_ ( snake_case_ ): '''simple docstring''' lowerCamelCase__ : List[str] = 'zstd' lowerCamelCase__ : Dict = 'zstd' lowerCamelCase__ : List[Any] = '.zst' def __init__(self, lowerCamelCase_, lowerCamelCase_ = "rb", lowerCamelCase_ = None, lowerCamelCase_ = None, lowerCamelCase_ = DEFAULT_BLOCK_SIZE, **lowerCamelCase_, ): '''simple docstring''' super().__init__( fo=lowerCamelCase_, mode=lowerCamelCase_, target_protocol=lowerCamelCase_, target_options=lowerCamelCase_, block_size=lowerCamelCase_, **lowerCamelCase_, ) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 lowerCamelCase__ : List[str] = self.file.__enter__ class a_ : '''simple docstring''' def __init__(self, lowerCamelCase_ ): '''simple docstring''' lowerCamelCase__ : Optional[int] = file_ def __enter__(self ): '''simple docstring''' self._file.__enter__() return self def __exit__(self, *lowerCamelCase_, **lowerCamelCase_ ): '''simple docstring''' self._file.__exit__(*lowerCamelCase_, **lowerCamelCase_ ) def __iter__(self ): '''simple docstring''' return iter(self._file ) def a__ (self ): '''simple docstring''' return next(self._file ) def __getattr__(self, lowerCamelCase_ ): '''simple docstring''' return getattr(self._file, lowerCamelCase_ ) def fixed_enter(*lowerCamelCase_, **lowerCamelCase_ ): return WrappedFile(_enter(*lowerCamelCase_, **lowerCamelCase_ ) ) lowerCamelCase__ : Tuple = fixed_enter
316
"""simple docstring""" # Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best hparam into the config # - generate model_cards - useful if you have multiple models from the same paper import argparse import json import os import re from collections import OrderedDict from os.path import basename, dirname import fairseq import torch from fairseq import hub_utils from fairseq.data.dictionary import Dictionary from transformers import FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() A_ : Optional[Any] = 2 # based on the results of a search on a range of `num_beams`, `length_penalty` and `early_stopping` # values against wmt19 test data to obtain the best BLEU scores, we will use the following defaults: # # * `num_beams`: 5 (higher scores better, but requires more memory/is slower, can be adjusted by users) # * `early_stopping`: `False` consistently scored better # * `length_penalty` varied, so will assign the best one depending on the model A_ : List[Any] = { # fairseq: "wmt19-ru-en": {"length_penalty": 1.1}, "wmt19-en-ru": {"length_penalty": 1.15}, "wmt19-en-de": {"length_penalty": 1.0}, "wmt19-de-en": {"length_penalty": 1.1}, # allenai: "wmt16-en-de-dist-12-1": {"length_penalty": 0.6}, "wmt16-en-de-dist-6-1": {"length_penalty": 0.6}, "wmt16-en-de-12-1": {"length_penalty": 0.8}, "wmt19-de-en-6-6-base": {"length_penalty": 0.6}, "wmt19-de-en-6-6-big": {"length_penalty": 0.6}, } # this remaps the different models to their organization names A_ : str = {} for m in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: A_ : str = "facebook" for m in [ "wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1", "wmt19-de-en-6-6-base", "wmt19-de-en-6-6-big", ]: A_ : Optional[Any] = "allenai" def lowerCamelCase_ ( _lowerCamelCase ): # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} lowerCamelCase__ : List[Any] = dict((re.sub(r'@@$' , '' , _lowerCamelCase ), v) if k.endswith('@@' ) else (re.sub(r'$' , '</w>' , _lowerCamelCase ), v) for k, v in d.items() ) lowerCamelCase__ : int = '<s> <pad> </s> <unk>'.split() # restore the special tokens for k in keep_keys: del da[f'''{k}</w>'''] lowerCamelCase__ : List[str] = d[k] # restore return da def lowerCamelCase_ ( _lowerCamelCase , _lowerCamelCase ): # prep assert os.path.exists(_lowerCamelCase ) os.makedirs(_lowerCamelCase , exist_ok=_lowerCamelCase ) print(f'''Writing results to {pytorch_dump_folder_path}''' ) # handle various types of models lowerCamelCase__ : Optional[int] = basename(_lowerCamelCase ) lowerCamelCase__ : str = dirname(_lowerCamelCase ) lowerCamelCase__ : Any = fairseq.model_parallel.models.transformer.ModelParallelTransformerModel lowerCamelCase__ : int = cls.hub_models() lowerCamelCase__ : str = {'bpe': 'fastbpe', 'tokenizer': 'moses'} lowerCamelCase__ : Optional[Any] = '.' # note: since the model dump is old, fairseq has upgraded its model some # time later, and it does a whole lot of rewrites and splits on the saved # weights, therefore we can't use torch.load() directly on the model file. # see: upgrade_state_dict(state_dict) in fairseq_model.py print(f'''using checkpoint {checkpoint_file}''' ) lowerCamelCase__ : Any = hub_utils.from_pretrained( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , archive_map=_lowerCamelCase , **_lowerCamelCase ) lowerCamelCase__ : List[str] = vars(chkpt['args']['model'] ) lowerCamelCase__ : Optional[Any] = args['source_lang'] lowerCamelCase__ : List[str] = args['target_lang'] lowerCamelCase__ : List[str] = dirname(_lowerCamelCase ) lowerCamelCase__ : Union[str, Any] = basename(_lowerCamelCase ) # dicts lowerCamelCase__ : Optional[Any] = os.path.join(_lowerCamelCase , f'''dict.{src_lang}.txt''' ) lowerCamelCase__ : Optional[Any] = os.path.join(_lowerCamelCase , f'''dict.{tgt_lang}.txt''' ) lowerCamelCase__ : Dict = Dictionary.load(_lowerCamelCase ) lowerCamelCase__ : List[Any] = rewrite_dict_keys(src_dict.indices ) lowerCamelCase__ : int = len(_lowerCamelCase ) lowerCamelCase__ : List[Any] = os.path.join(_lowerCamelCase , 'vocab-src.json' ) print(f'''Generating {src_vocab_file} of {src_vocab_size} of {src_lang} records''' ) with open(_lowerCamelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(_lowerCamelCase , ensure_ascii=_lowerCamelCase , indent=_lowerCamelCase ) ) # detect whether this is a do_lower_case situation, which can be derived by checking whether we # have at least one uppercase letter in the source vocab lowerCamelCase__ : Optional[int] = True for k in src_vocab.keys(): if not k.islower(): lowerCamelCase__ : int = False break lowerCamelCase__ : str = Dictionary.load(_lowerCamelCase ) lowerCamelCase__ : Union[str, Any] = rewrite_dict_keys(tgt_dict.indices ) lowerCamelCase__ : Optional[Any] = len(_lowerCamelCase ) lowerCamelCase__ : Dict = os.path.join(_lowerCamelCase , 'vocab-tgt.json' ) print(f'''Generating {tgt_vocab_file} of {tgt_vocab_size} of {tgt_lang} records''' ) with open(_lowerCamelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(_lowerCamelCase , ensure_ascii=_lowerCamelCase , indent=_lowerCamelCase ) ) # merges_file (bpecodes) lowerCamelCase__ : List[Any] = os.path.join(_lowerCamelCase , VOCAB_FILES_NAMES['merges_file'] ) for fn in ["bpecodes", "code"]: # older fairseq called the merges file "code" lowerCamelCase__ : Optional[int] = os.path.join(_lowerCamelCase , _lowerCamelCase ) if os.path.exists(_lowerCamelCase ): break with open(_lowerCamelCase , encoding='utf-8' ) as fin: lowerCamelCase__ : Union[str, Any] = fin.read() lowerCamelCase__ : Any = re.sub(r' \d+$' , '' , _lowerCamelCase , 0 , re.M ) # remove frequency number print(f'''Generating {merges_file}''' ) with open(_lowerCamelCase , 'w' , encoding='utf-8' ) as fout: fout.write(_lowerCamelCase ) # model config lowerCamelCase__ : Dict = os.path.join(_lowerCamelCase , 'config.json' ) # validate bpe/tokenizer config, as currently it's hardcoded to moses+fastbpe - # may have to modify the tokenizer if a different type is used by a future model assert args["bpe"] == "fastbpe", f'''need to extend tokenizer to support bpe={args['bpe']}''' assert args["tokenizer"] == "moses", f'''need to extend tokenizer to support bpe={args['tokenizer']}''' lowerCamelCase__ : Optional[int] = { 'architectures': ['FSMTForConditionalGeneration'], 'model_type': 'fsmt', 'activation_dropout': args['activation_dropout'], 'activation_function': 'relu', 'attention_dropout': args['attention_dropout'], 'd_model': args['decoder_embed_dim'], 'dropout': args['dropout'], 'init_std': 0.02, 'max_position_embeddings': args['max_source_positions'], 'num_hidden_layers': args['encoder_layers'], 'src_vocab_size': src_vocab_size, 'tgt_vocab_size': tgt_vocab_size, 'langs': [src_lang, tgt_lang], 'encoder_attention_heads': args['encoder_attention_heads'], 'encoder_ffn_dim': args['encoder_ffn_embed_dim'], 'encoder_layerdrop': args['encoder_layerdrop'], 'encoder_layers': args['encoder_layers'], 'decoder_attention_heads': args['decoder_attention_heads'], 'decoder_ffn_dim': args['decoder_ffn_embed_dim'], 'decoder_layerdrop': args['decoder_layerdrop'], 'decoder_layers': args['decoder_layers'], 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'is_encoder_decoder': True, 'scale_embedding': not args['no_scale_embedding'], 'tie_word_embeddings': args['share_all_embeddings'], } # good hparam defaults to start with lowerCamelCase__ : str = 5 lowerCamelCase__ : Tuple = False if model_dir in best_score_hparams and "length_penalty" in best_score_hparams[model_dir]: lowerCamelCase__ : List[str] = best_score_hparams[model_dir]['length_penalty'] else: lowerCamelCase__ : List[Any] = 1.0 print(f'''Generating {fsmt_model_config_file}''' ) with open(_lowerCamelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(_lowerCamelCase , ensure_ascii=_lowerCamelCase , indent=_lowerCamelCase ) ) # tokenizer config lowerCamelCase__ : Dict = os.path.join(_lowerCamelCase , _lowerCamelCase ) lowerCamelCase__ : int = { 'langs': [src_lang, tgt_lang], 'model_max_length': 1024, 'do_lower_case': do_lower_case, } print(f'''Generating {fsmt_tokenizer_config_file}''' ) with open(_lowerCamelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(_lowerCamelCase , ensure_ascii=_lowerCamelCase , indent=_lowerCamelCase ) ) # model lowerCamelCase__ : List[str] = chkpt['models'][0] lowerCamelCase__ : Optional[Any] = model.state_dict() # rename keys to start with 'model.' lowerCamelCase__ : str = OrderedDict(('model.' + k, v) for k, v in model_state_dict.items() ) # remove unneeded keys lowerCamelCase__ : int = [ 'model.model', 'model.encoder.version', 'model.decoder.version', 'model.encoder_embed_tokens.weight', 'model.decoder_embed_tokens.weight', 'model.encoder.embed_positions._float_tensor', 'model.decoder.embed_positions._float_tensor', ] for k in ignore_keys: model_state_dict.pop(_lowerCamelCase , _lowerCamelCase ) lowerCamelCase__ : Any = FSMTConfig.from_pretrained(_lowerCamelCase ) lowerCamelCase__ : Union[str, Any] = FSMTForConditionalGeneration(_lowerCamelCase ) # check that it loads ok model_new.load_state_dict(_lowerCamelCase , strict=_lowerCamelCase ) # save lowerCamelCase__ : List[Any] = os.path.join(_lowerCamelCase , _lowerCamelCase ) print(f'''Generating {pytorch_weights_dump_path}''' ) torch.save(_lowerCamelCase , _lowerCamelCase ) print('Conversion is done!' ) print('\nLast step is to upload the files to s3' ) print(f'''cd {data_root}''' ) print(f'''transformers-cli upload {model_dir}''' ) if __name__ == "__main__": A_ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--fsmt_checkpoint_path", default=None, type=str, required=True, help=( "Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts," " bpecodes, etc." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) A_ : Dict = parser.parse_args() convert_fsmt_checkpoint_to_pytorch(args.fsmt_checkpoint_path, args.pytorch_dump_folder_path)
316
1
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( A_ ): # if the collection is empty, returns empty if collection == []: return [] # get some information about the collection lowerCAmelCase__ : Optional[Any] = len(A_ ) lowerCAmelCase__ : Tuple = max(A_ ) lowerCAmelCase__ : Optional[int] = min(A_ ) # create the counting array lowerCAmelCase__ : Optional[int] = coll_max + 1 - coll_min lowerCAmelCase__ : List[Any] = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , A_ ): lowerCAmelCase__ : Dict = counting_arr[i] + counting_arr[i - 1] # create the output collection lowerCAmelCase__ : Dict = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , A_ ) ): lowerCAmelCase__ : List[str] = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def __SCREAMING_SNAKE_CASE ( A_ ): return "".join([chr(A_ ) for i in counting_sort([ord(A_ ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string('''thisisthestring''') == "eghhiiinrsssttt" __UpperCamelCase : Optional[int] = input('''Enter numbers separated by a comma:\n''').strip() __UpperCamelCase : Dict = [int(item) for item in user_input.split(''',''')] print(counting_sort(unsorted))
106
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def __SCREAMING_SNAKE_CASE ( A_=None ): if subparsers is not None: lowerCAmelCase__ : Optional[Any] = subparsers.add_parser('''test''' ) else: lowerCAmelCase__ : List[str] = argparse.ArgumentParser('''Accelerate test command''' ) parser.add_argument( '''--config_file''' , default=A_ , help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) , ) if subparsers is not None: parser.set_defaults(func=A_ ) return parser def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : Optional[int] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['''test_utils''', '''scripts''', '''test_script.py'''] ) if args.config_file is None: lowerCAmelCase__ : Optional[Any] = script_name else: lowerCAmelCase__ : Any = f'--config_file={args.config_file} {script_name}' lowerCAmelCase__ : List[Any] = ['''accelerate-launch'''] + test_args.split() lowerCAmelCase__ : int = execute_subprocess_async(A_ , env=os.environ.copy() ) if result.returncode == 0: print('''Test is a success! You are ready for your distributed training!''' ) def __SCREAMING_SNAKE_CASE ( ): lowerCAmelCase__ : Any = test_command_parser() lowerCAmelCase__ : List[Any] = parser.parse_args() test_command(A_ ) if __name__ == "__main__": main()
106
1
'''simple docstring''' import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class a__ ( lowerCamelCase_ ): def __init__( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ): """simple docstring""" super().__init__() _lowercase : Dict = value_function _lowercase : Union[str, Any] = unet _lowercase : Union[str, Any] = scheduler _lowercase : int = env _lowercase : Dict = env.get_dataset() _lowercase : int = {} for key in self.data.keys(): try: _lowercase : Any = self.data[key].mean() except: # noqa: E722 pass _lowercase : Dict = {} for key in self.data.keys(): try: _lowercase : Union[str, Any] = self.data[key].std() except: # noqa: E722 pass _lowercase : int = env.observation_space.shape[0] _lowercase : Optional[int] = env.action_space.shape[0] def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" return (x_in - self.means[key]) / self.stds[key] def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" return x_in * self.stds[key] + self.means[key] def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" if type(_UpperCamelCase ) is dict: return {k: self.to_torch(_UpperCamelCase ) for k, v in x_in.items()} elif torch.is_tensor(_UpperCamelCase ): return x_in.to(self.unet.device ) return torch.tensor(_UpperCamelCase , device=self.unet.device ) def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" for key, val in cond.items(): _lowercase : str = val.clone() return x_in def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase : Dict = x.shape[0] _lowercase : List[str] = None for i in tqdm.tqdm(self.scheduler.timesteps ): # create batch of timesteps to pass into model _lowercase : Tuple = torch.full((batch_size,) , _UpperCamelCase , device=self.unet.device , dtype=torch.long ) for _ in range(_UpperCamelCase ): with torch.enable_grad(): x.requires_grad_() # permute to match dimension for pre-trained models _lowercase : Any = self.value_function(x.permute(0 , 2 , 1 ) , _UpperCamelCase ).sample _lowercase : List[Any] = torch.autograd.grad([y.sum()] , [x] )[0] _lowercase : str = self.scheduler._get_variance(_UpperCamelCase ) _lowercase : Tuple = torch.exp(0.5 * posterior_variance ) _lowercase : List[str] = model_std * grad _lowercase : int = 0 _lowercase : Any = x.detach() _lowercase : int = x + scale * grad _lowercase : List[Any] = self.reset_xa(_UpperCamelCase , _UpperCamelCase , self.action_dim ) _lowercase : Union[str, Any] = self.unet(x.permute(0 , 2 , 1 ) , _UpperCamelCase ).sample.permute(0 , 2 , 1 ) # TODO: verify deprecation of this kwarg _lowercase : str = self.scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , predict_epsilon=_UpperCamelCase )["prev_sample"] # apply conditions to the trajectory (set the initial state) _lowercase : List[str] = self.reset_xa(_UpperCamelCase , _UpperCamelCase , self.action_dim ) _lowercase : Union[str, Any] = self.to_torch(_UpperCamelCase ) return x, y def __call__( self , _UpperCamelCase , _UpperCamelCase=64 , _UpperCamelCase=32 , _UpperCamelCase=2 , _UpperCamelCase=0.1 ): """simple docstring""" _lowercase : str = self.normalize(_UpperCamelCase , "observations" ) _lowercase : int = obs[None].repeat(_UpperCamelCase , axis=0 ) _lowercase : Optional[int] = {0: self.to_torch(_UpperCamelCase )} _lowercase : Dict = (batch_size, planning_horizon, self.state_dim + self.action_dim) # generate initial noise and apply our conditions (to make the trajectories start at current state) _lowercase : Union[str, Any] = randn_tensor(_UpperCamelCase , device=self.unet.device ) _lowercase : Union[str, Any] = self.reset_xa(_UpperCamelCase , _UpperCamelCase , self.action_dim ) _lowercase : Optional[int] = self.to_torch(_UpperCamelCase ) # run the diffusion process _lowercase , _lowercase : Optional[Any] = self.run_diffusion(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # sort output trajectories by value _lowercase : List[Any] = y.argsort(0 , descending=_UpperCamelCase ).squeeze() _lowercase : Optional[Any] = x[sorted_idx] _lowercase : Dict = sorted_values[:, :, : self.action_dim] _lowercase : List[str] = actions.detach().cpu().numpy() _lowercase : int = self.de_normalize(_UpperCamelCase , key="actions" ) # select the action with the highest value if y is not None: _lowercase : Optional[int] = 0 else: # if we didn't run value guiding, select a random action _lowercase : Optional[Any] = np.random.randint(0 , _UpperCamelCase ) _lowercase : Optional[int] = denorm_actions[selected_index, 0] return denorm_actions
199
'''simple docstring''' import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'spiece.model'} _snake_case = { 'vocab_file': { 'google/bigbird-roberta-base': 'https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model', 'google/bigbird-roberta-large': ( 'https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model' ), 'google/bigbird-base-trivia-itc': ( 'https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model' ), } } _snake_case = { 'google/bigbird-roberta-base': 4_096, 'google/bigbird-roberta-large': 4_096, 'google/bigbird-base-trivia-itc': 4_096, } class a__ ( lowerCamelCase_ ): _SCREAMING_SNAKE_CASE : List[Any] = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE : Any = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE : Optional[int] = ['input_ids', 'attention_mask'] _SCREAMING_SNAKE_CASE : List[int] = [] def __init__( self , _UpperCamelCase , _UpperCamelCase="<unk>" , _UpperCamelCase="<s>" , _UpperCamelCase="</s>" , _UpperCamelCase="<pad>" , _UpperCamelCase="[SEP]" , _UpperCamelCase="[MASK]" , _UpperCamelCase="[CLS]" , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" _lowercase : Optional[int] = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else bos_token _lowercase : Optional[int] = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else eos_token _lowercase : Optional[Any] = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else unk_token _lowercase : Union[str, Any] = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else pad_token _lowercase : Any = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else cls_token _lowercase : str = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else sep_token # Mask token behave like a normal word, i.e. include the space before it _lowercase : Any = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else mask_token _lowercase : Dict = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_UpperCamelCase , eos_token=_UpperCamelCase , unk_token=_UpperCamelCase , pad_token=_UpperCamelCase , sep_token=_UpperCamelCase , mask_token=_UpperCamelCase , cls_token=_UpperCamelCase , sp_model_kwargs=self.sp_model_kwargs , **_UpperCamelCase , ) _lowercase : str = vocab_file _lowercase : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_UpperCamelCase ) @property def _lowerCamelCase ( self ): """simple docstring""" return self.sp_model.get_piece_size() def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Dict = {self.convert_ids_to_tokens(_UpperCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): """simple docstring""" _lowercase : str = self.__dict__.copy() _lowercase : Union[str, Any] = None return state def __setstate__( self , _UpperCamelCase ): """simple docstring""" _lowercase : Dict = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): _lowercase : Optional[Any] = {} _lowercase : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" return self.sp_model.encode(_UpperCamelCase , out_type=_UpperCamelCase ) def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" return self.sp_model.piece_to_id(_UpperCamelCase ) def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" _lowercase : List[Any] = self.sp_model.IdToPiece(_UpperCamelCase ) return token def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" _lowercase : Optional[int] = [] _lowercase : int = "" _lowercase : Tuple = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(_UpperCamelCase ) + token _lowercase : Union[str, Any] = True _lowercase : Optional[int] = [] else: current_sub_tokens.append(_UpperCamelCase ) _lowercase : str = False out_string += self.sp_model.decode(_UpperCamelCase ) return out_string.strip() def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase = False , _UpperCamelCase = None , _UpperCamelCase = True , **_UpperCamelCase , ): """simple docstring""" _lowercase : Any = kwargs.pop("use_source_tokenizer" , _UpperCamelCase ) _lowercase : Dict = self.convert_ids_to_tokens(_UpperCamelCase , skip_special_tokens=_UpperCamelCase ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 _lowercase : Dict = [] _lowercase : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(_UpperCamelCase ) ) _lowercase : Optional[Any] = [] sub_texts.append(_UpperCamelCase ) else: current_sub_text.append(_UpperCamelCase ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(_UpperCamelCase ) ) # Mimic the behavior of the Rust tokenizer: # No space before [MASK] and [SEP] if spaces_between_special_tokens: _lowercase : List[str] = re.sub(R" (\[(MASK|SEP)\])" , R"\1" , " ".join(_UpperCamelCase ) ) else: _lowercase : List[Any] = "".join(_UpperCamelCase ) _lowercase : Optional[int] = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: _lowercase : Tuple = self.clean_up_tokenization(_UpperCamelCase ) return clean_text else: return text def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase = None ): """simple docstring""" if not os.path.isdir(_UpperCamelCase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return _lowercase : Optional[Any] = os.path.join( _UpperCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_UpperCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _UpperCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(_UpperCamelCase , "wb" ) as fi: _lowercase : List[str] = self.sp_model.serialized_model_proto() fi.write(_UpperCamelCase ) return (out_vocab_file,) def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _lowercase : str = [self.cls_token_id] _lowercase : List[str] = [self.sep_token_id] return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_UpperCamelCase , token_ids_a=_UpperCamelCase , already_has_special_tokens=_UpperCamelCase ) if token_ids_a is None: return [1] + ([0] * len(_UpperCamelCase )) + [1] return [1] + ([0] * len(_UpperCamelCase )) + [1] + ([0] * len(_UpperCamelCase )) + [1] def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase = None ): """simple docstring""" _lowercase : str = [self.sep_token_id] _lowercase : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1]
199
1
def a__ ( ) -> list[list[int]]: return [list(range(1_000 - i , -1_000 - i , -1 ) ) for i in range(1_000 )] _lowerCamelCase : List[str] = generate_large_matrix() _lowerCamelCase : Union[str, Any] = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def a__ ( UpperCAmelCase : list[list[int]] ) -> None: assert all(row == sorted(UpperCAmelCase , reverse=UpperCAmelCase ) for row in grid ) assert all(list(UpperCAmelCase ) == sorted(UpperCAmelCase , reverse=UpperCAmelCase ) for col in zip(*UpperCAmelCase ) ) def a__ ( UpperCAmelCase : list[int] ) -> int: UpperCAmelCase : Union[str, Any] = 0 UpperCAmelCase : Union[str, Any] = len(UpperCAmelCase ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: UpperCAmelCase : Any = (left + right) // 2 UpperCAmelCase : List[str] = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: UpperCAmelCase : List[Any] = mid + 1 else: UpperCAmelCase : Dict = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(UpperCAmelCase ) def a__ ( UpperCAmelCase : list[list[int]] ) -> int: UpperCAmelCase : Dict = 0 UpperCAmelCase : List[Any] = len(grid[0] ) for i in range(len(UpperCAmelCase ) ): UpperCAmelCase : Any = find_negative_index(grid[i][:bound] ) total += bound return (len(UpperCAmelCase ) * len(grid[0] )) - total def a__ ( UpperCAmelCase : list[list[int]] ) -> int: return len([number for row in grid for number in row if number < 0] ) def a__ ( UpperCAmelCase : list[list[int]] ) -> int: UpperCAmelCase : Tuple = 0 for row in grid: for i, number in enumerate(UpperCAmelCase ): if number < 0: total += len(UpperCAmelCase ) - i break return total def a__ ( ) -> None: from timeit import timeit print('''Running benchmarks''' ) UpperCAmelCase : Union[str, Any] = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): UpperCAmelCase : Any = timeit(f'''{func}(grid=grid)''' , setup=UpperCAmelCase , number=500 ) print(f'''{func}() took {time:0.4f} seconds''' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
336
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __UpperCAmelCase ( lowerCamelCase__ , unittest.TestCase ): UpperCamelCase = LayoutLMTokenizer UpperCamelCase = LayoutLMTokenizerFast UpperCamelCase = True UpperCamelCase = True def __magic_name__ ( self : Any ): super().setUp() UpperCAmelCase : Dict = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] UpperCAmelCase : int = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file, '''w''', encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __magic_name__ ( self : Union[str, Any], **__A : List[str] ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname, **__A ) def __magic_name__ ( self : Optional[int], __A : int ): UpperCAmelCase : Optional[Any] = '''UNwant\u00E9d,running''' UpperCAmelCase : Optional[int] = '''unwanted, running''' return input_text, output_text def __magic_name__ ( self : Any ): UpperCAmelCase : Union[str, Any] = self.tokenizer_class(self.vocab_file ) UpperCAmelCase : Optional[Any] = tokenizer.tokenize('''UNwant\u00E9d,running''' ) self.assertListEqual(__A, ['''un''', '''##want''', '''##ed''', ''',''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ), [7, 4, 5, 1_0, 8, 9] ) def __magic_name__ ( self : Optional[int] ): pass
336
1
'''simple docstring''' import importlib.metadata import operator import re import sys from typing import Optional from packaging import version lowerCAmelCase : Dict = { '<': operator.lt, '<=': operator.le, '==': operator.eq, '!=': operator.ne, '>=': operator.ge, '>': operator.gt, } def A_( A : Optional[Any] , A : Optional[Any] , A : Optional[int] , A : List[Any] , A : List[Any] , A : Optional[Any]): if got_ver is None or want_ver is None: raise ValueError( f'''Unable to compare versions for {requirement}: need={want_ver} found={got_ver}. This is unusual. Consider''' f''' reinstalling {pkg}.''') if not ops[op](version.parse(A) , version.parse(A)): raise ImportError( f'''{requirement} is required for a normal functioning of this module, but found {pkg}=={got_ver}.{hint}''') def A_( A : str , A : Optional[str] = None): UpperCamelCase = f'''\n{hint}''' if hint is not None else '' # non-versioned check if re.match(r'^[\w_\-\d]+$' , A): UpperCamelCase , UpperCamelCase , UpperCamelCase = requirement, None, None else: UpperCamelCase = re.findall(r'^([^!=<>\s]+)([\s!=<>]{1,2}.+)' , A) if not match: raise ValueError( 'requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23, but' f''' got {requirement}''') UpperCamelCase , UpperCamelCase = match[0] UpperCamelCase = want_full.split(',') # there could be multiple requirements UpperCamelCase = {} for w in want_range: UpperCamelCase = re.findall(r'^([\s!=<>]{1,2})(.+)' , A) if not match: raise ValueError( 'requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23,' f''' but got {requirement}''') UpperCamelCase , UpperCamelCase = match[0] UpperCamelCase = want_ver if op not in ops: raise ValueError(f'''{requirement}: need one of {list(ops.keys())}, but got {op}''') # special case if pkg == "python": UpperCamelCase = '.'.join([str(A) for x in sys.version_info[:3]]) for op, want_ver in wanted.items(): _compare_versions(A , A , A , A , A , A) return # check if any version is installed try: UpperCamelCase = importlib.metadata.version(A) except importlib.metadata.PackageNotFoundError: raise importlib.metadata.PackageNotFoundError( f'''The \'{requirement}\' distribution was not found and is required by this application. {hint}''') # check that the right version is installed if version number or a range was provided if want_ver is not None: for op, want_ver in wanted.items(): _compare_versions(A , A , A , A , A , A) def A_( A : int): UpperCamelCase = 'Try: pip install transformers -U or pip install -e \'.[dev]\' if you\'re working with git main' return require_version(A , A)
368
'''simple docstring''' import flax.linen as nn import jax import jax.numpy as jnp class SCREAMING_SNAKE_CASE__ ( nn.Module): lowerCAmelCase_ = 42 lowerCAmelCase_ = jnp.floataa def UpperCAmelCase_ ( self )-> int: '''simple docstring''' UpperCamelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , A_ )-> List[str]: '''simple docstring''' UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = hidden_states.shape UpperCamelCase = jax.image.resize( A_ , shape=(batch, height * 2, width * 2, channels) , method='nearest' , ) UpperCamelCase = self.conv(A_ ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module): lowerCAmelCase_ = 42 lowerCAmelCase_ = jnp.floataa def UpperCAmelCase_ ( self )-> List[Any]: '''simple docstring''' UpperCamelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , A_ )-> List[Any]: '''simple docstring''' UpperCamelCase = self.conv(A_ ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module): lowerCAmelCase_ = 42 lowerCAmelCase_ = None lowerCAmelCase_ = 0.0 lowerCAmelCase_ = None lowerCAmelCase_ = jnp.floataa def UpperCAmelCase_ ( self )-> Optional[int]: '''simple docstring''' UpperCamelCase = self.in_channels if self.out_channels is None else self.out_channels UpperCamelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCamelCase = nn.Conv( A_ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCamelCase = nn.Dense(A_ , dtype=self.dtype ) UpperCamelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCamelCase = nn.Dropout(self.dropout_prob ) UpperCamelCase = nn.Conv( A_ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) UpperCamelCase = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut UpperCamelCase = None if use_nin_shortcut: UpperCamelCase = nn.Conv( A_ , kernel_size=(1, 1) , strides=(1, 1) , padding='VALID' , dtype=self.dtype , ) def __call__( self , A_ , A_ , A_=True )-> str: '''simple docstring''' UpperCamelCase = hidden_states UpperCamelCase = self.norma(A_ ) UpperCamelCase = nn.swish(A_ ) UpperCamelCase = self.conva(A_ ) UpperCamelCase = self.time_emb_proj(nn.swish(A_ ) ) UpperCamelCase = jnp.expand_dims(jnp.expand_dims(A_ , 1 ) , 1 ) UpperCamelCase = hidden_states + temb UpperCamelCase = self.norma(A_ ) UpperCamelCase = nn.swish(A_ ) UpperCamelCase = self.dropout(A_ , A_ ) UpperCamelCase = self.conva(A_ ) if self.conv_shortcut is not None: UpperCamelCase = self.conv_shortcut(A_ ) return hidden_states + residual
251
0
from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class __lowerCAmelCase : __lowerCamelCase = PegasusConfig __lowerCamelCase = {} __lowerCamelCase = '''gelu''' def __init__( self , _snake_case , _snake_case=13 , _snake_case=7 , _snake_case=True , _snake_case=False , _snake_case=99 , _snake_case=32 , _snake_case=2 , _snake_case=4 , _snake_case=37 , _snake_case=0.1 , _snake_case=0.1 , _snake_case=40 , _snake_case=2 , _snake_case=1 , _snake_case=0 , ): """simple docstring""" _lowerCAmelCase = parent _lowerCAmelCase = batch_size _lowerCAmelCase = seq_length _lowerCAmelCase = is_training _lowerCAmelCase = use_labels _lowerCAmelCase = vocab_size _lowerCAmelCase = hidden_size _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = intermediate_size _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = attention_probs_dropout_prob _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = eos_token_id _lowerCAmelCase = pad_token_id _lowerCAmelCase = bos_token_id def snake_case ( self ): """simple docstring""" _lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) _lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _lowerCAmelCase = prepare_pegasus_inputs_dict(_snake_case , _snake_case , _snake_case ) return config, inputs_dict def snake_case ( self , _snake_case , _snake_case ): """simple docstring""" _lowerCAmelCase = TFPegasusModel(config=_snake_case ).get_decoder() _lowerCAmelCase = inputs_dict["""input_ids"""] _lowerCAmelCase = input_ids[:1, :] _lowerCAmelCase = inputs_dict["""attention_mask"""][:1, :] _lowerCAmelCase = inputs_dict["""head_mask"""] _lowerCAmelCase = 1 # first forward pass _lowerCAmelCase = model(_snake_case , attention_mask=_snake_case , head_mask=_snake_case , use_cache=_snake_case ) _lowerCAmelCase , _lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) _lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and _lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) _lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) _lowerCAmelCase = model(_snake_case , attention_mask=_snake_case )[0] _lowerCAmelCase = model(_snake_case , attention_mask=_snake_case , past_key_values=_snake_case )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice _lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) _lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] _lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_snake_case , _snake_case , rtol=1e-3 ) def _UpperCAmelCase ( snake_case , snake_case , snake_case , snake_case=None , snake_case=None , snake_case=None , snake_case=None , snake_case=None , ): """simple docstring""" if attention_mask is None: _lowerCAmelCase = tf.cast(tf.math.not_equal(snake_case , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _lowerCAmelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _lowerCAmelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _lowerCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _lowerCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __lowerCAmelCase ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): __lowerCamelCase = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () __lowerCamelCase = (TFPegasusForConditionalGeneration,) if is_tf_available() else () __lowerCamelCase = ( { '''conversational''': TFPegasusForConditionalGeneration, '''feature-extraction''': TFPegasusModel, '''summarization''': TFPegasusForConditionalGeneration, '''text2text-generation''': TFPegasusForConditionalGeneration, '''translation''': TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) __lowerCamelCase = True __lowerCamelCase = False __lowerCamelCase = False def snake_case ( self ): """simple docstring""" _lowerCAmelCase = TFPegasusModelTester(self ) _lowerCAmelCase = ConfigTester(self , config_class=_snake_case ) def snake_case ( self ): """simple docstring""" self.config_tester.run_common_tests() def snake_case ( self ): """simple docstring""" _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_snake_case ) @require_sentencepiece @require_tokenizers @require_tf class __lowerCAmelCase ( unittest.TestCase ): __lowerCamelCase = [ ''' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.''', ''' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning \'Oh I think you\'re nominated\'", said Dappy."And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around."At the end of the day we\'re grateful to be where we are in our careers."If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!" ''', ] __lowerCamelCase = [ '''California\'s largest electricity provider has cut power to hundreds of thousands of customers in an effort to''' ''' reduce the risk of wildfires.''', '''N-Dubz have revealed they\'re "grateful" to have been nominated for four Mobo Awards.''', ] # differs slightly from pytorch, likely due to numerical differences in linear layers __lowerCamelCase = '''google/pegasus-xsum''' @cached_property def snake_case ( self ): """simple docstring""" return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def snake_case ( self ): """simple docstring""" _lowerCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def snake_case ( self , **_snake_case ): """simple docstring""" _lowerCAmelCase = self.translate_src_text(**_snake_case ) assert self.expected_text == generated_words def snake_case ( self , **_snake_case ): """simple docstring""" _lowerCAmelCase = self.tokenizer(self.src_text , **_snake_case , padding=_snake_case , return_tensors="""tf""" ) _lowerCAmelCase = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=_snake_case , ) _lowerCAmelCase = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=_snake_case ) return generated_words @slow def snake_case ( self ): """simple docstring""" self._assert_generated_batch_equal_expected()
82
'''simple docstring''' import os import sys import unittest _lowerCamelCase : Optional[int] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import get_test_info # noqa: E402 from get_test_info import ( # noqa: E402 get_model_to_test_mapping, get_model_to_tester_mapping, get_test_to_tester_mapping, ) _lowerCamelCase : str = os.path.join('tests', 'models', 'bert', 'test_modeling_bert.py') _lowerCamelCase : Tuple = os.path.join('tests', 'models', 'blip', 'test_modeling_blip.py') class __UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def A (self : Any ): A = get_test_to_tester_mapping(_lowerCAmelCase ) A = get_test_to_tester_mapping(_lowerCAmelCase ) A = {"""BertModelTest""": """BertModelTester"""} A = { """BlipModelTest""": """BlipModelTester""", """BlipTextImageModelTest""": """BlipTextImageModelsModelTester""", """BlipTextModelTest""": """BlipTextModelTester""", """BlipTextRetrievalModelTest""": """BlipTextRetrievalModelTester""", """BlipVQAModelTest""": """BlipVQAModelTester""", """BlipVisionModelTest""": """BlipVisionModelTester""", } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) def A (self : str ): A = get_model_to_test_mapping(_lowerCAmelCase ) A = get_model_to_test_mapping(_lowerCAmelCase ) A = { """BertForMaskedLM""": ["""BertModelTest"""], """BertForMultipleChoice""": ["""BertModelTest"""], """BertForNextSentencePrediction""": ["""BertModelTest"""], """BertForPreTraining""": ["""BertModelTest"""], """BertForQuestionAnswering""": ["""BertModelTest"""], """BertForSequenceClassification""": ["""BertModelTest"""], """BertForTokenClassification""": ["""BertModelTest"""], """BertLMHeadModel""": ["""BertModelTest"""], """BertModel""": ["""BertModelTest"""], } A = { """BlipForConditionalGeneration""": ["""BlipTextImageModelTest"""], """BlipForImageTextRetrieval""": ["""BlipTextRetrievalModelTest"""], """BlipForQuestionAnswering""": ["""BlipVQAModelTest"""], """BlipModel""": ["""BlipModelTest"""], """BlipTextModel""": ["""BlipTextModelTest"""], """BlipVisionModel""": ["""BlipVisionModelTest"""], } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) def A (self : Union[str, Any] ): A = get_model_to_tester_mapping(_lowerCAmelCase ) A = get_model_to_tester_mapping(_lowerCAmelCase ) A = { """BertForMaskedLM""": ["""BertModelTester"""], """BertForMultipleChoice""": ["""BertModelTester"""], """BertForNextSentencePrediction""": ["""BertModelTester"""], """BertForPreTraining""": ["""BertModelTester"""], """BertForQuestionAnswering""": ["""BertModelTester"""], """BertForSequenceClassification""": ["""BertModelTester"""], """BertForTokenClassification""": ["""BertModelTester"""], """BertLMHeadModel""": ["""BertModelTester"""], """BertModel""": ["""BertModelTester"""], } A = { """BlipForConditionalGeneration""": ["""BlipTextImageModelsModelTester"""], """BlipForImageTextRetrieval""": ["""BlipTextRetrievalModelTester"""], """BlipForQuestionAnswering""": ["""BlipVQAModelTester"""], """BlipModel""": ["""BlipModelTester"""], """BlipTextModel""": ["""BlipTextModelTester"""], """BlipVisionModel""": ["""BlipVisionModelTester"""], } self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(get_test_info.to_json(_lowerCAmelCase ) , _lowerCAmelCase )
258
0
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed UpperCamelCase__ = """true""" def _a ( SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[str]=82 , SCREAMING_SNAKE_CASE_ : str=16 ): set_seed(42 ) __lowerCAmelCase = RegressionModel() __lowerCAmelCase = deepcopy(SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = RegressionDataset(length=SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = DataLoader(SCREAMING_SNAKE_CASE_ , batch_size=SCREAMING_SNAKE_CASE_ ) model.to(accelerator.device ) __lowerCAmelCase , __lowerCAmelCase = accelerator.prepare(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return model, ddp_model, dataloader def _a ( SCREAMING_SNAKE_CASE_ : Accelerator , SCREAMING_SNAKE_CASE_ : Optional[int]=False ): __lowerCAmelCase = AutoTokenizer.from_pretrained("hf-internal-testing/mrpc-bert-base-cased" ) __lowerCAmelCase = load_dataset("glue" , "mrpc" , split="validation" ) def tokenize_function(SCREAMING_SNAKE_CASE_ : List[Any] ): __lowerCAmelCase = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ ) return outputs with accelerator.main_process_first(): __lowerCAmelCase = dataset.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , remove_columns=["idx", "sentence1", "sentence2"] , ) __lowerCAmelCase = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(SCREAMING_SNAKE_CASE_ : Union[str, Any] ): if use_longest: return tokenizer.pad(SCREAMING_SNAKE_CASE_ , padding="longest" , return_tensors="pt" ) return tokenizer.pad(SCREAMING_SNAKE_CASE_ , padding="max_length" , max_length=1_28 , return_tensors="pt" ) return DataLoader(SCREAMING_SNAKE_CASE_ , shuffle=SCREAMING_SNAKE_CASE_ , collate_fn=SCREAMING_SNAKE_CASE_ , batch_size=16 ) def _a ( SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : int ): __lowerCAmelCase = Accelerator(dispatch_batches=SCREAMING_SNAKE_CASE_ , split_batches=SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = get_dataloader(SCREAMING_SNAKE_CASE_ , not dispatch_batches ) __lowerCAmelCase = AutoModelForSequenceClassification.from_pretrained( "hf-internal-testing/mrpc-bert-base-cased" , return_dict=SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase , __lowerCAmelCase = accelerator.prepare(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def _a ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : Optional[int] ): __lowerCAmelCase = [] for batch in dataloader: __lowerCAmelCase , __lowerCAmelCase = batch.values() with torch.no_grad(): __lowerCAmelCase = model(SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase , __lowerCAmelCase = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __lowerCAmelCase , __lowerCAmelCase = [], [] for logit, targ in logits_and_targets: logits.append(SCREAMING_SNAKE_CASE_ ) targs.append(SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase , __lowerCAmelCase = torch.cat(SCREAMING_SNAKE_CASE_ ), torch.cat(SCREAMING_SNAKE_CASE_ ) return logits, targs def _a ( SCREAMING_SNAKE_CASE_ : Accelerator , SCREAMING_SNAKE_CASE_ : Union[str, Any]=82 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=False , SCREAMING_SNAKE_CASE_ : Any=False , SCREAMING_SNAKE_CASE_ : Optional[int]=16 ): __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = get_basic_setup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase , __lowerCAmelCase = generate_predictions(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) assert ( len(SCREAMING_SNAKE_CASE_ ) == num_samples ), F"""Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(SCREAMING_SNAKE_CASE_ )}""" def _a ( SCREAMING_SNAKE_CASE_ : bool = False , SCREAMING_SNAKE_CASE_ : bool = False ): __lowerCAmelCase = evaluate.load("glue" , "mrpc" ) __lowerCAmelCase , __lowerCAmelCase = get_mrpc_setup(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # First do baseline __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = setup["no"] model.to(SCREAMING_SNAKE_CASE_ ) model.eval() for batch in dataloader: batch.to(SCREAMING_SNAKE_CASE_ ) with torch.inference_mode(): __lowerCAmelCase = model(**SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=SCREAMING_SNAKE_CASE_ , references=batch["labels"] ) __lowerCAmelCase = metric.compute() # Then do distributed __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = setup["ddp"] model.eval() for batch in dataloader: with torch.inference_mode(): __lowerCAmelCase = model(**SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = outputs.logits.argmax(dim=-1 ) __lowerCAmelCase = batch["labels"] __lowerCAmelCase , __lowerCAmelCase = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=SCREAMING_SNAKE_CASE_ , references=SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), F"""Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n""" def _a ( ): __lowerCAmelCase = Accelerator(split_batches=SCREAMING_SNAKE_CASE_ , dispatch_batches=SCREAMING_SNAKE_CASE_ ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("**Testing gather_for_metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`""" ) test_mrpc(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test torch metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __lowerCAmelCase = Accelerator(split_batches=SCREAMING_SNAKE_CASE_ , dispatch_batches=SCREAMING_SNAKE_CASE_ ) if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99""" ) test_torch_metrics(SCREAMING_SNAKE_CASE_ , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test last batch is not dropped when perfectly divisible**" ) __lowerCAmelCase = Accelerator() test_torch_metrics(SCREAMING_SNAKE_CASE_ , 5_12 ) accelerator.state._reset_state() def _a ( SCREAMING_SNAKE_CASE_ : List[str] ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
360
import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class a__ : def __init__( self , _A , _A=1_4 , _A=7 , _A=True , _A=True , _A=False , _A=True , _A=9_9 , _A=3_2 , _A=4 , _A=4 , _A=4 , _A=3_7 , _A="gelu" , _A=0.1 , _A=0.1 , _A=5_1_2 , _A=0.02 , ): """simple docstring""" __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_input_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = rotary_dim __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = initializer_range __lowerCAmelCase = None __lowerCAmelCase = vocab_size - 1 __lowerCAmelCase = vocab_size - 1 __lowerCAmelCase = vocab_size - 1 def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase = None if self.use_input_mask: __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=_A , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = config_and_inputs __lowerCAmelCase = {"input_ids": input_ids, "attention_mask": attention_mask} return config, inputs_dict def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = 2_0 __lowerCAmelCase = model_class_name(_A ) __lowerCAmelCase = model.init_cache(input_ids.shape[0] , _A ) __lowerCAmelCase = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype="i4" ) __lowerCAmelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __lowerCAmelCase = model( input_ids[:, :-1] , attention_mask=_A , past_key_values=_A , position_ids=_A , ) __lowerCAmelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="i4" ) __lowerCAmelCase = model( input_ids[:, -1:] , attention_mask=_A , past_key_values=outputs_cache.past_key_values , position_ids=_A , ) __lowerCAmelCase = model(_A ) __lowerCAmelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"""Max diff is {diff}""" ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = 2_0 __lowerCAmelCase = model_class_name(_A ) __lowerCAmelCase = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )] , axis=-1 , ) __lowerCAmelCase = model.init_cache(input_ids.shape[0] , _A ) __lowerCAmelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __lowerCAmelCase = model( input_ids[:, :-1] , attention_mask=_A , past_key_values=_A , position_ids=_A , ) __lowerCAmelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="i4" ) __lowerCAmelCase = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=_A , position_ids=_A , ) __lowerCAmelCase = model(_A , attention_mask=_A ) __lowerCAmelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"""Max diff is {diff}""" ) @require_flax class a__ ( snake_case__ , snake_case__ , unittest.TestCase ): _a : str = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () _a : Any = (FlaxGPTJForCausalLM,) if is_flax_available() else () def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = FlaxGPTJModelTester(self ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_class_name in self.all_model_classes: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(_A , _A , _A , _A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_class_name in self.all_model_classes: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( _A , _A , _A , _A ) @tooslow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = GPTaTokenizer.from_pretrained("gpt2" , pad_token="<|endoftext|>" , padding_side="left" ) __lowerCAmelCase = tokenizer(["Hello this is a long string", "Hey"] , return_tensors="np" , padding=_A , truncation=_A ) __lowerCAmelCase = FlaxGPTJForCausalLM.from_pretrained("EleutherAI/gpt-j-6B" ) __lowerCAmelCase = False __lowerCAmelCase = model.config.eos_token_id __lowerCAmelCase = jax.jit(model.generate ) __lowerCAmelCase = jit_generate( inputs["input_ids"] , attention_mask=inputs["attention_mask"] , pad_token_id=tokenizer.pad_token_id ).sequences __lowerCAmelCase = tokenizer.batch_decode(_A , skip_special_tokens=_A ) __lowerCAmelCase = [ "Hello this is a long string of text.\n\nI'm trying to get the text of the", "Hey, I'm a little late to the party. I'm going to", ] self.assertListEqual(_A , _A ) @is_pt_flax_cross_test def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __lowerCAmelCase = self._prepare_for_class(_A , _A ) __lowerCAmelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __lowerCAmelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning __lowerCAmelCase = getattr(_A , _A ) __lowerCAmelCase , __lowerCAmelCase = pt_inputs["input_ids"].shape __lowerCAmelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_A ): __lowerCAmelCase = 0 __lowerCAmelCase = 1 __lowerCAmelCase = 0 __lowerCAmelCase = 1 __lowerCAmelCase = pt_model_class(_A ).eval() __lowerCAmelCase = model_class(_A , dtype=jnp.floataa ) __lowerCAmelCase = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _A ) __lowerCAmelCase = fx_state with torch.no_grad(): __lowerCAmelCase = pt_model(**_A ).to_tuple() __lowerCAmelCase = fx_model(**_A ).to_tuple() self.assertEqual(len(_A ) , len(_A ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(_A , _A ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_A ) __lowerCAmelCase = model_class.from_pretrained(_A , from_pt=_A ) __lowerCAmelCase = fx_model_loaded(**_A ).to_tuple() self.assertEqual( len(_A ) , len(_A ) , "Output lengths differ between Flax and PyTorch" ) for fx_output_loaded, pt_output in zip(_A , _A ): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @is_pt_flax_cross_test def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __lowerCAmelCase = self._prepare_for_class(_A , _A ) __lowerCAmelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __lowerCAmelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning __lowerCAmelCase = getattr(_A , _A ) __lowerCAmelCase = pt_model_class(_A ).eval() __lowerCAmelCase = model_class(_A , dtype=jnp.floataa ) __lowerCAmelCase = load_flax_weights_in_pytorch_model(_A , fx_model.params ) __lowerCAmelCase , __lowerCAmelCase = pt_inputs["input_ids"].shape __lowerCAmelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_A ): __lowerCAmelCase = 0 __lowerCAmelCase = 1 __lowerCAmelCase = 0 __lowerCAmelCase = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): __lowerCAmelCase = pt_model(**_A ).to_tuple() __lowerCAmelCase = fx_model(**_A ).to_tuple() self.assertEqual(len(_A ) , len(_A ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(_A , _A ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_A ) __lowerCAmelCase = pt_model_class.from_pretrained(_A , from_flax=_A ) with torch.no_grad(): __lowerCAmelCase = pt_model_loaded(**_A ).to_tuple() self.assertEqual( len(_A ) , len(_A ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(_A , _A ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @tooslow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_class_name in self.all_model_classes: __lowerCAmelCase = model_class_name.from_pretrained("EleutherAI/gpt-j-6B" ) __lowerCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(_A )
102
0
import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class lowerCAmelCase__ ( unittest.TestCase ): def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : Dict[str, int] = None , SCREAMING_SNAKE_CASE__ : int = 32 , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : Union[int, float] = 1 / 2_55 , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = [0.48145466, 0.4578275, 0.40821073] , SCREAMING_SNAKE_CASE__ : Optional[Union[float, List[float]]] = [0.26862954, 0.26130258, 0.27577711] , SCREAMING_SNAKE_CASE__ : bool = True , SCREAMING_SNAKE_CASE__ : Optional[Any]=7 , SCREAMING_SNAKE_CASE__ : List[Any]=30 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=4_00 , SCREAMING_SNAKE_CASE__ : int=3 , ) -> List[str]: __lowerCamelCase = parent __lowerCamelCase = do_resize __lowerCamelCase = size if size is not None else {'''shortest_edge''': 2_88} __lowerCamelCase = size_divisor __lowerCamelCase = do_rescale __lowerCamelCase = rescale_factor __lowerCamelCase = do_normalize __lowerCamelCase = do_center_crop __lowerCamelCase = image_mean __lowerCamelCase = image_std __lowerCamelCase = do_pad __lowerCamelCase = batch_size __lowerCamelCase = num_channels __lowerCamelCase = min_resolution __lowerCamelCase = max_resolution def __A ( self : List[str] ) -> List[Any]: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def __A ( self : str , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : List[str]=False ) -> int: if not batched: __lowerCamelCase = self.size['''shortest_edge'''] __lowerCamelCase = image_inputs[0] if isinstance(SCREAMING_SNAKE_CASE__ , Image.Image ): __lowerCamelCase , __lowerCamelCase = image.size else: __lowerCamelCase , __lowerCamelCase = image.shape[1], image.shape[2] __lowerCamelCase = size / min(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if h < w: __lowerCamelCase , __lowerCamelCase = size, scale * w else: __lowerCamelCase , __lowerCamelCase = scale * h, size __lowerCamelCase = int((13_33 / 8_00) * size ) if max(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) > max_size: __lowerCamelCase = max_size / max(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = newh * scale __lowerCamelCase = neww * scale __lowerCamelCase , __lowerCamelCase = int(newh + 0.5 ), int(neww + 0.5 ) __lowerCamelCase , __lowerCamelCase = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: __lowerCamelCase = [] for image in image_inputs: __lowerCamelCase , __lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __lowerCamelCase = max(SCREAMING_SNAKE_CASE__ , key=lambda SCREAMING_SNAKE_CASE__ : item[0] )[0] __lowerCamelCase = max(SCREAMING_SNAKE_CASE__ , key=lambda SCREAMING_SNAKE_CASE__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class lowerCAmelCase__ ( __lowercase , unittest.TestCase ): a__ : Union[str, Any] = BridgeTowerImageProcessor if is_vision_available() else None def __A ( self : Dict ) -> Union[str, Any]: __lowerCamelCase = BridgeTowerImageProcessingTester(self ) @property def __A ( self : str ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def __A ( self : Union[str, Any] ) -> int: __lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ , '''image_mean''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ , '''image_std''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ , '''do_normalize''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ , '''do_resize''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ , '''size''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ , '''size_divisor''' ) ) def __A ( self : int ) -> Optional[int]: pass def __A ( self : str ) -> str: # Initialize image processor __lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE__ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE__ , Image.Image ) # Test not batched input __lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values __lowerCamelCase , __lowerCamelCase = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCamelCase = image_processing(SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).pixel_values __lowerCamelCase , __lowerCamelCase = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __A ( self : Dict ) -> List[str]: # Initialize image processor __lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE__ , numpify=SCREAMING_SNAKE_CASE__ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE__ , np.ndarray ) # Test not batched input __lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values __lowerCamelCase , __lowerCamelCase = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCamelCase = image_processing(SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).pixel_values __lowerCamelCase , __lowerCamelCase = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __A ( self : Dict ) -> str: # Initialize image processor __lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE__ , torchify=SCREAMING_SNAKE_CASE__ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE__ , torch.Tensor ) # Test not batched input __lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values __lowerCamelCase , __lowerCamelCase = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCamelCase = image_processing(SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).pixel_values __lowerCamelCase , __lowerCamelCase = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
270
import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets SCREAMING_SNAKE_CASE__ : Union[str, Any] = "\\n@inproceedings{popovic-2015-chrf,\n title = \"chr{F}: character n-gram {F}-score for automatic {MT} evaluation\",\n author = \"Popovi{\'c}, Maja\",\n booktitle = \"Proceedings of the Tenth Workshop on Statistical Machine Translation\",\n month = sep,\n year = \"2015\",\n address = \"Lisbon, Portugal\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/W15-3049\",\n doi = \"10.18653/v1/W15-3049\",\n pages = \"392--395\",\n}\n@inproceedings{popovic-2017-chrf,\n title = \"chr{F}++: words helping character n-grams\",\n author = \"Popovi{\'c}, Maja\",\n booktitle = \"Proceedings of the Second Conference on Machine Translation\",\n month = sep,\n year = \"2017\",\n address = \"Copenhagen, Denmark\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/W17-4770\",\n doi = \"10.18653/v1/W17-4770\",\n pages = \"612--618\",\n}\n@inproceedings{post-2018-call,\n title = \"A Call for Clarity in Reporting {BLEU} Scores\",\n author = \"Post, Matt\",\n booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\",\n month = oct,\n year = \"2018\",\n address = \"Belgium, Brussels\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W18-6319\",\n pages = \"186--191\",\n}\n" SCREAMING_SNAKE_CASE__ : int = "\\nChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches,\nand ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation\nthat is already present in sacrebleu.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information.\n" SCREAMING_SNAKE_CASE__ : Optional[Any] = "\nProduces ChrF(++) scores for hypotheses given reference translations.\n\nArgs:\n predictions (list of str): The predicted sentences.\n references (list of list of str): The references. There should be one reference sub-list for each prediction sentence.\n char_order (int): Character n-gram order. Defaults to `6`.\n word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`.\n beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`.\n lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`.\n whitespace (bool): If `True`, include whitespaces when extracting character n-grams.\n eps_smoothing (bool): If `True`, applies epsilon smoothing similar\n to reference chrF++.py, NLTK and Moses implementations. If `False`,\n it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`.\n\nReturns:\n 'score' (float): The chrF (chrF++) score,\n 'char_order' (int): The character n-gram order,\n 'word_order' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++,\n 'beta' (int): Determine the importance of recall w.r.t precision\n\nExamples:\n Example 1--a simple example of calculating chrF:\n >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"]\n >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]]\n >>> chrf = datasets.load_metric(\"chrf\")\n >>> results = chrf.compute(predictions=prediction, references=reference)\n >>> print(results)\n {'score': 84.64214891738334, 'char_order': 6, 'word_order': 0, 'beta': 2}\n\n Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF:\n >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"]\n >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]]\n >>> chrf = datasets.load_metric(\"chrf\")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2)\n >>> print(results)\n {'score': 82.87263732906315, 'char_order': 6, 'word_order': 2, 'beta': 2}\n\n Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case:\n >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"]\n >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]]\n >>> chrf = datasets.load_metric(\"chrf\")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2,\n ... lowercase=True)\n >>> print(results)\n {'score': 92.12853119829202, 'char_order': 6, 'word_order': 2, 'beta': 2}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def __A ( self : Dict ) -> Tuple: if version.parse(scb.__version__ ) < version.parse('''1.4.12''' ): raise ImportWarning( '''To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n''' '''You can install it with `pip install "sacrebleu>=1.4.12"`.''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://github.com/mjpost/sacreBLEU#chrf--chrf''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=['''https://github.com/mjpost/sacreBLEU#chrf--chrf'''] , reference_urls=[ '''https://github.com/m-popovic/chrF''', ] , ) def __A ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : int = CHRF.CHAR_ORDER , SCREAMING_SNAKE_CASE__ : int = CHRF.WORD_ORDER , SCREAMING_SNAKE_CASE__ : int = CHRF.BETA , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : bool = False , ) -> Optional[Any]: __lowerCamelCase = len(references[0] ) if any(len(SCREAMING_SNAKE_CASE__ ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) __lowerCamelCase = [[refs[i] for refs in references] for i in range(SCREAMING_SNAKE_CASE__ )] __lowerCamelCase = CHRF(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = sb_chrf.corpus_score(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
270
1
from __future__ import annotations import math def SCREAMING_SNAKE_CASE__ ( __a , __a , __a , __a , __a ): if depth < 0: raise ValueError('Depth cannot be less than 0' ) if not scores: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , __a , __a , __a ) , minimax(depth + 1 , node_index * 2 + 1 , __a , __a , __a ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , __a , __a , __a ) , minimax(depth + 1 , node_index * 2 + 1 , __a , __a , __a ) , ) ) def SCREAMING_SNAKE_CASE__ ( ): snake_case_ : Optional[int] = [90, 23, 6, 33, 21, 65, 1_23, 3_44_23] snake_case_ : Dict = math.log(len(__a ) , 2 ) print(f"""Optimal value : {minimax(0 , 0 , __a , __a , __a )}""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
88
from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def SCREAMING_SNAKE_CASE__ ( __a , __a ): snake_case_ : List[Any] = k_size // 2 snake_case_ ,snake_case_ : Any = mgrid[0 - center : k_size - center, 0 - center : k_size - center] snake_case_ : Optional[Any] = 1 / (2 * pi * sigma) * exp(-(square(__a ) + square(__a )) / (2 * square(__a )) ) return g def SCREAMING_SNAKE_CASE__ ( __a , __a , __a ): snake_case_ ,snake_case_ : Union[str, Any] = image.shape[0], image.shape[1] # dst image height and width snake_case_ : int = height - k_size + 1 snake_case_ : Optional[int] = width - k_size + 1 # im2col, turn the k_size*k_size pixels into a row and np.vstack all rows snake_case_ : Optional[Any] = zeros((dst_height * dst_width, k_size * k_size) ) snake_case_ : Tuple = 0 for i, j in product(range(__a ) , range(__a ) ): snake_case_ : Optional[int] = ravel(image[i : i + k_size, j : j + k_size] ) snake_case_ : str = window row += 1 # turn the kernel into shape(k*k, 1) snake_case_ : List[Any] = gen_gaussian_kernel(__a , __a ) snake_case_ : str = ravel(__a ) # reshape and get the dst image snake_case_ : Optional[int] = dot(__a , __a ).reshape(__a , __a ).astype(__a ) return dst if __name__ == "__main__": # read original image _SCREAMING_SNAKE_CASE = imread(R"""../image_data/lena.jpg""") # turn image in gray scale value _SCREAMING_SNAKE_CASE = cvtColor(img, COLOR_BGR2GRAY) # get values with two different mask size _SCREAMING_SNAKE_CASE = gaussian_filter(gray, 3, sigma=1) _SCREAMING_SNAKE_CASE = gaussian_filter(gray, 5, sigma=0.8) # show result images imshow("""gaussian filter with 3x3 mask""", gaussianaxa) imshow("""gaussian filter with 5x5 mask""", gaussianaxa) waitKey()
88
1
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def A_ ( A__ ) -> Tuple: # A local function to see if a dot lands in the circle. def is_in_circle(A__ , A__ ) -> bool: a__ : List[str] = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle a__ : List[str] = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(A__ ) ) # The ratio of the area for circle to square is pi/4. a__ : Optional[Any] = proportion * 4 print(F'The estimated value of pi is {pi_estimate}' ) print(F'The numpy value of pi is {pi}' ) print(F'The total error is {abs(pi - pi_estimate )}' ) def A_ ( A__ , A__ , A__ = 0.0 , A__ = 1.0 , ) -> float: return mean( function_to_integrate(uniform(A__ , A__ ) ) for _ in range(A__ ) ) * (max_value - min_value) def A_ ( A__ , A__ = 0.0 , A__ = 1.0 ) -> None: def identity_function(A__ ) -> float: return x a__ : List[Any] = area_under_curve_estimator( A__ , A__ , A__ , A__ ) a__ : Union[str, Any] = (max_value * max_value - min_value * min_value) / 2 print('******************' ) print(F'Estimating area under y=x where x varies from {min_value} to {max_value}' ) print(F'Estimated value is {estimated_value}' ) print(F'Expected value is {expected_value}' ) print(F'Total error is {abs(estimated_value - expected_value )}' ) print('******************' ) def A_ ( A__ ) -> None: def function_to_integrate(A__ ) -> float: return sqrt(4.0 - x * x ) a__ : Dict = area_under_curve_estimator( A__ , A__ , 0.0 , 2.0 ) print('******************' ) print('Estimating pi using area_under_curve_estimator' ) print(F'Estimated value is {estimated_value}' ) print(F'Expected value is {pi}' ) print(F'Total error is {abs(estimated_value - pi )}' ) print('******************' ) if __name__ == "__main__": import doctest doctest.testmod()
99
"""simple docstring""" from ..utils import DummyObject, requires_backends class _A ( metaclass=lowerCAmelCase ): snake_case__ : str = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : List[Any] = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : Optional[Any] = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : int = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : List[str] = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : Optional[Any] = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] )
197
0
import webbrowser from sys import argv from urllib.parse import parse_qs, quote import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": UpperCAmelCase_ = '%20'.join(argv[1:]) if len(argv) > 1 else quote(str(input('Search: '))) print('Googling.....') UpperCAmelCase_ = f"""https://www.google.com/search?q={query}&num=100""" UpperCAmelCase_ = requests.get( url, headers={'User-Agent': str(UserAgent().random)}, ) try: UpperCAmelCase_ = ( BeautifulSoup(res.text, 'html.parser') .find('div', attrs={'class': 'yuRUbf'}) .find('a') .get('href') ) except AttributeError: UpperCAmelCase_ = parse_qs( BeautifulSoup(res.text, 'html.parser') .find('div', attrs={'class': 'kCrYT'}) .find('a') .get('href') )['url'][0] webbrowser.open(link)
355
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Union[str, Any] = 'yolos' def __init__( self: Dict , UpperCamelCase_: List[Any]=7_68 , UpperCamelCase_: Tuple=12 , UpperCamelCase_: int=12 , UpperCamelCase_: int=30_72 , UpperCamelCase_: List[str]="gelu" , UpperCamelCase_: Union[str, Any]=0.0 , UpperCamelCase_: int=0.0 , UpperCamelCase_: Optional[int]=0.02 , UpperCamelCase_: Dict=1E-12 , UpperCamelCase_: List[Any]=[5_12, 8_64] , UpperCamelCase_: Optional[int]=16 , UpperCamelCase_: Any=3 , UpperCamelCase_: Union[str, Any]=True , UpperCamelCase_: List[str]=1_00 , UpperCamelCase_: List[str]=True , UpperCamelCase_: Any=False , UpperCamelCase_: Optional[Any]=1 , UpperCamelCase_: Any=5 , UpperCamelCase_: Any=2 , UpperCamelCase_: Tuple=5 , UpperCamelCase_: str=2 , UpperCamelCase_: Any=0.1 , **UpperCamelCase_: Any , ): super().__init__(**UpperCamelCase_ ) __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = qkv_bias __lowerCamelCase = num_detection_tokens __lowerCamelCase = use_mid_position_embeddings __lowerCamelCase = auxiliary_loss # Hungarian matcher __lowerCamelCase = class_cost __lowerCamelCase = bbox_cost __lowerCamelCase = giou_cost # Loss coefficients __lowerCamelCase = bbox_loss_coefficient __lowerCamelCase = giou_loss_coefficient __lowerCamelCase = eos_coefficient class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = version.parse('1.11') @property def lowerCAmelCase__ ( self: Any ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def lowerCAmelCase__ ( self: Dict ): return 1E-4 @property def lowerCAmelCase__ ( self: Dict ): return 12
29
0
def _a ( SCREAMING_SNAKE_CASE_ : list[list[int]] , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : list[int] ): # 1. Validate that path exists between current and next vertices if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def _a ( SCREAMING_SNAKE_CASE_ : list[list[int]] , SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : int ): # Base Case if curr_ind == len(SCREAMING_SNAKE_CASE_ ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(SCREAMING_SNAKE_CASE_ ) ): if valid_connection(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): # Insert current vertex into path as next transition __lowerCAmelCase = next_ver # Validate created path if util_hamilton_cycle(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , curr_ind + 1 ): return True # Backtrack __lowerCAmelCase = -1 return False def _a ( SCREAMING_SNAKE_CASE_ : list[list[int]] , SCREAMING_SNAKE_CASE_ : int = 0 ): __lowerCAmelCase = [-1] * (len(SCREAMING_SNAKE_CASE_ ) + 1) # initialize start and end of path with starting index __lowerCAmelCase = __lowerCAmelCase = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , 1 ) else []
92
import argparse from argparse import Namespace import torch from torch import nn from transformers import XGLMConfig, XGLMForCausalLM def __lowerCamelCase ( __a :Dict ) -> Any: """simple docstring""" A__ = [ """decoder.version""", """decoder.output_projection.weight""", """_float_tensor""", """decoder.embed_positions._float_tensor""", ] for k in ignore_keys: state_dict.pop(__a , __a ) def __lowerCamelCase ( __a :str ) -> Union[str, Any]: """simple docstring""" A__ , A__ = emb.weight.shape A__ = nn.Linear(__a , __a , bias=__a ) A__ = emb.weight.data return lin_layer def __lowerCamelCase ( __a :str ) -> List[str]: """simple docstring""" A__ = torch.load(__a , map_location="""cpu""" ) A__ = Namespace(**checkpoint["""cfg"""]["""model"""] ) A__ = checkpoint["""model"""] remove_ignore_keys_(__a ) A__ = state_dict["""decoder.embed_tokens.weight"""].shape[0] A__ = {key.replace("""decoder""" , """model""" ): val for key, val in state_dict.items()} A__ = XGLMConfig( vocab_size=__a , max_position_embeddings=args.max_target_positions , num_layers=args.decoder_layers , attention_heads=args.decoder_attention_heads , ffn_dim=args.decoder_ffn_embed_dim , d_model=args.decoder_embed_dim , layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="""gelu""" , scale_embedding=not args.no_scale_embedding , tie_word_embeddings=args.share_decoder_input_output_embed , ) A__ = XGLMForCausalLM(__a ) A__ = model.load_state_dict(__a , strict=__a ) print(__a ) A__ = make_linear_from_emb(model.model.embed_tokens ) return model if __name__ == "__main__": A : int = argparse.ArgumentParser() # Required parameters parser.add_argument('''fairseq_path''', type=str, help='''path to a model.pt on local filesystem.''') parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') A : str = parser.parse_args() A : str = convert_fairseq_xglm_checkpoint_from_disk(args.fairseq_path) model.save_pretrained(args.pytorch_dump_folder_path)
274
0
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowerCAmelCase_ : Tuple = False class __SCREAMING_SNAKE_CASE (unittest.TestCase ): """simple docstring""" pass @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE (unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self : List[str] ): _a = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion" ) pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) _a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) _a = torch.manual_seed(0 ) _a = pipe( image=__a , generator=__a , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images _a = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _a = np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
360
'''simple docstring''' import requests lowerCAmelCase_ : List[Any] = 'YOUR API KEY' def _lowerCamelCase ( lowercase : str , lowercase : str = giphy_api_key ) -> list: _a = "+".join(query.split() ) _a = F'https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}' _a = requests.get(lowercase ).json()["data"] return [gif["url"] for gif in gifs] if __name__ == "__main__": print('\n'.join(get_gifs('space ship')))
346
0
def __lowerCamelCase ( UpperCAmelCase_ : int ): """simple docstring""" if p < 2: raise ValueError('''p should not be less than 2!''' ) elif p == 2: return True a :int = 4 a :Union[str, Any] = (1 << p) - 1 for _ in range(p - 2 ): a :List[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
94
from __future__ import annotations def __lowerCamelCase ( UpperCAmelCase_ : dict , UpperCAmelCase_ : str ): """simple docstring""" a , a :Optional[Any] = set(UpperCAmelCase_ ), [start] while stack: a :Optional[int] = stack.pop() explored.add(UpperCAmelCase_ ) # Differences from BFS: # 1) pop last element instead of first one # 2) add adjacent elements to stack without exploring them for adj in reversed(graph[v] ): if adj not in explored: stack.append(UpperCAmelCase_ ) return explored snake_case : Optional[int] = { '''A''': ['''B''', '''C''', '''D'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F'''], '''D''': ['''B''', '''D'''], '''E''': ['''B''', '''F'''], '''F''': ['''C''', '''E''', '''G'''], '''G''': ['''F'''], } if __name__ == "__main__": import doctest doctest.testmod() print(depth_first_search(G, '''A'''))
94
1
from __future__ import annotations def __lowercase ( __lowerCAmelCase : list[int] ): if len(__lowerCAmelCase ) == 0: return array a__ , a__ = min(__lowerCAmelCase ), max(__lowerCAmelCase ) # Compute the variables a__ = _max - _min + 1 a__ , a__ = [0] * holes_range, [0] * holes_range # Make the sorting. for i in array: a__ = i - _min a__ = i holes_repeat[index] += 1 # Makes the array back by replacing the numbers. a__ = 0 for i in range(__lowerCAmelCase ): while holes_repeat[i] > 0: a__ = holes[i] index += 1 holes_repeat[i] -= 1 # Returns the sorted array. return array if __name__ == "__main__": import doctest doctest.testmod() snake_case : List[str] = input('''Enter numbers separated by comma:\n''') snake_case : Optional[Any] = [int(x) for x in user_input.split(''',''')] print(pigeon_sort(unsorted))
358
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case : Optional[int] = { '''configuration_clipseg''': [ '''CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CLIPSegConfig''', '''CLIPSegTextConfig''', '''CLIPSegVisionConfig''', ], '''processing_clipseg''': ['''CLIPSegProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case : List[str] = [ '''CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CLIPSegModel''', '''CLIPSegPreTrainedModel''', '''CLIPSegTextModel''', '''CLIPSegVisionModel''', '''CLIPSegForImageSegmentation''', ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys snake_case : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
109
0
"""simple docstring""" import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = "" lowercase = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) lowercase = None # compression type in fsspec. ex: "gzip" lowercase = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self , __UpperCAmelCase = "" , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase ): '''simple docstring''' super().__init__(self , **__UpperCAmelCase ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode __UpperCamelCase = fsspec.open( __UpperCAmelCase , mode='rb' , protocol=__UpperCAmelCase , compression=self.compression , client_kwargs={ 'requote_redirect_url': False, # see https://github.com/huggingface/datasets/pull/5459 'trust_env': True, # Enable reading proxy env variables. **(target_options or {}).pop('client_kwargs' , {} ), # To avoid issues if it was already passed. } , **(target_options or {}) , ) __UpperCamelCase = os.path.basename(self.file.path.split('::' )[0] ) __UpperCamelCase = ( self.compressed_name[: self.compressed_name.rindex('.' )] if '.' in self.compressed_name else self.compressed_name ) __UpperCamelCase = None @classmethod def UpperCAmelCase ( cls , __UpperCAmelCase ): '''simple docstring''' return super()._strip_protocol(__UpperCAmelCase ).lstrip('/' ) def UpperCAmelCase ( self ): '''simple docstring''' if self.dir_cache is None: __UpperCamelCase = {**self.file.fs.info(self.file.path ), 'name': self.uncompressed_name} __UpperCamelCase = {f['name']: f} def UpperCAmelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.file.open().read() def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = "rb" , __UpperCAmelCase=None , __UpperCAmelCase=True , __UpperCAmelCase=None , **__UpperCAmelCase , ): '''simple docstring''' __UpperCamelCase = self._strip_protocol(__UpperCAmelCase ) if mode != "rb": raise ValueError(F'Tried to read with mode {mode} on file {self.file.path} opened with mode \'rb\'' ) return self.file.open() class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = "bz2" lowercase = "bz2" lowercase = ".bz2" class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = "gzip" lowercase = "gzip" lowercase = ".gz" class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = "lz4" lowercase = "lz4" lowercase = ".lz4" class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = "xz" lowercase = "xz" lowercase = ".xz" class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = "zstd" lowercase = "zstd" lowercase = ".zst" def __init__( self , __UpperCAmelCase , __UpperCAmelCase = "rb" , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = DEFAULT_BLOCK_SIZE , **__UpperCAmelCase , ): '''simple docstring''' super().__init__( fo=__UpperCAmelCase , mode=__UpperCAmelCase , target_protocol=__UpperCAmelCase , target_options=__UpperCAmelCase , block_size=__UpperCAmelCase , **__UpperCAmelCase , ) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 __UpperCamelCase = self.file.__enter__ class __lowerCAmelCase : def __init__( self , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = file_ def __enter__( self ): '''simple docstring''' self._file.__enter__() return self def __exit__( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' self._file.__exit__(*__UpperCAmelCase , **__UpperCAmelCase ) def __iter__( self ): '''simple docstring''' return iter(self._file ) def UpperCAmelCase ( self ): '''simple docstring''' return next(self._file ) def __getattr__( self , __UpperCAmelCase ): '''simple docstring''' return getattr(self._file , __UpperCAmelCase ) def fixed_enter(*__UpperCAmelCase , **__UpperCAmelCase ): return WrappedFile(_enter(*__UpperCAmelCase , **__UpperCAmelCase ) ) __UpperCamelCase = fixed_enter
316
"""simple docstring""" UpperCamelCase : dict[str, float] = { "km/h": 1.0, "m/s": 3.6, "mph": 1.60_93_44, "knot": 1.8_52, } UpperCamelCase : dict[str, float] = { "km/h": 1.0, "m/s": 0.2_77_77_77_78, "mph": 0.6_21_37_11_92, "knot": 0.5_39_95_68_03, } def A ( snake_case :float , snake_case :str , snake_case :str ) -> float: if unit_to not in speed_chart or unit_from not in speed_chart_inverse: __UpperCamelCase = ( f'Incorrect \'from_type\' or \'to_type\' value: {unit_from!r}, {unit_to!r}\n' f'Valid values are: {", ".join(snake_case )}' ) raise ValueError(snake_case ) return round(speed * speed_chart[unit_from] * speed_chart_inverse[unit_to] , 3 ) if __name__ == "__main__": import doctest doctest.testmod()
316
1
"""simple docstring""" import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_image_processing import CustomImageProcessor # noqa E402 lowercase__ = get_tests_dir('fixtures') class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[Any] = mock.Mock() a__: Any = 5_00 a__: Union[str, Any] = {} a__: List[str] = HTTPError a__: Optional[Any] = {} # Download this model to make sure it's in the cache. a__: List[Any] = ViTImageProcessor.from_pretrained('hf-internal-testing/tiny-random-vit') # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('requests.Session.request' , return_value=lowercase) as mock_head: a__: Optional[Any] = ViTImageProcessor.from_pretrained('hf-internal-testing/tiny-random-vit') # This check we did call the fake head request mock_head.assert_called() def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[Any] = ViTImageProcessor.from_pretrained( 'https://huggingface.co/hf-internal-testing/tiny-random-vit/resolve/main/preprocessor_config.json') def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' with self.assertRaises(lowercase): # config is in subfolder, the following should not work without specifying the subfolder a__: int = AutoImageProcessor.from_pretrained('hf-internal-testing/stable-diffusion-all-variants') a__: List[str] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/stable-diffusion-all-variants' , subfolder='feature_extractor') self.assertIsNotNone(lowercase) @is_staging_test class __snake_case ( unittest.TestCase ): @classmethod def lowerCamelCase_ ( cls) -> Optional[Any]: '''simple docstring''' a__: Any = TOKEN HfFolder.save_token(lowercase) @classmethod def lowerCamelCase_ ( cls) -> Tuple: '''simple docstring''' try: delete_repo(token=cls._token , repo_id='test-image-processor') except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-image-processor-org') except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-image-processor') except HTTPError: pass def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Optional[int] = ViTImageProcessor.from_pretrained(lowercase) image_processor.push_to_hub('test-image-processor' , use_auth_token=self._token) a__: Optional[int] = ViTImageProcessor.from_pretrained(f'{USER}/test-image-processor') for k, v in image_processor.__dict__.items(): self.assertEqual(lowercase , getattr(lowercase , lowercase)) # Reset repo delete_repo(token=self._token , repo_id='test-image-processor') # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( lowercase , repo_id='test-image-processor' , push_to_hub=lowercase , use_auth_token=self._token) a__: Optional[Any] = ViTImageProcessor.from_pretrained(f'{USER}/test-image-processor') for k, v in image_processor.__dict__.items(): self.assertEqual(lowercase , getattr(lowercase , lowercase)) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: str = ViTImageProcessor.from_pretrained(lowercase) image_processor.push_to_hub('valid_org/test-image-processor' , use_auth_token=self._token) a__: List[Any] = ViTImageProcessor.from_pretrained('valid_org/test-image-processor') for k, v in image_processor.__dict__.items(): self.assertEqual(lowercase , getattr(lowercase , lowercase)) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-image-processor') # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( lowercase , repo_id='valid_org/test-image-processor-org' , push_to_hub=lowercase , use_auth_token=self._token) a__: Optional[Any] = ViTImageProcessor.from_pretrained('valid_org/test-image-processor-org') for k, v in image_processor.__dict__.items(): self.assertEqual(lowercase , getattr(lowercase , lowercase)) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' CustomImageProcessor.register_for_auto_class() a__: Dict = CustomImageProcessor.from_pretrained(lowercase) image_processor.push_to_hub('test-dynamic-image-processor' , use_auth_token=self._token) # This has added the proper auto_map field to the config self.assertDictEqual( image_processor.auto_map , {'AutoImageProcessor': 'custom_image_processing.CustomImageProcessor'} , ) a__: Optional[int] = AutoImageProcessor.from_pretrained( f'{USER}/test-dynamic-image-processor' , trust_remote_code=lowercase) # Can't make an isinstance check because the new_image_processor is from the CustomImageProcessor class of a dynamic module self.assertEqual(new_image_processor.__class__.__name__ , 'CustomImageProcessor')
366
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: lowercase__ = None lowercase__ = logging.get_logger(__name__) lowercase__ = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} lowercase__ = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'facebook/mbart-large-en-ro': 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json', 'facebook/mbart-large-cc25': 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json', }, } lowercase__ = { 'facebook/mbart-large-en-ro': 1024, 'facebook/mbart-large-cc25': 1024, } # fmt: off lowercase__ = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class __snake_case ( __lowerCAmelCase ): a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = ["""input_ids""", """attention_mask"""] a__ = MBartTokenizer a__ = [] a__ = [] def __init__( self , lowercase=None , lowercase=None , lowercase="<s>" , lowercase="</s>" , lowercase="</s>" , lowercase="<s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase="<mask>" , lowercase=None , lowercase=None , lowercase=None , **lowercase , ) -> Tuple: '''simple docstring''' a__: List[str] = AddedToken(lowercase , lstrip=lowercase , rstrip=lowercase) if isinstance(lowercase , lowercase) else mask_token super().__init__( vocab_file=lowercase , tokenizer_file=lowercase , bos_token=lowercase , eos_token=lowercase , sep_token=lowercase , cls_token=lowercase , unk_token=lowercase , pad_token=lowercase , mask_token=lowercase , src_lang=lowercase , tgt_lang=lowercase , additional_special_tokens=lowercase , **lowercase , ) a__: Tuple = vocab_file a__: Union[str, Any] = False if not self.vocab_file else True a__: Union[str, Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens}) a__: int = { lang_code: self.convert_tokens_to_ids(lowercase) for lang_code in FAIRSEQ_LANGUAGE_CODES } a__: List[Any] = src_lang if src_lang is not None else 'en_XX' a__: Tuple = self.convert_tokens_to_ids(self._src_lang) a__: str = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def lowerCamelCase_ ( self) -> str: '''simple docstring''' return self._src_lang @src_lang.setter def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def lowerCamelCase_ ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowerCamelCase_ ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' a__: Any = [self.sep_token_id] a__: List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , **lowercase) -> Union[str, Any]: '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') a__: Union[str, Any] = src_lang a__: Any = self(lowercase , add_special_tokens=lowercase , return_tensors=lowercase , **lowercase) a__: str = self.convert_tokens_to_ids(lowercase) a__: Any = tgt_lang_id return inputs def lowerCamelCase_ ( self , lowercase , lowercase = "en_XX" , lowercase = None , lowercase = "ro_RO" , **lowercase , ) -> BatchEncoding: '''simple docstring''' a__: Any = src_lang a__: List[Any] = tgt_lang return super().prepare_seqaseq_batch(lowercase , lowercase , **lowercase) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: int = self.convert_tokens_to_ids(lowercase) a__: List[Any] = [] a__: List[str] = [self.eos_token_id, self.cur_lang_code] a__: Dict = self.convert_ids_to_tokens(self.prefix_tokens) a__: Any = self.convert_ids_to_tokens(self.suffix_tokens) a__: int = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: str = self.convert_tokens_to_ids(lowercase) a__: List[Any] = [] a__: Dict = [self.eos_token_id, self.cur_lang_code] a__: Any = self.convert_ids_to_tokens(self.prefix_tokens) a__: Optional[Any] = self.convert_ids_to_tokens(self.suffix_tokens) a__: str = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def lowerCamelCase_ ( self , lowercase , lowercase = None) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.') if not os.path.isdir(lowercase): logger.error(f'Vocabulary path ({save_directory}) should be a directory.') return a__: Any = os.path.join( lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase): copyfile(self.vocab_file , lowercase) return (out_vocab_file,)
203
0
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class A ( UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): UpperCamelCase__ : str =IFImgaImgSuperResolutionPipeline UpperCamelCase__ : Dict =TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'width', 'height'} UpperCamelCase__ : Optional[Any] =TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'original_image'} ) UpperCamelCase__ : List[str] =PipelineTesterMixin.required_optional_params - {'latents'} def lowerCamelCase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" return self._get_superresolution_dummy_components() def lowerCamelCase ( self : Tuple , lowercase_ : str , lowercase_ : Optional[int]=0 ) -> Dict: """simple docstring""" if str(lowercase_ ).startswith('mps' ): _lowerCamelCase : Dict =torch.manual_seed(lowercase_ ) else: _lowerCamelCase : Tuple =torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) _lowerCamelCase : Dict =floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) _lowerCamelCase : str =floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) _lowerCamelCase : Optional[Any] ={ 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def lowerCamelCase ( self : Dict ) -> Tuple: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def lowerCamelCase ( self : Tuple ) -> List[str]: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def lowerCamelCase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1E-1 ) def lowerCamelCase ( self : int ) -> List[Any]: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def lowerCamelCase ( self : Union[str, Any] ) -> Any: """simple docstring""" self._test_save_load_local() def lowerCamelCase ( self : List[str] ) -> Dict: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
199
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer lowerCamelCase = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast lowerCamelCase = TaTokenizerFast lowerCamelCase = {'configuration_mt5': ['MT5Config', 'MT5OnnxConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ 'MT5EncoderModel', 'MT5ForConditionalGeneration', 'MT5ForQuestionAnswering', 'MT5Model', 'MT5PreTrainedModel', 'MT5Stack', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['TFMT5EncoderModel', 'TFMT5ForConditionalGeneration', 'TFMT5Model'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['FlaxMT5EncoderModel', 'FlaxMT5ForConditionalGeneration', 'FlaxMT5Model'] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys lowerCamelCase = _LazyModule( __name__, globals()['__file__'], _import_structure, extra_objects={'MT5Tokenizer': MTaTokenizer, 'MT5TokenizerFast': MTaTokenizerFast}, module_spec=__spec__, )
199
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass class lowercase__ : # setable values _UpperCAmelCase :Optional[int] = None _UpperCAmelCase :Optional[jnp.ndarray] = None _UpperCAmelCase :Optional[jnp.ndarray] = None # sigma(t_i) @classmethod def UpperCAmelCase__ ( cls : Any ): return cls() @dataclass class lowercase__ ( snake_case__ ): _UpperCAmelCase :jnp.ndarray _UpperCAmelCase :jnp.ndarray _UpperCAmelCase :KarrasVeSchedulerState class lowercase__ ( snake_case__, snake_case__ ): @property def UpperCAmelCase__ ( self : Dict ): return True @register_to_config def __init__( self : Dict , snake_case__ : float = 0.02 , snake_case__ : float = 100 , snake_case__ : float = 1.007 , snake_case__ : float = 80 , snake_case__ : float = 0.05 , snake_case__ : float = 50 , ): pass def UpperCAmelCase__ ( self : Optional[int] ): return KarrasVeSchedulerState.create() def UpperCAmelCase__ ( self : Optional[Any] , snake_case__ : KarrasVeSchedulerState , snake_case__ : int , snake_case__ : Tuple = () ): lowerCamelCase_ : str =jnp.arange(0 , snake_case__ )[::-1].copy() lowerCamelCase_ : str =[ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in timesteps ] return state.replace( num_inference_steps=snake_case__ , schedule=jnp.array(snake_case__ , dtype=jnp.floataa ) , timesteps=snake_case__ , ) def UpperCAmelCase__ ( self : Optional[int] , snake_case__ : KarrasVeSchedulerState , snake_case__ : jnp.ndarray , snake_case__ : float , snake_case__ : random.KeyArray , ): if self.config.s_min <= sigma <= self.config.s_max: lowerCamelCase_ : List[Any] =min(self.config.s_churn / state.num_inference_steps , 2**0.5 - 1 ) else: lowerCamelCase_ : Tuple =0 # sample eps ~ N(0, S_noise^2 * I) lowerCamelCase_ : int =random.split(snake_case__ , num=1 ) lowerCamelCase_ : List[str] =self.config.s_noise * random.normal(key=snake_case__ , shape=sample.shape ) lowerCamelCase_ : Optional[Any] =sigma + gamma * sigma lowerCamelCase_ : Tuple =sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def UpperCAmelCase__ ( self : List[str] , snake_case__ : KarrasVeSchedulerState , snake_case__ : jnp.ndarray , snake_case__ : float , snake_case__ : float , snake_case__ : jnp.ndarray , snake_case__ : bool = True , ): lowerCamelCase_ : Optional[Any] =sample_hat + sigma_hat * model_output lowerCamelCase_ : Dict =(sample_hat - pred_original_sample) / sigma_hat lowerCamelCase_ : Dict =sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=snake_case__ , derivative=snake_case__ , state=snake_case__ ) def UpperCAmelCase__ ( self : Optional[Any] , snake_case__ : KarrasVeSchedulerState , snake_case__ : jnp.ndarray , snake_case__ : float , snake_case__ : float , snake_case__ : jnp.ndarray , snake_case__ : jnp.ndarray , snake_case__ : jnp.ndarray , snake_case__ : bool = True , ): lowerCamelCase_ : List[Any] =sample_prev + sigma_prev * model_output lowerCamelCase_ : Dict =(sample_prev - pred_original_sample) / sigma_prev lowerCamelCase_ : Union[str, Any] =sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=snake_case__ , derivative=snake_case__ , state=snake_case__ ) def UpperCAmelCase__ ( self : Union[str, Any] , snake_case__ : KarrasVeSchedulerState , snake_case__ : Optional[Any] , snake_case__ : List[str] , snake_case__ : Tuple ): raise NotImplementedError()
209
"""simple docstring""" import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor A__ : str = logging.get_logger(__name__) class lowercase__ ( snake_case__ ): def __init__( self : Optional[Any] , *snake_case__ : int , **snake_case__ : Any ): warnings.warn( "The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use PerceiverImageProcessor instead." , snake_case__ , ) super().__init__(*snake_case__ , **snake_case__ )
209
1
import unittest from knapsack import knapsack as k class UpperCAmelCase ( unittest.TestCase ): def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' snake_case : Optional[int] = 0 snake_case : Dict = [0] snake_case : List[str] = [0] snake_case : Optional[int] = len(snake_case__ ) self.assertEqual(k.knapsack(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) , 0 ) snake_case : Any = [60] snake_case : Optional[int] = [10] snake_case : List[Any] = len(snake_case__ ) self.assertEqual(k.knapsack(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) , 0 ) def _SCREAMING_SNAKE_CASE (self : Dict ) -> List[str]: '''simple docstring''' snake_case : Optional[Any] = 3 snake_case : Dict = [1, 2, 3] snake_case : Any = [3, 2, 1] snake_case : str = len(snake_case__ ) self.assertEqual(k.knapsack(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) , 5 ) def _SCREAMING_SNAKE_CASE (self : Dict ) -> List[Any]: '''simple docstring''' snake_case : Any = 50 snake_case : str = [60, 1_00, 1_20] snake_case : int = [10, 20, 30] snake_case : List[str] = len(snake_case__ ) self.assertEqual(k.knapsack(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) , 2_20 ) if __name__ == "__main__": unittest.main()
59
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase_ = { "configuration_funnel": ["FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP", "FunnelConfig"], "convert_funnel_original_tf_checkpoint_to_pytorch": [], "tokenization_funnel": ["FunnelTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = ["FunnelTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ "FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST", "FunnelBaseModel", "FunnelForMaskedLM", "FunnelForMultipleChoice", "FunnelForPreTraining", "FunnelForQuestionAnswering", "FunnelForSequenceClassification", "FunnelForTokenClassification", "FunnelModel", "FunnelPreTrainedModel", "load_tf_weights_in_funnel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ "TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST", "TFFunnelBaseModel", "TFFunnelForMaskedLM", "TFFunnelForMultipleChoice", "TFFunnelForPreTraining", "TFFunnelForQuestionAnswering", "TFFunnelForSequenceClassification", "TFFunnelForTokenClassification", "TFFunnelModel", "TFFunnelPreTrainedModel", ] if TYPE_CHECKING: from .configuration_funnel import FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP, FunnelConfig from .tokenization_funnel import FunnelTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_funnel_fast import FunnelTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_funnel import ( FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, FunnelBaseModel, FunnelForMaskedLM, FunnelForMultipleChoice, FunnelForPreTraining, FunnelForQuestionAnswering, FunnelForSequenceClassification, FunnelForTokenClassification, FunnelModel, FunnelPreTrainedModel, load_tf_weights_in_funnel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_funnel import ( TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, TFFunnelPreTrainedModel, ) else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
251
0
import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( """split_dict""" , [ SplitDict(), SplitDict({"""train""": SplitInfo(name="""train""" , num_bytes=1337 , num_examples=42 , dataset_name="""my_dataset""" )} ), SplitDict({"""train""": SplitInfo(name="""train""" , num_bytes=1337 , num_examples=42 )} ), SplitDict({"""train""": SplitInfo()} ), ] , ) def lowerCamelCase_ ( _a : SplitDict ): '''simple docstring''' UpperCAmelCase_ : str = split_dict._to_yaml_list() assert len(_a ) == len(_a ) UpperCAmelCase_ : int = SplitDict._from_yaml_list(_a ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump UpperCAmelCase_ : Union[str, Any] = None # the split name of split_dict takes over the name of the split info object UpperCAmelCase_ : Optional[int] = split_name assert split_dict == reloaded @pytest.mark.parametrize( """split_info""" , [SplitInfo(), SplitInfo(dataset_name=_a ), SplitInfo(dataset_name="""my_dataset""" )] ) def lowerCamelCase_ ( _a : Optional[int] ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = asdict(SplitDict({"""train""": split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
360
from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar UpperCamelCase_ = TypeVar('''T''') UpperCamelCase_ = TypeVar('''U''') class _snake_case ( Generic[T, U] ): '''simple docstring''' def __init__( self: List[str] ,lowerCamelCase_: T | None ,lowerCamelCase_: U | None ) -> Optional[int]: UpperCAmelCase_ : Any = key UpperCAmelCase_ : List[str] = val UpperCAmelCase_ : DoubleLinkedListNode[T, U] | None = None UpperCAmelCase_ : DoubleLinkedListNode[T, U] | None = None def __repr__( self: Optional[Any] ) -> str: return ( F'''Node: key: {self.key}, val: {self.val}, ''' F'''has next: {bool(self.next )}, has prev: {bool(self.prev )}''' ) class _snake_case ( Generic[T, U] ): '''simple docstring''' def __init__( self: Tuple ) -> None: UpperCAmelCase_ : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(lowerCamelCase_ ,lowerCamelCase_ ) UpperCAmelCase_ : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(lowerCamelCase_ ,lowerCamelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = self.rear, self.head def __repr__( self: Tuple ) -> str: UpperCAmelCase_ : List[str] = ["""DoubleLinkedList"""] UpperCAmelCase_ : Optional[Any] = self.head while node.next is not None: rep.append(str(lowerCamelCase_ ) ) UpperCAmelCase_ : str = node.next rep.append(str(self.rear ) ) return ",\n ".join(lowerCamelCase_ ) def A__ ( self: Tuple ,lowerCamelCase_: DoubleLinkedListNode[T, U] ) -> None: UpperCAmelCase_ : str = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None UpperCAmelCase_ : Any = node UpperCAmelCase_ : Tuple = previous UpperCAmelCase_ : List[Any] = node UpperCAmelCase_ : List[str] = self.rear def A__ ( self: str ,lowerCamelCase_: DoubleLinkedListNode[T, U] ) -> DoubleLinkedListNode[T, U] | None: if node.prev is None or node.next is None: return None UpperCAmelCase_ : str = node.next UpperCAmelCase_ : Optional[int] = node.prev UpperCAmelCase_ : List[str] = None UpperCAmelCase_ : int = None return node class _snake_case ( Generic[T, U] ): '''simple docstring''' A__ : dict[Callable[[T], U], LRUCache[T, U]] = {} def __init__( self: Any ,lowerCamelCase_: int ) -> Union[str, Any]: UpperCAmelCase_ : DoubleLinkedList[T, U] = DoubleLinkedList() UpperCAmelCase_ : Optional[Any] = capacity UpperCAmelCase_ : Any = 0 UpperCAmelCase_ : Any = 0 UpperCAmelCase_ : Optional[Any] = 0 UpperCAmelCase_ : dict[T, DoubleLinkedListNode[T, U]] = {} def __repr__( self: Any ) -> str: return ( F'''CacheInfo(hits={self.hits}, misses={self.miss}, ''' F'''capacity={self.capacity}, current size={self.num_keys})''' ) def __contains__( self: Union[str, Any] ,lowerCamelCase_: T ) -> bool: return key in self.cache def A__ ( self: Any ,lowerCamelCase_: T ) -> U | None: # Note: pythonic interface would throw KeyError rather than return None if key in self.cache: self.hits += 1 UpperCAmelCase_ : DoubleLinkedListNode[T, U] = self.cache[key] UpperCAmelCase_ : Any = self.list.remove(self.cache[key] ) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(lowerCamelCase_ ) return node.val self.miss += 1 return None def A__ ( self: Optional[Any] ,lowerCamelCase_: T ,lowerCamelCase_: U ) -> None: if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity UpperCAmelCase_ : Tuple = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(lowerCamelCase_ ) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 UpperCAmelCase_ : Tuple = DoubleLinkedListNode(lowerCamelCase_ ,lowerCamelCase_ ) self.list.add(self.cache[key] ) self.num_keys += 1 else: # bump node to the end of the list, update value UpperCAmelCase_ : Tuple = self.list.remove(self.cache[key] ) assert node is not None # node guaranteed to be in list UpperCAmelCase_ : int = value self.list.add(lowerCamelCase_ ) @classmethod def A__ ( cls: List[str] ,lowerCamelCase_: int = 128 ) -> Callable[[Callable[[T], U]], Callable[..., U]]: def cache_decorator_inner(lowerCamelCase_: Callable[[T], U] ) -> Callable[..., U]: def cache_decorator_wrapper(*lowerCamelCase_: T ) -> U: if func not in cls.decorator_function_to_instance_map: UpperCAmelCase_ : Union[str, Any] = LRUCache(lowerCamelCase_ ) UpperCAmelCase_ : Any = cls.decorator_function_to_instance_map[func].get(args[0] ) if result is None: UpperCAmelCase_ : Optional[Any] = func(*lowerCamelCase_ ) cls.decorator_function_to_instance_map[func].put(args[0] ,lowerCamelCase_ ) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(lowerCamelCase_ ,"""cache_info""" ,lowerCamelCase_ ) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
59
0
import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = inspect.getfile(accelerate.test_utils) lowercase = os.path.sep.join( mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''external_deps''', '''test_metrics.py''']) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 lowercase = test_metrics @require_cpu def A__ ( self): debug_launcher(self.test_metrics.main ,num_processes=1) @require_cpu def A__ ( self): debug_launcher(self.test_metrics.main) @require_single_gpu def A__ ( self): self.test_metrics.main() @require_multi_gpu def A__ ( self): print(f'Found {torch.cuda.device_count()} devices.') lowercase = ['''torchrun''', f'--nproc_per_node={torch.cuda.device_count()}', self.test_file_path] with patch_environment(omp_num_threads=1): execute_subprocess_async(a_ ,env=os.environ.copy())
101
"""simple docstring""" from __future__ import annotations def lowercase ( _snake_case : int | str ) ->bool: """simple docstring""" __snake_case : List[str] = str(_snake_case ) return n == n[::-1] def lowercase ( _snake_case : int = 1_000_000 ) ->str: """simple docstring""" __snake_case : Union[str, Any] = 0 for i in range(1 , _snake_case ): if is_palindrome(_snake_case ) and is_palindrome(bin(_snake_case ).split('''b''' )[1] ): total += i return total if __name__ == "__main__": print(solution(int(str(input().strip()))))
102
0
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class lowercase : @property def a ( self ): return self.get_dummy_input() @property def a ( self ): if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(F'''\'{self.block_type}\' is not a supported block_type. Set it to \'up\', \'mid\', or \'down\'.''' ) def a ( self , snake_case=True , snake_case=False , snake_case=False , snake_case=False , ): snake_case_ = 4 snake_case_ = 32 snake_case_ = (32, 32) snake_case_ = torch.manual_seed(0 ) snake_case_ = torch.device(snake_case ) snake_case_ = (batch_size, num_channels) + sizes snake_case_ = randn_tensor(snake_case , generator=snake_case , device=snake_case ) snake_case_ = {'hidden_states': hidden_states} if include_temb: snake_case_ = 128 snake_case_ = randn_tensor((batch_size, temb_channels) , generator=snake_case , device=snake_case ) if include_res_hidden_states_tuple: snake_case_ = torch.manual_seed(1 ) snake_case_ = (randn_tensor(snake_case , generator=snake_case , device=snake_case ),) if include_encoder_hidden_states: snake_case_ = floats_tensor((batch_size, 32, 32) ).to(snake_case ) if include_skip_sample: snake_case_ = randn_tensor(((batch_size, 3) + sizes) , generator=snake_case , device=snake_case ) return dummy_input def a ( self ): snake_case_ = { 'in_channels': 32, 'out_channels': 32, 'temb_channels': 128, } if self.block_type == "up": snake_case_ = 32 if self.block_type == "mid": init_dict.pop('out_channels' ) snake_case_ = self.dummy_input return init_dict, inputs_dict def a ( self , snake_case ): snake_case_ , snake_case_ = self.prepare_init_args_and_inputs_for_common() snake_case_ = self.block_class(**snake_case ) unet_block.to(snake_case ) unet_block.eval() with torch.no_grad(): snake_case_ = unet_block(**snake_case ) if isinstance(snake_case , snake_case ): snake_case_ = output[0] self.assertEqual(output.shape , self.output_shape ) snake_case_ = output[0, -1, -3:, -3:] snake_case_ = torch.tensor(snake_case ).to(snake_case ) assert torch_all_close(output_slice.flatten() , snake_case , atol=5e-3 ) @unittest.skipIf(torch_device == 'mps' , 'Training is not supported in mps' ) def a ( self ): snake_case_ , snake_case_ = self.prepare_init_args_and_inputs_for_common() snake_case_ = self.block_class(**snake_case ) model.to(snake_case ) model.train() snake_case_ = model(**snake_case ) if isinstance(snake_case , snake_case ): snake_case_ = output[0] snake_case_ = torch.device(snake_case ) snake_case_ = randn_tensor(output.shape , device=snake_case ) snake_case_ = torch.nn.functional.mse_loss(snake_case , snake_case ) loss.backward()
200
def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' return int((input_a, input_a).count(0 ) != 0 ) def __lowerCamelCase ( ): '''simple docstring''' assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
200
1
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_A ) class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = field(default="""summarization""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) a__ = Features({"""text""": Value("""string""" )} ) a__ = Features({"""summary""": Value("""string""" )} ) a__ = "text" a__ = "summary" @property def _lowercase ( self : Any ) -> Dict[str, str]: """simple docstring""" return {self.text_column: "text", self.summary_column: "summary"}
88
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _lowercase ( self : List[str] , UpperCamelCase__ : int ) -> str: """simple docstring""" for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] , model_result["""ss"""] ): __magic_name__ = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(UpperCamelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __magic_name__ = """sgugger/tiny-distilbert-classification""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , only_pretrain_model=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Tuple ) -> List[Any]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = AutoConfig.from_pretrained(UpperCamelCase__ ) __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ , [config] ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = AutoConfig.from_pretrained(UpperCamelCase__ ) __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ , [config] ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = AutoConfig.from_pretrained(UpperCamelCase__ ) __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ , [config] ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" __magic_name__ = """patrickvonplaten/t5-tiny-random""" __magic_name__ = AutoConfig.from_pretrained(UpperCamelCase__ ) __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ , configs=[config] ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , """Cannot do xla on CPU.""" ) def _lowercase ( self : Tuple ) -> int: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , use_xla=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self : Union[str, Any] ) -> Dict: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCamelCase__ , save_to_csv=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCamelCase__ , """inf_time.csv""" ) , inference_memory_csv_file=os.path.join(UpperCamelCase__ , """inf_mem.csv""" ) , env_info_csv_file=os.path.join(UpperCamelCase__ , """env.csv""" ) , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCamelCase__ , """inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , """inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , """env.csv""" ) ).exists() ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" __magic_name__ = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(UpperCamelCase__ : Dict ): self.assertTrue(hasattr(UpperCamelCase__ , """sequential""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """cumulative""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """current""" ) ) self.assertTrue(hasattr(UpperCamelCase__ , """total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: __magic_name__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCamelCase__ , """log.txt""" ) , log_print=UpperCamelCase__ , trace_memory_line_by_line=UpperCamelCase__ , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) __magic_name__ = TensorFlowBenchmark(UpperCamelCase__ ) __magic_name__ = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , """log.txt""" ) ).exists() )
88
1
from collections.abc import Callable def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> float: """simple docstring""" A : float = a A : float = b if function(_lowerCAmelCase ) == 0: # one of the a or b is a root for the function return a elif function(_lowerCAmelCase ) == 0: return b elif ( function(_lowerCAmelCase ) * function(_lowerCAmelCase ) > 0 ): # if none of these are root and they are both positive or negative, # then this algorithm can't find the root raise ValueError("""could not find root in given interval.""" ) else: A : float = start + (end - start) / 2.0 while abs(start - mid ) > 10**-7: # until precisely equals to 10^-7 if function(_lowerCAmelCase ) == 0: return mid elif function(_lowerCAmelCase ) * function(_lowerCAmelCase ) < 0: A : Optional[Any] = mid else: A : Any = mid A : Dict = start + (end - start) / 2.0 return mid def __UpperCamelCase ( _lowerCAmelCase ) -> float: """simple docstring""" return x**3 - 2 * x - 5 if __name__ == "__main__": print(bisection(f, 1, 1_000)) import doctest doctest.testmod()
353
from random import randint, random def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = False , _lowerCAmelCase = False , _lowerCAmelCase = 5 , ) -> list: """simple docstring""" A : Any = [[-1] * number_of_cells] # Create a highway without any car A : Tuple = 0 A : Dict = max(_lowerCAmelCase , 0 ) while i < number_of_cells: A : Any = ( randint(0 , _lowerCAmelCase ) if random_speed else initial_speed ) # Place the cars i += ( randint(1 , max_speed * 2 ) if random_frequency else frequency ) # Arbitrary number, may need tuning return highway def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ) -> int: """simple docstring""" A : List[str] = 0 A : Dict = highway_now[car_index + 1 :] for cell in range(len(_lowerCAmelCase ) ): # May need a better name for this if cells[cell] != -1: # If the cell is not empty then return distance # we have the distance we wanted distance += 1 # Here if the car is near the end of the highway return distance + get_distance(_lowerCAmelCase , -1 ) def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> list: """simple docstring""" A : str = len(_lowerCAmelCase ) # Beforce calculations, the highway is empty A : Any = [-1] * number_of_cells for car_index in range(_lowerCAmelCase ): if highway_now[car_index] != -1: # Add 1 to the current speed of the car and cap the speed A : str = min(highway_now[car_index] + 1 , _lowerCAmelCase ) # Number of empty cell before the next car A : Optional[int] = get_distance(_lowerCAmelCase , _lowerCAmelCase ) - 1 # We can't have the car causing an accident A : Any = min(next_highway[car_index] , _lowerCAmelCase ) if random() < probability: # Randomly, a driver will slow down A : Optional[Any] = max(next_highway[car_index] - 1 , 0 ) return next_highway def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> list: """simple docstring""" A : Any = len(highway[0] ) for i in range(_lowerCAmelCase ): A : Optional[int] = update(highway[i] , _lowerCAmelCase , _lowerCAmelCase ) A : Tuple = [-1] * number_of_cells for car_index in range(_lowerCAmelCase ): A : Dict = next_speeds_calculated[car_index] if speed != -1: # Change the position based on the speed (with % to create the loop) A : Optional[int] = (car_index + speed) % number_of_cells # Commit the change of position A : Dict = speed highway.append(_lowerCAmelCase ) return highway if __name__ == "__main__": import doctest doctest.testmod()
115
0
'''simple docstring''' __snake_case ={ "joule": 1.0, "kilojoule": 1_000, "megajoule": 1_000_000, "gigajoule": 1_000_000_000, "wattsecond": 1.0, "watthour": 3_600, "kilowatthour": 3_600_000, "newtonmeter": 1.0, "calorie_nutr": 4_186.8, "kilocalorie_nutr": 4_186_800.00, "electronvolt": 1.6_02_17_66_34e-19, "britishthermalunit_it": 1_055.05_585, "footpound": 1.3_5_5_8_1_8, } def a_ ( lowerCamelCase : str , lowerCamelCase : str , lowerCamelCase : float ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowerCAmelCase = ( f'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' f'''Valid values are: {', '.join(lowerCamelCase )}''' ) raise ValueError(lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
4
def lowercase__ ( __snake_case : Dict ): '''simple docstring''' if not head: return True # split the list to two parts UpperCAmelCase_ , UpperCAmelCase_ : Any = head.next, head while fast and fast.next: UpperCAmelCase_ : str = fast.next.next UpperCAmelCase_ : Union[str, Any] = slow.next UpperCAmelCase_ : int = slow.next UpperCAmelCase_ : List[Any] = None # Don't forget here! But forget still works! # reverse the second part UpperCAmelCase_ : Tuple = None while second: UpperCAmelCase_ : int = second.next UpperCAmelCase_ : Any = node UpperCAmelCase_ : Optional[Any] = second UpperCAmelCase_ : Tuple = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False UpperCAmelCase_ : Optional[Any] = node.next UpperCAmelCase_ : Dict = head.next return True def lowercase__ ( __snake_case : Union[str, Any] ): '''simple docstring''' if not head or not head.next: return True # 1. Get the midpoint (slow) UpperCAmelCase_ : Any = head while fast and fast.next: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = fast.next.next, slow.next # 2. Push the second half into the stack UpperCAmelCase_ : List[str] = [slow.val] while slow.next: UpperCAmelCase_ : List[str] = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False UpperCAmelCase_ : int = cur.next return True def lowercase__ ( __snake_case : Dict ): '''simple docstring''' if not head or not head.next: return True UpperCAmelCase_ : Tuple = {} UpperCAmelCase_ : int = 0 while head: if head.val in d: d[head.val].append(__snake_case ) else: UpperCAmelCase_ : List[Any] = [pos] UpperCAmelCase_ : Any = head.next pos += 1 UpperCAmelCase_ : Dict = pos - 1 UpperCAmelCase_ : Optional[int] = 0 for v in d.values(): if len(__snake_case ) % 2 != 0: middle += 1 else: UpperCAmelCase_ : int = 0 for i in range(0 , len(__snake_case ) ): if v[i] + v[len(__snake_case ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
29
0
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets __UpperCAmelCase : Any = "\\n@inproceedings{lin-2004-rouge,\n title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\",\n author = \"Lin, Chin-Yew\",\n booktitle = \"Text Summarization Branches Out\",\n month = jul,\n year = \"2004\",\n address = \"Barcelona, Spain\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W04-1013\",\n pages = \"74--81\",\n}\n" __UpperCAmelCase : Union[str, Any] = "\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n" __UpperCAmelCase : Any = "\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring,\n `\"rougeL\"`: Longest common subsequence based scoring.\n `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric(\'rouge\')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\']\n >>> print(results[\"rouge1\"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results[\"rouge1\"].mid.fmeasure)\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __snake_case ( datasets.Metric ): '''simple docstring''' def UpperCAmelCase__ ( self : Tuple ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/google-research/google-research/tree/master/rouge"""] , reference_urls=[ """https://en.wikipedia.org/wiki/ROUGE_(metric)""", """https://github.com/google-research/google-research/tree/master/rouge""", ] , ) def UpperCAmelCase__ ( self : List[Any] , A : Optional[Any] , A : Tuple , A : List[Any]=None , A : Optional[int]=True , A : Optional[int]=False ): if rouge_types is None: __snake_case: List[Any] = ['rouge1', 'rouge2', 'rougeL', 'rougeLsum'] __snake_case: List[Any] = rouge_scorer.RougeScorer(rouge_types=_UpperCAmelCase , use_stemmer=_UpperCAmelCase ) if use_aggregator: __snake_case: Optional[int] = scoring.BootstrapAggregator() else: __snake_case: List[str] = [] for ref, pred in zip(_UpperCAmelCase , _UpperCAmelCase ): __snake_case: List[str] = scorer.score(_UpperCAmelCase , _UpperCAmelCase ) if use_aggregator: aggregator.add_scores(_UpperCAmelCase ) else: scores.append(_UpperCAmelCase ) if use_aggregator: __snake_case: Tuple = aggregator.aggregate() else: __snake_case: List[Any] = {} for key in scores[0]: __snake_case: str = [score[key] for score in scores] return result
369
import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __snake_case ( __lowerCamelCase , unittest.TestCase ): '''simple docstring''' pass @nightly @require_onnxruntime @require_torch_gpu class __snake_case ( unittest.TestCase ): '''simple docstring''' @property def UpperCAmelCase__ ( self : Dict ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCAmelCase__ ( self : Dict ): __snake_case: Optional[int] = ort.SessionOptions() __snake_case: List[Any] = False return options def UpperCAmelCase__ ( self : Optional[Any] ): __snake_case: Optional[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) __snake_case: Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) __snake_case: List[str] = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , safety_checker=A , feature_extractor=A , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=A ) __snake_case: int = """A red cat sitting on a park bench""" __snake_case: Any = np.random.RandomState(0 ) __snake_case: Optional[Any] = pipe( prompt=A , image=A , mask_image=A , guidance_scale=7.5 , num_inference_steps=10 , generator=A , output_type="""np""" , ) __snake_case: List[Any] = output.images __snake_case: str = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) __snake_case: Any = np.array([0.2514, 0.3007, 0.3517, 0.1790, 0.2382, 0.3167, 0.1944, 0.2273, 0.2464] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def UpperCAmelCase__ ( self : Optional[Any] ): __snake_case: Optional[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) __snake_case: Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) __snake_case: Optional[int] = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-inpainting""" , subfolder="""scheduler""" , revision="""onnx""" ) __snake_case: List[Any] = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , scheduler=A , safety_checker=A , feature_extractor=A , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=A ) __snake_case: Optional[int] = """A red cat sitting on a park bench""" __snake_case: Dict = np.random.RandomState(0 ) __snake_case: Optional[Any] = pipe( prompt=A , image=A , mask_image=A , guidance_scale=7.5 , num_inference_steps=20 , generator=A , output_type="""np""" , ) __snake_case: List[str] = output.images __snake_case: str = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) __snake_case: Union[str, Any] = np.array([0.0086, 0.0077, 0.0083, 0.0093, 0.0107, 0.0139, 0.0094, 0.0097, 0.0125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
293
0
"""simple docstring""" def snake_case ( A__ = 1_00_00_00 ): UpperCAmelCase_ : List[str] = 1 UpperCAmelCase_ : Union[str, Any] = 1 UpperCAmelCase_ : Tuple = {1: 1} for inputa in range(2 ,A__ ): UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ : Dict = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: UpperCAmelCase_ : List[Any] = (3 * number) + 1 counter += 1 if inputa not in counters: UpperCAmelCase_ : List[Any] = counter if counter > pre_counter: UpperCAmelCase_ : List[str] = inputa UpperCAmelCase_ : Dict = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
268
"""simple docstring""" from torch import nn def snake_case ( A__ ): if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F"""Unsupported activation function: {act_fn}""" )
268
1
from pathlib import PurePosixPath from typing import Optional import fsspec from fsspec import AbstractFileSystem from huggingface_hub.hf_api import DatasetInfo from ..utils.file_utils import get_authentication_headers_for_url from ..utils.hub import hf_hub_url class _snake_case ( A__ ): _lowercase : Tuple = '''''' _lowercase : Optional[int] = '''hf-legacy''' # "hf://"" is reserved for hffs def __init__( self , a = None , a = None , **a , ) -> Any: super().__init__(self , **a) SCREAMING_SNAKE_CASE = repo_info SCREAMING_SNAKE_CASE = token SCREAMING_SNAKE_CASE = None def SCREAMING_SNAKE_CASE__ ( self) -> Dict: if self.dir_cache is None: SCREAMING_SNAKE_CASE = {} for hf_file in self.repo_info.siblings: # TODO(QL): add sizes SCREAMING_SNAKE_CASE = { 'name': hf_file.rfilename, 'size': None, 'type': 'file', } self.dir_cache.update( { str(a): {'name': str(a), 'size': None, 'type': 'directory'} for d in list(PurePosixPath(hf_file.rfilename).parents)[:-1] }) def SCREAMING_SNAKE_CASE__ ( self , a , a = "rb" , **a , ) -> Tuple: if not isinstance(self.repo_info , a): raise NotImplementedError(f'''Open is only implemented for dataset repositories, but got {self.repo_info}''') SCREAMING_SNAKE_CASE = hf_hub_url(self.repo_info.id , a , revision=self.repo_info.sha) return fsspec.open( a , mode=a , headers=get_authentication_headers_for_url(a , use_auth_token=self.token) , client_kwargs={'trust_env': True} , ).open() def SCREAMING_SNAKE_CASE__ ( self , a , **a) -> List[Any]: self._get_dirs() SCREAMING_SNAKE_CASE = self._strip_protocol(a) if path in self.dir_cache: return self.dir_cache[path] else: raise FileNotFoundError(a) def SCREAMING_SNAKE_CASE__ ( self , a , a=False , **a) -> List[str]: self._get_dirs() SCREAMING_SNAKE_CASE = PurePosixPath(path.strip('/')) SCREAMING_SNAKE_CASE = {} for p, f in self.dir_cache.items(): SCREAMING_SNAKE_CASE = PurePosixPath(p.strip('/')) SCREAMING_SNAKE_CASE = p.parent if root == path: SCREAMING_SNAKE_CASE = f SCREAMING_SNAKE_CASE = list(paths.values()) if detail: return out else: return sorted(f['name'] for f in out)
355
import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint a_ : Optional[int] = { '169M': 12, '430M': 24, '1B5': 24, '3B': 32, '7B': 32, '14B': 40, } a_ : Optional[int] = { '169M': 7_68, '430M': 10_24, '1B5': 20_48, '3B': 25_60, '7B': 40_96, '14B': 51_20, } def lowerCamelCase__ (_UpperCAmelCase): SCREAMING_SNAKE_CASE = list(state_dict.keys()) for name in state_dict_keys: SCREAMING_SNAKE_CASE = state_dict.pop(_UpperCAmelCase) # emb -> embedding if name.startswith('emb.'): SCREAMING_SNAKE_CASE = name.replace('emb.' , 'embeddings.') # ln_0 -> pre_ln (only present at block 0) if name.startswith('blocks.0.ln0'): SCREAMING_SNAKE_CASE = name.replace('blocks.0.ln0' , 'blocks.0.pre_ln') # att -> attention SCREAMING_SNAKE_CASE = re.sub(R'blocks\.(\d+)\.att' , R'blocks.\1.attention' , _UpperCAmelCase) # ffn -> feed_forward SCREAMING_SNAKE_CASE = re.sub(R'blocks\.(\d+)\.ffn' , R'blocks.\1.feed_forward' , _UpperCAmelCase) # time_mix_k -> time_mix_key and reshape if name.endswith('.time_mix_k'): SCREAMING_SNAKE_CASE = name.replace('.time_mix_k' , '.time_mix_key') # time_mix_v -> time_mix_value and reshape if name.endswith('.time_mix_v'): SCREAMING_SNAKE_CASE = name.replace('.time_mix_v' , '.time_mix_value') # time_mix_r -> time_mix_key and reshape if name.endswith('.time_mix_r'): SCREAMING_SNAKE_CASE = name.replace('.time_mix_r' , '.time_mix_receptance') if name != "head.weight": SCREAMING_SNAKE_CASE = 'rwkv.' + name SCREAMING_SNAKE_CASE = weight return state_dict def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=False , _UpperCAmelCase=None): # 1. If possible, build the tokenizer. if tokenizer_file is None: print('No `--tokenizer_file` provided, we will use the default tokenizer.') SCREAMING_SNAKE_CASE = 5_0277 SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained('EleutherAI/gpt-neox-20b') else: SCREAMING_SNAKE_CASE = PreTrainedTokenizerFast(tokenizer_file=_UpperCAmelCase) SCREAMING_SNAKE_CASE = len(_UpperCAmelCase) tokenizer.save_pretrained(_UpperCAmelCase) # 2. Build the config SCREAMING_SNAKE_CASE = list(NUM_HIDDEN_LAYERS_MAPPING.keys()) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: SCREAMING_SNAKE_CASE = candidate break if size is None: raise ValueError('Could not infer the size, please provide it with the `--size` argument.') if size not in possible_sizes: raise ValueError(F'''`size` should be one of {possible_sizes}, got {size}.''') SCREAMING_SNAKE_CASE = RwkvConfig( vocab_size=_UpperCAmelCase , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(_UpperCAmelCase) # 3. Download model file then convert state_dict SCREAMING_SNAKE_CASE = hf_hub_download(_UpperCAmelCase , _UpperCAmelCase) SCREAMING_SNAKE_CASE = torch.load(_UpperCAmelCase , map_location='cpu') SCREAMING_SNAKE_CASE = convert_state_dict(_UpperCAmelCase) # 4. Split in shards and save SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = shard_checkpoint(_UpperCAmelCase) for shard_file, shard in shards.items(): torch.save(_UpperCAmelCase , os.path.join(_UpperCAmelCase , _UpperCAmelCase)) if index is not None: SCREAMING_SNAKE_CASE = os.path.join(_UpperCAmelCase , _UpperCAmelCase) # Save the index as well with open(_UpperCAmelCase , 'w' , encoding='utf-8') as f: SCREAMING_SNAKE_CASE = json.dumps(_UpperCAmelCase , indent=2 , sort_keys=_UpperCAmelCase) + '\n' f.write(_UpperCAmelCase) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( 'Cleaning up shards. This may error with an OOM error, it this is the case don\'t worry you still have converted the model.') SCREAMING_SNAKE_CASE = list(shards.keys()) del state_dict del shards gc.collect() for shard_file in shard_files: SCREAMING_SNAKE_CASE = torch.load(os.path.join(_UpperCAmelCase , _UpperCAmelCase)) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(_UpperCAmelCase , _UpperCAmelCase)) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError('Please provide a `model_name` to push the model to the Hub.') SCREAMING_SNAKE_CASE = AutoModelForCausalLM.from_pretrained(_UpperCAmelCase) model.push_to_hub(_UpperCAmelCase , max_shard_size='2GB') tokenizer.push_to_hub(_UpperCAmelCase) if __name__ == "__main__": a_ : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '--repo_id', default=None, type=str, required=True, help='Repo ID from which to pull the checkpoint.' ) parser.add_argument( '--checkpoint_file', default=None, type=str, required=True, help='Name of the checkpoint file in the repo.' ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='Where to save the converted model.' ) parser.add_argument( '--tokenizer_file', default=None, type=str, help='Path to the tokenizer file to use (if not provided, only the model is converted).', ) parser.add_argument( '--size', default=None, type=str, help='Size of the model. Will be inferred from the `checkpoint_file` if not passed.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Push to the Hub the converted model.', ) parser.add_argument( '--model_name', default=None, type=str, help='Name of the pushed model on the Hub, including the username / organization.', ) a_ : Tuple = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
327
0
"""simple docstring""" import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): UpperCAmelCase_ : Optional[Any] = "pt" elif is_tf_available(): UpperCAmelCase_ : Tuple = "tf" else: UpperCAmelCase_ : Dict = "jax" class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase = ByTaTokenizer __UpperCamelCase = False def _SCREAMING_SNAKE_CASE ( self : Tuple): '''simple docstring''' super().setUp() SCREAMING_SNAKE_CASE_ : Union[str, Any] = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname) @cached_property def _SCREAMING_SNAKE_CASE ( self : List[Any]): '''simple docstring''' return ByTaTokenizer.from_pretrained('''google/byt5-small''') def _SCREAMING_SNAKE_CASE ( self : Tuple , **lowercase_ : Optional[int]): '''simple docstring''' return self.tokenizer_class.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowercase_ : List[str] , lowercase_ : Optional[Any]=False , lowercase_ : Dict=20 , lowercase_ : Union[str, Any]=5): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = [] for i in range(len(_SCREAMING_SNAKE_CASE)): try: SCREAMING_SNAKE_CASE_ : Optional[Any] = tokenizer.decode([i] , clean_up_tokenization_spaces=_SCREAMING_SNAKE_CASE) except UnicodeDecodeError: pass toks.append((i, tok)) SCREAMING_SNAKE_CASE_ : int = list(filter(lambda lowercase_: re.match(r'''^[ a-zA-Z]+$''' , t[1]) , _SCREAMING_SNAKE_CASE)) SCREAMING_SNAKE_CASE_ : Optional[int] = list(filter(lambda lowercase_: [t[0]] == tokenizer.encode(t[1] , add_special_tokens=_SCREAMING_SNAKE_CASE) , _SCREAMING_SNAKE_CASE)) if max_length is not None and len(_SCREAMING_SNAKE_CASE) > max_length: SCREAMING_SNAKE_CASE_ : Optional[Any] = toks[:max_length] if min_length is not None and len(_SCREAMING_SNAKE_CASE) < min_length and len(_SCREAMING_SNAKE_CASE) > 0: while len(_SCREAMING_SNAKE_CASE) < min_length: SCREAMING_SNAKE_CASE_ : Dict = toks + toks # toks_str = [t[1] for t in toks] SCREAMING_SNAKE_CASE_ : str = [t[0] for t in toks] # Ensure consistency SCREAMING_SNAKE_CASE_ : int = tokenizer.decode(_SCREAMING_SNAKE_CASE , clean_up_tokenization_spaces=_SCREAMING_SNAKE_CASE) if " " not in output_txt and len(_SCREAMING_SNAKE_CASE) > 1: SCREAMING_SNAKE_CASE_ : Optional[Any] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=_SCREAMING_SNAKE_CASE) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=_SCREAMING_SNAKE_CASE) ) if with_prefix_space: SCREAMING_SNAKE_CASE_ : Optional[int] = """ """ + output_txt SCREAMING_SNAKE_CASE_ : int = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) return output_txt, output_ids def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : List[Any] = tokenizer(['''hi</s>''', '''I went to the gym</s>''', '''</s>''']) SCREAMING_SNAKE_CASE_ : Any = tokenizer(['''hi''', '''I went to the gym''', '''''']) self.assertListEqual(batch_with_eos_added['''input_ids'''] , batch_without_eos_added['''input_ids''']) def _SCREAMING_SNAKE_CASE ( self : Any): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : Optional[Any] = """Unicode €.""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = tokenizer(_SCREAMING_SNAKE_CASE) SCREAMING_SNAKE_CASE_ : Union[str, Any] = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded['''input_ids'''] , _SCREAMING_SNAKE_CASE) # decoding SCREAMING_SNAKE_CASE_ : Tuple = tokenizer.decode(_SCREAMING_SNAKE_CASE) self.assertEqual(_SCREAMING_SNAKE_CASE , '''Unicode €.</s>''') SCREAMING_SNAKE_CASE_ : Optional[int] = tokenizer('''e è é ê ë''') SCREAMING_SNAKE_CASE_ : int = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded['''input_ids'''] , _SCREAMING_SNAKE_CASE) # decoding SCREAMING_SNAKE_CASE_ : List[str] = tokenizer.decode(_SCREAMING_SNAKE_CASE) self.assertEqual(_SCREAMING_SNAKE_CASE , '''e è é ê ë</s>''') # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''')) , '''e è é ê ë</s>''') def _SCREAMING_SNAKE_CASE ( self : Tuple): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[Any] = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : str = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] # fmt: off SCREAMING_SNAKE_CASE_ : int = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on SCREAMING_SNAKE_CASE_ : Optional[int] = tokenizer(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) if FRAMEWORK != "jax": SCREAMING_SNAKE_CASE_ : int = list(batch.input_ids.numpy()[0]) else: SCREAMING_SNAKE_CASE_ : str = list(batch.input_ids.tolist()[0]) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.assertEqual((2, 37) , batch.input_ids.shape) self.assertEqual((2, 37) , batch.attention_mask.shape) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : Optional[int] = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] SCREAMING_SNAKE_CASE_ : Dict = tokenizer(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , _SCREAMING_SNAKE_CASE) self.assertIn('''attention_mask''' , _SCREAMING_SNAKE_CASE) self.assertNotIn('''decoder_input_ids''' , _SCREAMING_SNAKE_CASE) self.assertNotIn('''decoder_attention_mask''' , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self : int): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : List[str] = [ """Summary of the text.""", """Another summary.""", ] SCREAMING_SNAKE_CASE_ : Tuple = tokenizer( text_target=_SCREAMING_SNAKE_CASE , max_length=32 , padding='''max_length''' , truncation=_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE) self.assertEqual(32 , targets['''input_ids'''].shape[1]) def _SCREAMING_SNAKE_CASE ( self : int): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ : str = ["""A long paragraph for summarization. </s>"""] SCREAMING_SNAKE_CASE_ : List[Any] = ["""Summary of the text. </s>"""] # fmt: off SCREAMING_SNAKE_CASE_ : Dict = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] SCREAMING_SNAKE_CASE_ : int = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on SCREAMING_SNAKE_CASE_ : Any = tokenizer(_SCREAMING_SNAKE_CASE , text_target=_SCREAMING_SNAKE_CASE) self.assertEqual(_SCREAMING_SNAKE_CASE , batch['''input_ids'''][0]) self.assertEqual(_SCREAMING_SNAKE_CASE , batch['''labels'''][0]) def _SCREAMING_SNAKE_CASE ( self : int): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): self.assertNotEqual(tokenizer.model_max_length , 42) # Now let's start the test SCREAMING_SNAKE_CASE_ : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ : Optional[int] = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ : List[str] = """ He is very happy, UNwant\u00E9d,running""" SCREAMING_SNAKE_CASE_ : List[str] = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) tokenizer.save_pretrained(_SCREAMING_SNAKE_CASE) SCREAMING_SNAKE_CASE_ : Tuple = tokenizer.__class__.from_pretrained(_SCREAMING_SNAKE_CASE) SCREAMING_SNAKE_CASE_ : Any = after_tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) shutil.rmtree(_SCREAMING_SNAKE_CASE) SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_tokenizers(model_max_length=42) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ : Any = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ : List[str] = """ He is very happy, UNwant\u00E9d,running""" tokenizer.add_tokens(['''bim''', '''bambam''']) SCREAMING_SNAKE_CASE_ : Dict = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''') tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens}) SCREAMING_SNAKE_CASE_ : List[str] = tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) tokenizer.save_pretrained(_SCREAMING_SNAKE_CASE) SCREAMING_SNAKE_CASE_ : Dict = tokenizer.__class__.from_pretrained(_SCREAMING_SNAKE_CASE) SCREAMING_SNAKE_CASE_ : str = after_tokenizer.encode(_SCREAMING_SNAKE_CASE , add_special_tokens=_SCREAMING_SNAKE_CASE) self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens) self.assertEqual(after_tokenizer.model_max_length , 42) SCREAMING_SNAKE_CASE_ : Optional[int] = tokenizer.__class__.from_pretrained(_SCREAMING_SNAKE_CASE , model_max_length=43) self.assertEqual(tokenizer.model_max_length , 43) shutil.rmtree(_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self : Optional[int]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer())) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer())) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_SCREAMING_SNAKE_CASE) with open(os.path.join(_SCREAMING_SNAKE_CASE , '''special_tokens_map.json''') , encoding='''utf-8''') as json_file: SCREAMING_SNAKE_CASE_ : int = json.load(_SCREAMING_SNAKE_CASE) with open(os.path.join(_SCREAMING_SNAKE_CASE , '''tokenizer_config.json''') , encoding='''utf-8''') as json_file: SCREAMING_SNAKE_CASE_ : List[Any] = json.load(_SCREAMING_SNAKE_CASE) SCREAMING_SNAKE_CASE_ : List[Any] = [F'<extra_id_{i}>' for i in range(125)] SCREAMING_SNAKE_CASE_ : Optional[Any] = added_tokens_extra_ids + [ """an_additional_special_token""" ] SCREAMING_SNAKE_CASE_ : List[str] = added_tokens_extra_ids + [ """an_additional_special_token""" ] with open(os.path.join(_SCREAMING_SNAKE_CASE , '''special_tokens_map.json''') , '''w''' , encoding='''utf-8''') as outfile: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) with open(os.path.join(_SCREAMING_SNAKE_CASE , '''tokenizer_config.json''') , '''w''' , encoding='''utf-8''') as outfile: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files SCREAMING_SNAKE_CASE_ : List[str] = tokenizer_class.from_pretrained( _SCREAMING_SNAKE_CASE , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''])) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained SCREAMING_SNAKE_CASE_ : List[str] = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=_SCREAMING_SNAKE_CASE)] SCREAMING_SNAKE_CASE_ : List[str] = tokenizer_class.from_pretrained( _SCREAMING_SNAKE_CASE , additional_special_tokens=_SCREAMING_SNAKE_CASE , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''])) , ) def _SCREAMING_SNAKE_CASE ( self : Optional[int]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer())) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer())) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_SCREAMING_SNAKE_CASE) SCREAMING_SNAKE_CASE_ : Optional[int] = tokenizer_class.from_pretrained(_SCREAMING_SNAKE_CASE) self.assertTrue(tokenizer.decode([255]) == '''''') def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): '''simple docstring''' pass def _SCREAMING_SNAKE_CASE ( self : str): '''simple docstring''' pass def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): '''simple docstring''' pass def _SCREAMING_SNAKE_CASE ( self : Optional[int]): '''simple docstring''' pass def _SCREAMING_SNAKE_CASE ( self : str): '''simple docstring''' SCREAMING_SNAKE_CASE_ : int = self.get_tokenizers(fast=_SCREAMING_SNAKE_CASE , do_lower_case=_SCREAMING_SNAKE_CASE) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): SCREAMING_SNAKE_CASE_ : Union[str, Any] = ["""t""", """h""", """i""", """s""", """ """, """i""", """s""", """ """, """a""", """ """, """t""", """e""", """x""", """t""", """</s>"""] SCREAMING_SNAKE_CASE_ : Union[str, Any] = tokenizer.convert_tokens_to_string(_SCREAMING_SNAKE_CASE) self.assertIsInstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : int = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): SCREAMING_SNAKE_CASE_ : Dict = [ """bos_token""", """eos_token""", """unk_token""", """sep_token""", """pad_token""", """cls_token""", """mask_token""", ] SCREAMING_SNAKE_CASE_ : Dict = 0 SCREAMING_SNAKE_CASE_ : Union[str, Any] = tokenizer.convert_ids_to_tokens( _SCREAMING_SNAKE_CASE , skip_special_tokens=_SCREAMING_SNAKE_CASE) for attr in attributes_list: setattr(_SCREAMING_SNAKE_CASE , attr + '''_id''' , _SCREAMING_SNAKE_CASE) self.assertEqual(getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) , _SCREAMING_SNAKE_CASE) self.assertEqual(getattr(_SCREAMING_SNAKE_CASE , attr + '''_id''') , _SCREAMING_SNAKE_CASE) setattr(_SCREAMING_SNAKE_CASE , attr + '''_id''' , _SCREAMING_SNAKE_CASE) self.assertEqual(getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) , _SCREAMING_SNAKE_CASE) self.assertEqual(getattr(_SCREAMING_SNAKE_CASE , attr + '''_id''') , _SCREAMING_SNAKE_CASE) setattr(_SCREAMING_SNAKE_CASE , '''additional_special_tokens_ids''' , []) self.assertListEqual(getattr(_SCREAMING_SNAKE_CASE , '''additional_special_tokens''') , []) self.assertListEqual(getattr(_SCREAMING_SNAKE_CASE , '''additional_special_tokens_ids''') , []) setattr(_SCREAMING_SNAKE_CASE , '''additional_special_tokens_ids''' , [token_id_to_test_setters]) self.assertListEqual(getattr(_SCREAMING_SNAKE_CASE , '''additional_special_tokens''') , [token_to_test_setters]) self.assertListEqual(getattr(_SCREAMING_SNAKE_CASE , '''additional_special_tokens_ids''') , [token_id_to_test_setters])
91
"""simple docstring""" import datasets from .evaluate import evaluate A: Optional[Any] = "\\n@article{hendrycks2021cuad,\n title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review},\n author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball},\n journal={arXiv preprint arXiv:2103.06268},\n year={2021}\n}\n" A: Optional[int] = "\nThis metric wrap the official scoring script for version 1 of the Contract\nUnderstanding Atticus Dataset (CUAD).\nContract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510\ncommercial legal contracts that have been manually labeled to identify 41 categories of important\nclauses that lawyers look for when reviewing contracts in connection with corporate transactions.\n" A: int = "\nComputes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair as given in the references (see below)\n - 'prediction_text': list of possible texts for the answer, as a list of strings\n depending on a threshold on the confidence probability of each prediction.\n references: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair (see above),\n - 'answers': a Dict in the CUAD dataset format\n {\n 'text': list of possible texts for the answer, as a list of strings\n 'answer_start': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n 'exact_match': Exact match (the normalized answer exactly match the gold answer)\n 'f1': The F-score of predicted tokens versus the gold answer\n 'aupr': Area Under the Precision-Recall curve\n 'prec_at_80_recall': Precision at 80% recall\n 'prec_at_90_recall': Precision at 90% recall\nExamples:\n >>> predictions = [{'prediction_text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.'], 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}]\n >>> references = [{'answers': {'answer_start': [143, 49], 'text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.']}, 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}]\n >>> cuad_metric = datasets.load_metric(\"cuad\")\n >>> results = cuad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 100.0, 'f1': 100.0, 'aupr': 0.0, 'prec_at_80_recall': 1.0, 'prec_at_90_recall': 1.0}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": { """id""": datasets.Value("""string""" ), """prediction_text""": datasets.features.Sequence(datasets.Value("""string""" ) ), }, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) , codebase_urls=["""https://www.atticusprojectai.org/cuad"""] , reference_urls=["""https://www.atticusprojectai.org/cuad"""] , ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Dict: '''simple docstring''' UpperCAmelCase : int = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} UpperCAmelCase : Tuple = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] UpperCAmelCase : Optional[Any] = evaluate(dataset=_SCREAMING_SNAKE_CASE , predictions=_SCREAMING_SNAKE_CASE ) return score
109
0
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class a ( __lowerCamelCase ): def __lowerCamelCase ( self :Tuple ): snake_case__ : int = tempfile.mkdtemp() snake_case__ : Union[str, Any] = 8 # DPR tok snake_case__ : Union[str, Any] = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] snake_case__ : str = os.path.join(self.tmpdirname ,'''dpr_tokenizer''' ) os.makedirs(__lowercase ,exist_ok=__lowercase ) snake_case__ : Optional[int] = os.path.join(__lowercase ,DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file ,'''w''' ,encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok snake_case__ : List[str] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] snake_case__ : Optional[Any] = dict(zip(__lowercase ,range(len(__lowercase ) ) ) ) snake_case__ : List[Any] = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] snake_case__ : int = {'''unk_token''': '''<unk>'''} snake_case__ : Optional[int] = os.path.join(self.tmpdirname ,'''bart_tokenizer''' ) os.makedirs(__lowercase ,exist_ok=__lowercase ) snake_case__ : Union[str, Any] = os.path.join(__lowercase ,BART_VOCAB_FILES_NAMES['''vocab_file'''] ) snake_case__ : Any = os.path.join(__lowercase ,BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file ,'''w''' ,encoding='''utf-8''' ) as fp: fp.write(json.dumps(__lowercase ) + '''\n''' ) with open(self.merges_file ,'''w''' ,encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__lowercase ) ) def __lowerCamelCase ( self :Dict ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname ,'''dpr_tokenizer''' ) ) def __lowerCamelCase ( self :int ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname ,'''dpr_tokenizer''' ) ) def __lowerCamelCase ( self :Tuple ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname ,'''bart_tokenizer''' ) ) def __lowerCamelCase ( self :Union[str, Any] ): shutil.rmtree(self.tmpdirname ) def __lowerCamelCase ( self :Optional[Any] ): snake_case__ : int = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' ,string_factory='''Flat''' ,metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def __lowerCamelCase ( self :str ): snake_case__ : Dict = self.get_dummy_dataset() snake_case__ : Union[str, Any] = RagConfig( retrieval_vector_size=self.retrieval_vector_size ,question_encoder=DPRConfig().to_dict() ,generator=BartConfig().to_dict() ,) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: snake_case__ : Union[str, Any] = dataset snake_case__ : str = RagRetriever( __lowercase ,question_encoder_tokenizer=self.get_dpr_tokenizer() ,generator_tokenizer=self.get_bart_tokenizer() ,) return retriever def __lowerCamelCase ( self :Dict ,__lowercase :bool ): snake_case__ : Optional[Any] = self.get_dummy_dataset() snake_case__ : int = RagConfig( retrieval_vector_size=self.retrieval_vector_size ,question_encoder=DPRConfig().to_dict() ,generator=BartConfig().to_dict() ,index_name='''custom''' ,) if from_disk: snake_case__ : List[Any] = os.path.join(self.tmpdirname ,'''dataset''' ) snake_case__ : Dict = os.path.join(self.tmpdirname ,'''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname ,'''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname ,'''dataset''' ) ) del dataset snake_case__ : List[Any] = RagRetriever( __lowercase ,question_encoder_tokenizer=self.get_dpr_tokenizer() ,generator_tokenizer=self.get_bart_tokenizer() ,) else: snake_case__ : Dict = RagRetriever( __lowercase ,question_encoder_tokenizer=self.get_dpr_tokenizer() ,generator_tokenizer=self.get_bart_tokenizer() ,index=CustomHFIndex(config.retrieval_vector_size ,__lowercase ) ,) return retriever def __lowerCamelCase ( self :int ): snake_case__ : Optional[int] = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' ,string_factory='''Flat''' ,metric_type=faiss.METRIC_INNER_PRODUCT ) snake_case__ : Dict = os.path.join(self.tmpdirname ,'''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' ,index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] ,open(index_file_name + '''.index_meta.dpr''' ,'''wb''' ) ) snake_case__ : Any = os.path.join(self.tmpdirname ,'''psgs_w100.tsv.pkl''' ) snake_case__ : Any = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(__lowercase ,open(__lowercase ,'''wb''' ) ) snake_case__ : Union[str, Any] = RagConfig( retrieval_vector_size=self.retrieval_vector_size ,question_encoder=DPRConfig().to_dict() ,generator=BartConfig().to_dict() ,index_name='''legacy''' ,index_path=self.tmpdirname ,) snake_case__ : Any = RagRetriever( __lowercase ,question_encoder_tokenizer=self.get_dpr_tokenizer() ,generator_tokenizer=self.get_bart_tokenizer() ) return retriever def __lowerCamelCase ( self :Any ): snake_case__ : List[str] = 1 snake_case__ : Optional[Any] = self.get_dummy_canonical_hf_index_retriever() snake_case__ : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ , snake_case__ , snake_case__ : Optional[int] = retriever.retrieve(__lowercase ,n_docs=__lowercase ) self.assertEqual(retrieved_doc_embeds.shape ,(2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__lowercase ) ,2 ) self.assertEqual(sorted(doc_dicts[0] ) ,['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) ,__lowercase ) self.assertEqual(doc_dicts[0]['''id'''][0] ,'''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] ,'''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() ,[[1], [0]] ) def __lowerCamelCase ( self :Optional[int] ): snake_case__ : Any = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: snake_case__ : Union[str, Any] = self.get_dummy_dataset() retriever.save_pretrained(__lowercase ) snake_case__ : int = RagRetriever.from_pretrained(__lowercase ) self.assertIsInstance(__lowercase ,__lowercase ) snake_case__ : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ : Optional[int] = retriever.retrieve(__lowercase ,n_docs=1 ) self.assertTrue(out is not None ) def __lowerCamelCase ( self :Optional[int] ): snake_case__ : List[Any] = 1 snake_case__ : Union[str, Any] = self.get_dummy_custom_hf_index_retriever(from_disk=__lowercase ) snake_case__ : Union[str, Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ , snake_case__ , snake_case__ : List[str] = retriever.retrieve(__lowercase ,n_docs=__lowercase ) self.assertEqual(retrieved_doc_embeds.shape ,(2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__lowercase ) ,2 ) self.assertEqual(sorted(doc_dicts[0] ) ,['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) ,__lowercase ) self.assertEqual(doc_dicts[0]['''id'''][0] ,'''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] ,'''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() ,[[1], [0]] ) def __lowerCamelCase ( self :Any ): snake_case__ : Optional[Any] = self.get_dummy_custom_hf_index_retriever(from_disk=__lowercase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__lowercase ) snake_case__ : List[Any] = RagRetriever.from_pretrained(__lowercase ) self.assertIsInstance(__lowercase ,__lowercase ) snake_case__ : int = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ : str = retriever.retrieve(__lowercase ,n_docs=1 ) self.assertTrue(out is not None ) def __lowerCamelCase ( self :List[Any] ): snake_case__ : Optional[Any] = 1 snake_case__ : Optional[Any] = self.get_dummy_custom_hf_index_retriever(from_disk=__lowercase ) snake_case__ : Tuple = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ , snake_case__ , snake_case__ : List[str] = retriever.retrieve(__lowercase ,n_docs=__lowercase ) self.assertEqual(retrieved_doc_embeds.shape ,(2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__lowercase ) ,2 ) self.assertEqual(sorted(doc_dicts[0] ) ,['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) ,__lowercase ) self.assertEqual(doc_dicts[0]['''id'''][0] ,'''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] ,'''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() ,[[1], [0]] ) def __lowerCamelCase ( self :Any ): snake_case__ : Dict = self.get_dummy_custom_hf_index_retriever(from_disk=__lowercase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__lowercase ) snake_case__ : Any = RagRetriever.from_pretrained(__lowercase ) self.assertIsInstance(__lowercase ,__lowercase ) snake_case__ : Dict = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ : List[Any] = retriever.retrieve(__lowercase ,n_docs=1 ) self.assertTrue(out is not None ) def __lowerCamelCase ( self :Any ): snake_case__ : Tuple = 1 snake_case__ : Dict = self.get_dummy_legacy_index_retriever() snake_case__ : Union[str, Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ , snake_case__ , snake_case__ : int = retriever.retrieve(__lowercase ,n_docs=__lowercase ) self.assertEqual(retrieved_doc_embeds.shape ,(2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__lowercase ) ,2 ) self.assertEqual(sorted(doc_dicts[0] ) ,['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) ,__lowercase ) self.assertEqual(doc_dicts[0]['''text'''][0] ,'''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] ,'''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() ,[[1], [0]] ) def __lowerCamelCase ( self :Tuple ): snake_case__ : Optional[Any] = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__lowercase ) snake_case__ : Tuple = RagRetriever.from_pretrained(__lowercase ) self.assertIsInstance(__lowercase ,__lowercase ) snake_case__ : Any = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ : str = retriever.retrieve(__lowercase ,n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def __lowerCamelCase ( self :str ): import torch snake_case__ : Any = 1 snake_case__ : Any = self.get_dummy_canonical_hf_index_retriever() snake_case__ : List[Any] = [[5, 7], [1_0, 1_1]] snake_case__ : Union[str, Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ : Tuple = retriever(__lowercase ,__lowercase ,prefix=retriever.config.generator.prefix ,n_docs=__lowercase ) snake_case__ , snake_case__ , snake_case__ : str = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape ,(2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__lowercase ,__lowercase ) self.assertIsInstance(__lowercase ,__lowercase ) self.assertIsInstance(__lowercase ,np.ndarray ) snake_case__ : Union[str, Any] = retriever( __lowercase ,__lowercase ,prefix=retriever.config.generator.prefix ,n_docs=__lowercase ,return_tensors='''pt''' ,) snake_case__ , snake_case__ , snake_case__ , snake_case__ : Dict = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape ,(2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__lowercase ,torch.Tensor ) self.assertIsInstance(__lowercase ,torch.Tensor ) self.assertIsInstance(__lowercase ,torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def __lowerCamelCase ( self :List[Any] ): snake_case__ : List[Any] = self.get_dpr_ctx_encoder_tokenizer() snake_case__ : List[Any] = 1 snake_case__ : Any = self.get_dummy_custom_hf_index_retriever(from_disk=__lowercase ) retriever.set_ctx_encoder_tokenizer(__lowercase ) snake_case__ : List[str] = [[5, 7], [1_0, 1_1]] snake_case__ : Any = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] ,dtype=np.floataa ) snake_case__ : Tuple = retriever(__lowercase ,__lowercase ,prefix=retriever.config.generator.prefix ,n_docs=__lowercase ) self.assertEqual( len(__lowercase ) ,6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) ,__lowercase ) # check for doc token related keys in dictionary.
44
def _lowerCAmelCase ( __lowerCAmelCase ) -> list: """simple docstring""" for i in range(len(__lowerCAmelCase ) - 1 , 0 , -1 ): snake_case__ : List[Any] = False for j in range(__lowerCAmelCase , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: snake_case__ , snake_case__ : Optional[int] = unsorted[j - 1], unsorted[j] snake_case__ : Any = True for j in range(__lowerCAmelCase ): if unsorted[j] > unsorted[j + 1]: snake_case__ , snake_case__ : Tuple = unsorted[j + 1], unsorted[j] snake_case__ : int = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() A__ = input('''Enter numbers separated by a comma:\n''').strip() A__ = [int(item) for item in user_input.split(''',''')] print(f"""{cocktail_shaker_sort(unsorted) = }""")
44
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_lxmert import LxmertTokenizer __lowerCAmelCase : Optional[int] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} __lowerCAmelCase : List[Any] = { 'vocab_file': { 'unc-nlp/lxmert-base-uncased': 'https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/vocab.txt', }, 'tokenizer_file': { 'unc-nlp/lxmert-base-uncased': ( 'https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/tokenizer.json' ), }, } __lowerCAmelCase : Tuple = { 'unc-nlp/lxmert-base-uncased': 512, } __lowerCAmelCase : int = { 'unc-nlp/lxmert-base-uncased': {'do_lower_case': True}, } class snake_case__ (snake_case_ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : int = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : Any = PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE_ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : Optional[Any] = LxmertTokenizer def __init__( self : Optional[Any] , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : List[Any]=None , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="[UNK]" , __lowerCamelCase : Tuple="[SEP]" , __lowerCamelCase : Tuple="[PAD]" , __lowerCamelCase : List[Any]="[CLS]" , __lowerCamelCase : Union[str, Any]="[MASK]" , __lowerCamelCase : List[Any]=True , __lowerCamelCase : List[Any]=None , **__lowerCamelCase : Any , ) -> Tuple: super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , do_lower_case=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , tokenize_chinese_chars=UpperCamelCase__ , strip_accents=UpperCamelCase__ , **UpperCamelCase__ , ) a = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , UpperCamelCase__ ) != do_lower_case or normalizer_state.get("strip_accents" , UpperCamelCase__ ) != strip_accents or normalizer_state.get("handle_chinese_chars" , UpperCamelCase__ ) != tokenize_chinese_chars ): a = getattr(UpperCamelCase__ , normalizer_state.pop("type" ) ) a = do_lower_case a = strip_accents a = tokenize_chinese_chars a = normalizer_class(**UpperCamelCase__ ) a = do_lower_case def __UpperCAmelCase ( self : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : Dict=None ) -> int: a = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCAmelCase ( self : str , __lowerCamelCase : List[Any] , __lowerCamelCase : List[str] = None ) -> List[int]: a = [self.sep_token_id] a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCAmelCase ( self : Any , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[str] = None ) -> Tuple[str]: a = self._tokenizer.model.save(UpperCamelCase__ , name=UpperCamelCase__ ) return tuple(UpperCamelCase__ )
107
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { """facebook/xmod-base""": """https://huggingface.co/facebook/xmod-base/resolve/main/config.json""", """facebook/xmod-large-prenorm""": """https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json""", """facebook/xmod-base-13-125k""": """https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json""", """facebook/xmod-base-30-125k""": """https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json""", """facebook/xmod-base-30-195k""": """https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json""", """facebook/xmod-base-60-125k""": """https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json""", """facebook/xmod-base-60-265k""": """https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json""", """facebook/xmod-base-75-125k""": """https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json""", """facebook/xmod-base-75-269k""": """https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json""", } class _lowerCAmelCase ( snake_case_ ): __UpperCAmelCase : Optional[Any] = '''xmod''' def __init__( self , UpperCamelCase__=3_0522 , UpperCamelCase__=768 , UpperCamelCase__=12 , UpperCamelCase__=12 , UpperCamelCase__=3072 , UpperCamelCase__="gelu" , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=512 , UpperCamelCase__=2 , UpperCamelCase__=0.02 , UpperCamelCase__=1e-12 , UpperCamelCase__=1 , UpperCamelCase__=0 , UpperCamelCase__=2 , UpperCamelCase__="absolute" , UpperCamelCase__=True , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=2 , UpperCamelCase__=False , UpperCamelCase__=True , UpperCamelCase__=True , UpperCamelCase__=("en_XX",) , UpperCamelCase__=None , **UpperCamelCase__ , ) -> int: '''simple docstring''' super().__init__(pad_token_id=UpperCamelCase__ , bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ ) snake_case : List[Any] = vocab_size snake_case : List[Any] = hidden_size snake_case : Union[str, Any] = num_hidden_layers snake_case : Any = num_attention_heads snake_case : List[str] = hidden_act snake_case : Union[str, Any] = intermediate_size snake_case : int = hidden_dropout_prob snake_case : Any = attention_probs_dropout_prob snake_case : Optional[int] = max_position_embeddings snake_case : Tuple = type_vocab_size snake_case : List[str] = initializer_range snake_case : int = layer_norm_eps snake_case : Optional[Any] = position_embedding_type snake_case : int = use_cache snake_case : Dict = classifier_dropout snake_case : Dict = pre_norm snake_case : Union[str, Any] = adapter_reduction_factor snake_case : Any = adapter_layer_norm snake_case : Optional[int] = adapter_reuse_layer_norm snake_case : List[Any] = ln_before_adapter snake_case : str = list(UpperCamelCase__ ) snake_case : int = default_language class _lowerCAmelCase ( snake_case_ ): @property def lowerCamelCase ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": snake_case : Optional[int] = {0: "batch", 1: "choice", 2: "sequence"} else: snake_case : Optional[Any] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
203
0
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to properly calculate the metrics on the # validation dataset when in a distributed system, and builds off the # `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## lowerCAmelCase__ = 16 lowerCAmelCase__ = 32 def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 1_6 ): lowerCAmelCase__ = AutoTokenizer.from_pretrained('bert-base-cased' ) lowerCAmelCase__ = load_dataset('glue' , 'mrpc' ) def tokenize_function(lowerCAmelCase__ ): # max_length=None => use the model max length (it's actually the default) lowerCAmelCase__ = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowerCAmelCase__ = datasets.map( lowerCAmelCase__ , batched=lowerCAmelCase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowerCAmelCase__ = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowerCAmelCase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. lowerCAmelCase__ = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowerCAmelCase__ = 1_6 elif accelerator.mixed_precision != "no": lowerCAmelCase__ = 8 else: lowerCAmelCase__ = None return tokenizer.pad( lowerCAmelCase__ , padding='longest' , max_length=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_tensors='pt' , ) # Instantiate dataloaders. lowerCAmelCase__ = DataLoader( tokenized_datasets['train'] , shuffle=lowerCAmelCase__ , collate_fn=lowerCAmelCase__ , batch_size=lowerCAmelCase__ ) lowerCAmelCase__ = DataLoader( tokenized_datasets['validation'] , shuffle=lowerCAmelCase__ , collate_fn=lowerCAmelCase__ , batch_size=lowerCAmelCase__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders lowerCAmelCase__ = mocked_dataloaders # noqa: F811 def __lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS' , lowerCAmelCase__ ) == "1": lowerCAmelCase__ = 2 # Initialize accelerator lowerCAmelCase__ = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowerCAmelCase__ = config['lr'] lowerCAmelCase__ = int(config['num_epochs'] ) lowerCAmelCase__ = int(config['seed'] ) lowerCAmelCase__ = int(config['batch_size'] ) lowerCAmelCase__ = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation lowerCAmelCase__ = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: lowerCAmelCase__ = batch_size // MAX_GPU_BATCH_SIZE lowerCAmelCase__ = MAX_GPU_BATCH_SIZE set_seed(lowerCAmelCase__ ) lowerCAmelCase__ , lowerCAmelCase__ = get_dataloaders(lowerCAmelCase__ , lowerCAmelCase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowerCAmelCase__ = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=lowerCAmelCase__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowerCAmelCase__ = model.to(accelerator.device ) # Instantiate optimizer lowerCAmelCase__ = AdamW(params=model.parameters() , lr=lowerCAmelCase__ ) # Instantiate scheduler lowerCAmelCase__ = get_linear_schedule_with_warmup( optimizer=lowerCAmelCase__ , num_warmup_steps=1_0_0 , num_training_steps=(len(lowerCAmelCase__ ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = accelerator.prepare( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Now we train the model for epoch in range(lowerCAmelCase__ ): model.train() for step, batch in enumerate(lowerCAmelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowerCAmelCase__ = model(**lowerCAmelCase__ ) lowerCAmelCase__ = outputs.loss lowerCAmelCase__ = loss / gradient_accumulation_steps accelerator.backward(lowerCAmelCase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() lowerCAmelCase__ = 0 for step, batch in enumerate(lowerCAmelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowerCAmelCase__ = model(**lowerCAmelCase__ ) lowerCAmelCase__ = outputs.logits.argmax(dim=-1 ) lowerCAmelCase__ , lowerCAmelCase__ = accelerator.gather((predictions, batch['labels']) ) # New Code # # First we check if it's a distributed system if accelerator.use_distributed: # Then see if we're on the last batch of our eval dataloader if step == len(lowerCAmelCase__ ) - 1: # Last batch needs to be truncated on distributed systems as it contains additional samples lowerCAmelCase__ = predictions[: len(eval_dataloader.dataset ) - samples_seen] lowerCAmelCase__ = references[: len(eval_dataloader.dataset ) - samples_seen] else: # Otherwise we add the number of samples seen samples_seen += references.shape[0] # All of this can be avoided if you use `Accelerator.gather_for_metrics` instead of `Accelerator.gather`: # accelerator.gather_for_metrics((predictions, batch["labels"])) metric.add_batch( predictions=lowerCAmelCase__ , references=lowerCAmelCase__ , ) lowerCAmelCase__ = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , lowerCAmelCase__ ) def __lowerCamelCase ( ): lowerCAmelCase__ = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = {'lr': 2e-5, 'num_epochs': 3, 'seed': 4_2, 'batch_size': 1_6} training_function(lowerCAmelCase__ , lowerCAmelCase__ ) if __name__ == "__main__": main()
119
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class a_ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : str , lowercase__ : pyspark.sql.DataFrame , lowercase__ : Optional[NamedSplit] = None , lowercase__ : Optional[Features] = None , lowercase__ : bool = True , lowercase__ : str = None , lowercase__ : bool = False , lowercase__ : str = None , lowercase__ : bool = True , lowercase__ : str = "arrow" , **lowercase__ : Any , ): '''simple docstring''' super().__init__( split=lowercase__ , features=lowercase__ , cache_dir=lowercase__ , keep_in_memory=lowercase__ , streaming=lowercase__ , **lowercase__ , ) lowerCAmelCase__ = load_from_cache_file lowerCAmelCase__ = file_format lowerCAmelCase__ = Spark( df=lowercase__ , features=lowercase__ , cache_dir=lowercase__ , working_dir=lowercase__ , **lowercase__ , ) def __snake_case ( self : Tuple): '''simple docstring''' if self.streaming: return self.builder.as_streaming_dataset(split=self.split) lowerCAmelCase__ = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=lowercase__ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split)
119
1
from __future__ import annotations _a = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def lowerCAmelCase__(__snake_case ,__snake_case ,__snake_case ,__snake_case ,__snake_case ,) -> tuple[list[list[int]], list[list[int]]]: '''simple docstring''' lowerCamelCase__ = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__snake_case ) ) ] # the reference grid lowerCamelCase__ = 1 lowerCamelCase__ = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__snake_case ) ) ] # the action grid lowerCamelCase__ = init[0] lowerCamelCase__ = init[1] lowerCamelCase__ = 0 lowerCamelCase__ = g + heuristic[x][y] # cost from starting cell to destination cell lowerCamelCase__ = [[f, g, x, y]] lowerCamelCase__ = False # flag that is set when search is complete lowerCamelCase__ = False # flag set if we can't find expand while not found and not resign: if len(__snake_case ) == 0: raise ValueError('''Algorithm is unable to find solution''' ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() lowerCamelCase__ = cell.pop() lowerCamelCase__ = next_cell[2] lowerCamelCase__ = next_cell[3] lowerCamelCase__ = next_cell[1] if x == goal[0] and y == goal[1]: lowerCamelCase__ = True else: for i in range(len(__snake_case ) ): # to try out different valid actions lowerCamelCase__ = x + DIRECTIONS[i][0] lowerCamelCase__ = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(__snake_case ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: lowerCamelCase__ = g + cost lowerCamelCase__ = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) lowerCamelCase__ = 1 lowerCamelCase__ = i lowerCamelCase__ = [] lowerCamelCase__ = goal[0] lowerCamelCase__ = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: lowerCamelCase__ = x - DIRECTIONS[action[x][y]][0] lowerCamelCase__ = y - DIRECTIONS[action[x][y]][1] lowerCamelCase__ = xa lowerCamelCase__ = ya invpath.append([x, y] ) lowerCamelCase__ = [] for i in range(len(__snake_case ) ): path.append(invpath[len(__snake_case ) - 1 - i] ) return path, action if __name__ == "__main__": _a = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] _a = [0, 0] # all coordinates are given in format [y,x] _a = [len(grid) - 1, len(grid[0]) - 1] _a = 1 # the cost map which pushes the path closer to the goal _a = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): _a = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map _a = 99 _a , _a = search(grid, init, goal, cost, heuristic) print("ACTION MAP") for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
209
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _a = logging.get_logger(__name__) _a = {"vocab_file": "spiece.model"} _a = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } _a = { "AI-Sweden/gpt-sw3-126m": 2_048, "AI-Sweden/gpt-sw3-350m": 2_048, "AI-Sweden/gpt-sw3-1.6b": 2_048, "AI-Sweden/gpt-sw3-6.7b": 2_048, "AI-Sweden/gpt-sw3-20b": 2_048, } class __A ( lowerCAmelCase ): '''simple docstring''' lowerCAmelCase_ = VOCAB_FILES_NAMES lowerCAmelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ = ["""input_ids""", """attention_mask"""] def __init__( self , __lowerCAmelCase , __lowerCAmelCase=False , __lowerCAmelCase=False , __lowerCAmelCase=False , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase = None , **__lowerCAmelCase , ): '''simple docstring''' lowerCamelCase__ = {} if sp_model_kwargs is None else sp_model_kwargs lowerCamelCase__ = kwargs.get('''name_or_path''' ) if name_or_path is None: logger.warning( '''name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,''' ''' you are testing the model, this can safely be ignored''' ) lowerCamelCase__ = '''None''' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing lowerCamelCase__ = '''<|endoftext|>''' if eos_token is None else eos_token lowerCamelCase__ = '''<unk>''' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: lowerCamelCase__ = unk_token if pad_token is None else pad_token lowerCamelCase__ = eos_token if bos_token is None else bos_token else: lowerCamelCase__ = '''<pad>''' if pad_token is None else pad_token lowerCamelCase__ = '''<s>''' if bos_token is None else bos_token super().__init__( do_lower_case=__lowerCAmelCase , remove_space=__lowerCAmelCase , keep_accents=__lowerCAmelCase , bos_token=__lowerCAmelCase , eos_token=__lowerCAmelCase , unk_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__lowerCAmelCase , ) lowerCamelCase__ = do_lower_case lowerCamelCase__ = remove_space lowerCamelCase__ = keep_accents lowerCamelCase__ = vocab_file lowerCamelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__lowerCAmelCase ) # Used for whitespace normalization in input texts # fmt : off lowerCamelCase__ = {''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', '''''', '''„'''} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing lowerCamelCase__ = re.compile( F'[{"".join(map(__lowerCAmelCase , list(range(0 , 9 ) ) + list(range(1_1 , 3_2 ) ) + list(range(1_2_7 , 1_6_0 ) ) + [1_6_0, 1_7_3, 8_2_0_3] ) )}]' ) def __getstate__( self ): '''simple docstring''' lowerCamelCase__ = self.__dict__.copy() lowerCamelCase__ = None return state def __setstate__( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowerCamelCase__ = {} lowerCamelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def __lowerCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = self.non_printing_characters_re.sub('''''' , __lowerCAmelCase ) # Normalize whitespaces lowerCamelCase__ = ''''''.join([char if char not in self.whitespaces else ''' ''' for char in text] ) # NFC Unicode normalization lowerCamelCase__ = unicodedata.normalize('''NFC''' , __lowerCAmelCase ) return text def __lowerCamelCase ( self , __lowerCAmelCase , **__lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = self.preprocess_text(__lowerCAmelCase ) return self.sp_model.encode(__lowerCAmelCase , out_type=__lowerCAmelCase ) def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' return self.sp_model.PieceToId(__lowerCAmelCase ) def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' return self.sp_model.IdToPiece(__lowerCAmelCase ) @staticmethod def __lowerCamelCase ( __lowerCAmelCase ): '''simple docstring''' return out_string def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = [] lowerCamelCase__ = '''''' lowerCamelCase__ = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__lowerCAmelCase ) + token lowerCamelCase__ = True lowerCamelCase__ = [] else: current_sub_tokens.append(__lowerCAmelCase ) lowerCamelCase__ = False out_string += self.sp_model.decode(__lowerCAmelCase ) return out_string def __lowerCamelCase ( self ): '''simple docstring''' lowerCamelCase__ = {self.convert_ids_to_tokens(__lowerCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None ): '''simple docstring''' if not os.path.isdir(__lowerCAmelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return lowerCamelCase__ = os.path.join( __lowerCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowerCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __lowerCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__lowerCAmelCase , '''wb''' ) as fi: lowerCamelCase__ = self.sp_model.serialized_model_proto() fi.write(__lowerCAmelCase ) return (out_vocab_file,) def __lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = False ): '''simple docstring''' if isinstance(__lowerCAmelCase , __lowerCAmelCase ): lowerCamelCase__ = self.preprocess_text(__lowerCAmelCase ) lowerCamelCase__ = self.sp_model.encode(__lowerCAmelCase ) else: lowerCamelCase__ = [self.preprocess_text(__lowerCAmelCase ) for t in text] lowerCamelCase__ = self.sp_model.encode(__lowerCAmelCase ) if return_tensors is True or return_tensors == "pt": lowerCamelCase__ = torch.tensor(__lowerCAmelCase ) return token_ids def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' return self.sp_model.decode(__lowerCAmelCase ) def __lowerCamelCase ( self , __lowerCAmelCase ): '''simple docstring''' lowerCamelCase__ = [F'User: {text}' if is_user else F'Bot: {text}' for is_user, text in conversation.iter_texts()] lowerCamelCase__ = ( F'{self.eos_token}{self.bos_token}' + F'{self.bos_token}'.join(__lowerCAmelCase ) + F'{self.bos_token}Bot:' ) return self.encode(text=__lowerCAmelCase )
209
1
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __lowercase (UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): """simple docstring""" _snake_case = StableDiffusionDiffEditPipeline _snake_case = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""height""", """width""", """image"""} | {"""image_latents"""} _snake_case = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {"""image"""} | {"""image_latents"""} _snake_case = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _snake_case = frozenset([] ) def UpperCAmelCase ( self ) -> Tuple: torch.manual_seed(0 ) snake_case : List[Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=__A , ) snake_case : Any = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__A , set_alpha_to_one=__A , ) snake_case : Optional[int] = DDIMInverseScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__A , set_alpha_to_zero=__A , ) torch.manual_seed(0 ) snake_case : Any = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) snake_case : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="""gelu""" , projection_dim=5_1_2 , ) snake_case : List[Any] = CLIPTextModel(__A ) snake_case : int = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) snake_case : Union[str, Any] = { """unet""": unet, """scheduler""": scheduler, """inverse_scheduler""": inverse_scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def UpperCAmelCase ( self , A , A=0 ) -> Any: snake_case : List[str] = floats_tensor((1, 1_6, 1_6) , rng=random.Random(__A ) ).to(__A ) snake_case : List[Any] = floats_tensor((1, 2, 4, 1_6, 1_6) , rng=random.Random(__A ) ).to(__A ) if str(__A ).startswith("""mps""" ): snake_case : Optional[Any] = torch.manual_seed(__A ) else: snake_case : Optional[int] = torch.Generator(device=__A ).manual_seed(__A ) snake_case : int = { """prompt""": """a dog and a newt""", """mask_image""": mask, """image_latents""": latents, """generator""": generator, """num_inference_steps""": 2, """inpaint_strength""": 1.0, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCAmelCase ( self , A , A=0 ) -> Union[str, Any]: snake_case : Tuple = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__A ) ).to(__A ) snake_case : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case : Tuple = Image.fromarray(np.uinta(__A ) ).convert("""RGB""" ) if str(__A ).startswith("""mps""" ): snake_case : Tuple = torch.manual_seed(__A ) else: snake_case : Tuple = torch.Generator(device=__A ).manual_seed(__A ) snake_case : List[Any] = { """image""": image, """source_prompt""": """a cat and a frog""", """target_prompt""": """a dog and a newt""", """generator""": generator, """num_inference_steps""": 2, """num_maps_per_mask""": 2, """mask_encode_strength""": 1.0, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCAmelCase ( self , A , A=0 ) -> int: snake_case : Tuple = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__A ) ).to(__A ) snake_case : Any = image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case : Optional[Any] = Image.fromarray(np.uinta(__A ) ).convert("""RGB""" ) if str(__A ).startswith("""mps""" ): snake_case : int = torch.manual_seed(__A ) else: snake_case : List[str] = torch.Generator(device=__A ).manual_seed(__A ) snake_case : Union[str, Any] = { """image""": image, """prompt""": """a cat and a frog""", """generator""": generator, """num_inference_steps""": 2, """inpaint_strength""": 1.0, """guidance_scale""": 6.0, """decode_latents""": True, """output_type""": """numpy""", } return inputs def UpperCAmelCase ( self ) -> Optional[int]: if not hasattr(self.pipeline_class , """_optional_components""" ): return snake_case : str = self.get_dummy_components() snake_case : Optional[int] = self.pipeline_class(**__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(__A , __A , __A ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) snake_case : Union[str, Any] = self.get_dummy_inputs(__A ) snake_case : Any = pipe(**__A )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__A ) snake_case : Optional[Any] = self.pipeline_class.from_pretrained(__A ) pipe_loaded.to(__A ) pipe_loaded.set_progress_bar_config(disable=__A ) for optional_component in pipe._optional_components: self.assertTrue( getattr(__A , __A ) is None , f"""`{optional_component}` did not stay set to None after loading.""" , ) snake_case : Union[str, Any] = self.get_dummy_inputs(__A ) snake_case : Tuple = pipe_loaded(**__A )[0] snake_case : List[Any] = np.abs(output - output_loaded ).max() self.assertLess(__A , 1e-4 ) def UpperCAmelCase ( self ) -> Union[str, Any]: snake_case : Any = """cpu""" snake_case : str = self.get_dummy_components() snake_case : str = self.pipeline_class(**__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) snake_case : Any = self.get_dummy_mask_inputs(__A ) snake_case : List[Any] = pipe.generate_mask(**__A ) snake_case : int = mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 1_6, 1_6) ) snake_case : List[str] = np.array([0] * 9 ) snake_case : Union[str, Any] = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(__A , 1e-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def UpperCAmelCase ( self ) -> Optional[int]: snake_case : Optional[Any] = """cpu""" snake_case : int = self.get_dummy_components() snake_case : Tuple = self.pipeline_class(**__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) snake_case : Optional[int] = self.get_dummy_inversion_inputs(__A ) snake_case : str = pipe.invert(**__A ).images snake_case : Any = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 3_2, 3_2, 3) ) snake_case : Optional[int] = np.array( [0.51_50, 0.51_34, 0.50_43, 0.53_76, 0.46_94, 0.5_10_50, 0.50_15, 0.44_07, 0.47_99] , ) snake_case : Optional[Any] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__A , 1e-3 ) def UpperCAmelCase ( self ) -> Any: super().test_inference_batch_single_identical(expected_max_diff=5e-3 ) def UpperCAmelCase ( self ) -> Union[str, Any]: snake_case : str = """cpu""" snake_case : Tuple = self.get_dummy_components() snake_case : List[Any] = {"""beta_start""": 0.0_00_85, """beta_end""": 0.0_12, """beta_schedule""": """scaled_linear"""} snake_case : List[Any] = DPMSolverMultistepScheduler(**__A ) snake_case : List[str] = DPMSolverMultistepInverseScheduler(**__A ) snake_case : Union[str, Any] = self.pipeline_class(**__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) snake_case : Optional[int] = self.get_dummy_inversion_inputs(__A ) snake_case : Optional[int] = pipe.invert(**__A ).images snake_case : Optional[Any] = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 3_2, 3_2, 3) ) snake_case : Optional[Any] = np.array( [0.51_50, 0.51_34, 0.50_43, 0.53_76, 0.46_94, 0.5_10_50, 0.50_15, 0.44_07, 0.47_99] , ) snake_case : Dict = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__A , 1e-3 ) @require_torch_gpu @slow class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Tuple: super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def UpperCAmelCase ( cls ) -> Union[str, Any]: snake_case : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png""" ) snake_case : Any = raw_image.convert("""RGB""" ).resize((7_6_8, 7_6_8) ) snake_case : List[Any] = raw_image def UpperCAmelCase ( self ) -> str: snake_case : Any = torch.manual_seed(0 ) snake_case : str = StableDiffusionDiffEditPipeline.from_pretrained( """stabilityai/stable-diffusion-2-1""" , safety_checker=__A , torch_dtype=torch.floataa ) snake_case : Union[str, Any] = DDIMScheduler.from_config(pipe.scheduler.config ) snake_case : List[Any] = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__A ) snake_case : List[Any] = """a bowl of fruit""" snake_case : int = """a bowl of pears""" snake_case : Any = pipe.generate_mask( image=self.raw_image , source_prompt=__A , target_prompt=__A , generator=__A , ) snake_case : str = pipe.invert( prompt=__A , image=self.raw_image , inpaint_strength=0.7 , generator=__A ).latents snake_case : Optional[int] = pipe( prompt=__A , mask_image=__A , image_latents=__A , generator=__A , negative_prompt=__A , inpaint_strength=0.7 , output_type="""numpy""" , ).images[0] snake_case : List[Any] = ( np.array( load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/diffedit/pears.png""" ).resize((7_6_8, 7_6_8) ) ) / 2_5_5 ) assert np.abs((expected_image - image).max() ) < 5e-1 def UpperCAmelCase ( self ) -> Optional[int]: snake_case : Optional[Any] = torch.manual_seed(0 ) snake_case : Tuple = StableDiffusionDiffEditPipeline.from_pretrained( """stabilityai/stable-diffusion-2-1""" , safety_checker=__A , torch_dtype=torch.floataa ) snake_case : Dict = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) snake_case : Optional[int] = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__A ) snake_case : Union[str, Any] = """a bowl of fruit""" snake_case : Dict = """a bowl of pears""" snake_case : str = pipe.generate_mask( image=self.raw_image , source_prompt=__A , target_prompt=__A , generator=__A , ) snake_case : Tuple = pipe.invert( prompt=__A , image=self.raw_image , inpaint_strength=0.7 , generator=__A , num_inference_steps=2_5 , ).latents snake_case : Tuple = pipe( prompt=__A , mask_image=__A , image_latents=__A , generator=__A , negative_prompt=__A , inpaint_strength=0.7 , num_inference_steps=2_5 , output_type="""numpy""" , ).images[0] snake_case : Union[str, Any] = ( np.array( load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/diffedit/pears.png""" ).resize((7_6_8, 7_6_8) ) ) / 2_5_5 ) assert np.abs((expected_image - image).max() ) < 5e-1
359
import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging lowerCamelCase : Optional[Any] = logging.get_logger(__name__) lowerCamelCase : Optional[int] = { 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/config.json', # See all BART models at https://huggingface.co/models?filter=bart } class __lowercase (UpperCamelCase__ ): """simple docstring""" _snake_case = """bart""" _snake_case = ["""past_key_values"""] _snake_case = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , A=5_0_2_6_5 , A=1_0_2_4 , A=1_2 , A=4_0_9_6 , A=1_6 , A=1_2 , A=4_0_9_6 , A=1_6 , A=0.0 , A=0.0 , A="gelu" , A=1_0_2_4 , A=0.1 , A=0.0 , A=0.0 , A=0.02 , A=0.0 , A=False , A=True , A=3 , A=1 , A=0 , A=2 , A=True , A=2 , A=2 , **A , ) -> Any: snake_case : Optional[int] = vocab_size snake_case : Union[str, Any] = max_position_embeddings snake_case : List[str] = d_model snake_case : List[Any] = encoder_ffn_dim snake_case : Optional[Any] = encoder_layers snake_case : Union[str, Any] = encoder_attention_heads snake_case : str = decoder_ffn_dim snake_case : Union[str, Any] = decoder_layers snake_case : Any = decoder_attention_heads snake_case : Union[str, Any] = dropout snake_case : List[str] = attention_dropout snake_case : List[Any] = activation_dropout snake_case : Optional[int] = activation_function snake_case : Union[str, Any] = init_std snake_case : List[str] = encoder_layerdrop snake_case : int = decoder_layerdrop snake_case : str = classifier_dropout snake_case : List[str] = use_cache snake_case : Tuple = encoder_layers snake_case : int = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=A , pad_token_id=A , bos_token_id=A , eos_token_id=A , is_encoder_decoder=A , decoder_start_token_id=A , forced_eos_token_id=A , **A , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , A ): snake_case : Any = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" ) class __lowercase (UpperCamelCase__ ): """simple docstring""" @property def UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: snake_case : Optional[Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: snake_case : Tuple = {0: """batch"""} snake_case : List[Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: snake_case : Union[str, Any] = {0: """batch""", 1: """decoder_sequence"""} snake_case : Any = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(A , direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. snake_case : Union[str, Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: snake_case , snake_case : List[Any] = self.num_layers for i in range(A ): snake_case : List[Any] = {0: """batch""", 2: """past_sequence + sequence"""} snake_case : Optional[int] = {0: """batch""", 2: """past_sequence + sequence"""} else: snake_case : Union[str, Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property def UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: snake_case : Any = super().outputs else: snake_case : Any = super(A , self ).outputs if self.use_past: snake_case , snake_case : Any = self.num_layers for i in range(A ): snake_case : Any = {0: """batch""", 2: """past_sequence + sequence"""} snake_case : Union[str, Any] = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: snake_case : Optional[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , A , A , A , A ) # Generate decoder inputs snake_case : Any = seq_length if not self.use_past else 1 snake_case : List[str] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , A , A , A , A ) snake_case : Optional[int] = {f"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} snake_case : List[str] = dict(**A , **A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch snake_case , snake_case : Optional[int] = common_inputs["""input_ids"""].shape snake_case : Any = common_inputs["""decoder_input_ids"""].shape[1] snake_case , snake_case : Optional[Any] = self.num_attention_heads snake_case : Optional[int] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) snake_case : Any = decoder_seq_length + 3 snake_case : List[Any] = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) snake_case : str = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(A , A )] , dim=1 ) snake_case : str = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered snake_case , snake_case : Any = self.num_layers snake_case : List[str] = min(A , A ) snake_case : Dict = max(A , A ) - min_num_layers snake_case : List[str] = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(A ): common_inputs["past_key_values"].append( ( torch.zeros(A ), torch.zeros(A ), torch.zeros(A ), torch.zeros(A ), ) ) # TODO: test this. snake_case : Tuple = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(A , A ): common_inputs["past_key_values"].append((torch.zeros(A ), torch.zeros(A )) ) return common_inputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: snake_case : str = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , A , A , A , A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch snake_case , snake_case : str = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values snake_case : Optional[int] = seqlen + 2 snake_case , snake_case : Tuple = self.num_layers snake_case , snake_case : Optional[Any] = self.num_attention_heads snake_case : Union[str, Any] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) snake_case : Optional[Any] = common_inputs["""attention_mask"""].dtype snake_case : int = torch.cat( [common_inputs["""attention_mask"""], torch.ones(A , A , dtype=A )] , dim=1 ) snake_case : Union[str, Any] = [ (torch.zeros(A ), torch.zeros(A )) for _ in range(A ) ] return common_inputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX snake_case : int = compute_effective_axis_dimension( A , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX snake_case : int = tokenizer.num_special_tokens_to_add(A ) snake_case : Tuple = compute_effective_axis_dimension( A , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=A ) # Generate dummy inputs according to compute batch and sequence snake_case : int = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size snake_case : str = dict(tokenizer(A , return_tensors=A ) ) return common_inputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: snake_case : Optional[Any] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( A , batch_size=A , seq_length=A , is_pair=A , framework=A ) elif self.task == "causal-lm": snake_case : Optional[int] = self._generate_dummy_inputs_for_causal_lm( A , batch_size=A , seq_length=A , is_pair=A , framework=A ) else: snake_case : List[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , batch_size=A , seq_length=A , is_pair=A , framework=A ) return common_inputs def UpperCAmelCase ( self , A , A , A , A ) -> Union[str, Any]: if self.task in ["default", "seq2seq-lm"]: snake_case : Optional[Any] = super()._flatten_past_key_values_(A , A , A , A ) else: snake_case : Union[str, Any] = super(A , self )._flatten_past_key_values_( A , A , A , A )
176
0
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __A =logging.get_logger(__name__) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__=False ): lowerCamelCase_ = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight') ) rename_keys.append((F'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias', F'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((F'blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" lowerCamelCase_ = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): for i in range(config.num_hidden_layers ): if base_model: lowerCamelCase_ = "" else: lowerCamelCase_ = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase_ = state_dict.pop(F'blocks.{i}.attn.qkv.weight' ) lowerCamelCase_ = state_dict.pop(F'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase_ = in_proj_weight[ : config.hidden_size, : ] lowerCamelCase_ = in_proj_bias[: config.hidden_size] lowerCamelCase_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase_ = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase_ = in_proj_bias[-config.hidden_size :] def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(__lowerCamelCase , __lowerCamelCase ) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = dct.pop(__lowerCamelCase ) lowerCamelCase_ = val def lowerCamelCase_ ( ): lowerCamelCase_ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowerCamelCase_ = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): lowerCamelCase_ = BitConfig( global_padding="same" , layer_type="bottleneck" , depths=(3, 4, 9) , out_features=["stage3"] , embedding_dynamic_padding=__lowerCamelCase , ) lowerCamelCase_ = ViTHybridConfig(backbone_config=__lowerCamelCase , image_size=3_8_4 , num_labels=1_0_0_0 ) lowerCamelCase_ = False # load original model from timm lowerCamelCase_ = timm.create_model(__lowerCamelCase , pretrained=__lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys lowerCamelCase_ = timm_model.state_dict() if base_model: remove_classification_head_(__lowerCamelCase ) lowerCamelCase_ = create_rename_keys(__lowerCamelCase , __lowerCamelCase ) for src, dest in rename_keys: rename_key(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) read_in_q_k_v(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) lowerCamelCase_ = "huggingface/label-files" lowerCamelCase_ = "imagenet-1k-id2label.json" lowerCamelCase_ = json.load(open(hf_hub_download(__lowerCamelCase , __lowerCamelCase , repo_type="dataset" ) , "r" ) ) lowerCamelCase_ = {int(__lowerCamelCase ): v for k, v in idalabel.items()} lowerCamelCase_ = idalabel lowerCamelCase_ = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": lowerCamelCase_ = ViTHybridModel(__lowerCamelCase ).eval() else: lowerCamelCase_ = ViTHybridForImageClassification(__lowerCamelCase ).eval() model.load_state_dict(__lowerCamelCase ) # create image processor lowerCamelCase_ = create_transform(**resolve_data_config({} , model=__lowerCamelCase ) ) lowerCamelCase_ = transform.transforms lowerCamelCase_ = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } lowerCamelCase_ = ViTHybridImageProcessor( do_resize=__lowerCamelCase , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=__lowerCamelCase , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=__lowerCamelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowerCamelCase_ = prepare_img() lowerCamelCase_ = transform(__lowerCamelCase ).unsqueeze(0 ) lowerCamelCase_ = processor(__lowerCamelCase , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(__lowerCamelCase , __lowerCamelCase ) # verify logits with torch.no_grad(): lowerCamelCase_ = model(__lowerCamelCase ) lowerCamelCase_ = outputs.logits print("Predicted class:" , logits.argmax(-1 ).item() ) if base_model: lowerCamelCase_ = timm_model.forward_features(__lowerCamelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__lowerCamelCase , outputs.pooler_output , atol=1e-3 ) else: lowerCamelCase_ = timm_model(__lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCamelCase , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(F'Saving model {vit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__lowerCamelCase ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(__lowerCamelCase ) if push_to_hub: print(F'Pushing model and processor to the hub {vit_name}' ) model.push_to_hub(F'ybelkada/{vit_name}' ) processor.push_to_hub(F'ybelkada/{vit_name}' ) if __name__ == "__main__": __A =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--vit_name''', default='''vit_base_r50_s16_384''', type=str, help='''Name of the hybrid ViT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to upload the model to the HuggingFace hub.''' ) __A =parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
19
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def UpperCamelCase ( __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : PreTrainedTokenizer , __lowerCamelCase : int , __lowerCamelCase : Optional[int] = None , ): snake_case : int = {} if train_file is not None: snake_case : List[Any] = [train_file] if eval_file is not None: snake_case : Optional[int] = [eval_file] if test_file is not None: snake_case : Any = [test_file] snake_case : int = datasets.load_dataset("csv" , data_files=__lowerCamelCase ) snake_case : str = list(ds[list(files.keys() )[0]].features.keys() ) snake_case : int = features_name.pop(__lowerCamelCase ) snake_case : str = list(set(ds[list(files.keys() )[0]][label_name] ) ) snake_case : str = {label: i for i, label in enumerate(__lowerCamelCase )} snake_case : List[Any] = tokenizer.model_input_names snake_case : List[Any] = {} if len(__lowerCamelCase ) == 1: for k in files.keys(): snake_case : Tuple = ds[k].map( lambda __lowerCamelCase : tokenizer.batch_encode_plus( example[features_name[0]] , truncation=__lowerCamelCase , max_length=__lowerCamelCase , padding="max_length" ) , batched=__lowerCamelCase , ) elif len(__lowerCamelCase ) == 2: for k in files.keys(): snake_case : List[Any] = ds[k].map( lambda __lowerCamelCase : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=__lowerCamelCase , max_length=__lowerCamelCase , padding="max_length" , ) , batched=__lowerCamelCase , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: snake_case : Dict = {k: v for k, v in ex.items() if k in input_names} snake_case : Union[str, Any] = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: snake_case : str = {k: v for k, v in ex.items() if k in input_names} snake_case : Any = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: snake_case : str = {k: v for k, v in ex.items() if k in input_names} snake_case : List[str] = labelaid[ex[label_name]] yield (d, label) snake_case : int = ( tf.data.Dataset.from_generator( __lowerCamelCase , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: snake_case : Optional[Any] = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) snake_case : Tuple = ( tf.data.Dataset.from_generator( __lowerCamelCase , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: snake_case : List[str] = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) snake_case : Optional[int] = ( tf.data.Dataset.from_generator( __lowerCamelCase , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: snake_case : str = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid __lowerCamelCase = logging.getLogger(__name__) @dataclass class UpperCAmelCase : A__ : int = field(metadata={"help": "Which column contains the label"} ) A__ : str = field(default=A_ ,metadata={"help": "The path of the training file"} ) A__ : Optional[str] = field(default=A_ ,metadata={"help": "The path of the development file"} ) A__ : Optional[str] = field(default=A_ ,metadata={"help": "The path of the test file"} ) A__ : int = field( default=1_28 ,metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } ,) A__ : bool = field( default=A_ ,metadata={"help": "Overwrite the cached training and evaluation sets"} ) @dataclass class UpperCAmelCase : A__ : str = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) A__ : Optional[str] = field( default=A_ ,metadata={"help": "Pretrained config name or path if not the same as model_name"} ) A__ : Optional[str] = field( default=A_ ,metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) A__ : bool = field(default=A_ ,metadata={"help": "Set this flag to use fast tokenization."} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. A__ : Optional[str] = field( default=A_ ,metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} ,) def UpperCamelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. snake_case : List[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) snake_case , snake_case , snake_case : int = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" " --overwrite_output_dir to overcome." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.info( f"""n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, """ f"""16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case : Tuple = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) snake_case , snake_case , snake_case , snake_case : Tuple = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=__lowerCamelCase , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) snake_case : Optional[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(__lowerCamelCase ) , labelaid=__lowerCamelCase , idalabel={id: label for label, id in labelaid.items()} , finetuning_task="text-classification" , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): snake_case : int = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool(".bin" in model_args.model_name_or_path ) , config=__lowerCamelCase , cache_dir=model_args.cache_dir , ) def compute_metrics(__lowerCamelCase : EvalPrediction ) -> Dict: snake_case : Optional[int] = np.argmax(p.predictions , axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer snake_case : int = TFTrainer( model=__lowerCamelCase , args=__lowerCamelCase , train_dataset=__lowerCamelCase , eval_dataset=__lowerCamelCase , compute_metrics=__lowerCamelCase , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case : int = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) snake_case : Any = trainer.evaluate() snake_case : List[Any] = os.path.join(training_args.output_dir , "eval_results.txt" ) with open(__lowerCamelCase , "w" ) as writer: logger.info("***** Eval results *****" ) for key, value in result.items(): logger.info(f""" {key} = {value}""" ) writer.write(f"""{key} = {value}\n""" ) results.update(__lowerCamelCase ) return results if __name__ == "__main__": main()
59
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : List[Any] = logging.get_logger(__name__) lowerCamelCase : List[Any] = { "google/switch-base-8": "https://huggingface.co/google/switch-base-8/blob/main/config.json", } class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = '''switch_transformers''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''hidden_size''': '''d_model''', '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers'''} def __init__( self : str , A_ : int=32128 , A_ : Any=768 , A_ : List[Any]=64 , A_ : Optional[int]=2048 , A_ : Tuple=64 , A_ : Optional[Any]=12 , A_ : Dict=3 , A_ : Tuple=12 , A_ : int=3 , A_ : Any=12 , A_ : str=8 , A_ : int=False , A_ : Union[str, Any]=0.01 , A_ : int="float32" , A_ : Optional[Any]=False , A_ : Union[str, Any]=32 , A_ : Any=128 , A_ : List[str]=0.1 , A_ : Dict=1E-6 , A_ : Optional[Any]=0.001 , A_ : Dict=0.001 , A_ : Any=1.0 , A_ : Optional[Any]="relu" , A_ : Any=True , A_ : Dict=False , A_ : Tuple=True , A_ : Any=0 , A_ : Union[str, Any]=1 , **A_ : List[str] , ) -> str: """simple docstring""" lowerCamelCase_ = vocab_size lowerCamelCase_ = d_model lowerCamelCase_ = d_kv lowerCamelCase_ = d_ff lowerCamelCase_ = num_sparse_encoder_layers lowerCamelCase_ = num_layers lowerCamelCase_ = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry lowerCamelCase_ = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: lowerCamelCase_ = self.num_layers // self.num_sparse_encoder_layers else: lowerCamelCase_ = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: lowerCamelCase_ = self.num_decoder_layers // self.num_sparse_decoder_layers else: lowerCamelCase_ = self.num_decoder_layers # HACK: this will create 0 sparse layers lowerCamelCase_ = num_heads lowerCamelCase_ = num_experts lowerCamelCase_ = expert_capacity lowerCamelCase_ = router_bias lowerCamelCase_ = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(f"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) lowerCamelCase_ = router_dtype lowerCamelCase_ = router_ignore_padding_tokens lowerCamelCase_ = relative_attention_num_buckets lowerCamelCase_ = relative_attention_max_distance lowerCamelCase_ = dropout_rate lowerCamelCase_ = layer_norm_epsilon lowerCamelCase_ = initializer_factor lowerCamelCase_ = feed_forward_proj lowerCamelCase_ = use_cache lowerCamelCase_ = add_router_probs lowerCamelCase_ = router_z_loss_coef lowerCamelCase_ = router_aux_loss_coef lowerCamelCase_ = self.feed_forward_proj.split('-' ) lowerCamelCase_ = act_info[-1] lowerCamelCase_ = act_info[0] == 'gated' if len(A_ ) > 1 and act_info[0] != "gated" or len(A_ ) > 2: raise ValueError( f"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": lowerCamelCase_ = 'gelu_new' super().__init__( pad_token_id=A_ , eos_token_id=A_ , is_encoder_decoder=A_ , **A_ , )
370
import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel from transformers.file_utils import cached_path from transformers.testing_utils import TestCasePlus, require_torch_gpu, slow from utils import load_json lowerCamelCase : Tuple = "sshleifer/mar_enro_6_3_student" class A( UpperCamelCase ): '''simple docstring''' def a__ ( self : Union[str, Any] ) -> str: """simple docstring""" super().setUp() lowerCamelCase_ = cached_path( 'https://cdn-datasets.huggingface.co/translation/wmt_en_ro-tr40k-va0.5k-te0.5k.tar.gz' , extract_compressed_file=A_ , ) lowerCamelCase_ = f"""{data_cached}/wmt_en_ro-tr40k-va0.5k-te0.5k""" @slow @require_torch_gpu def a__ ( self : List[str] ) -> int: """simple docstring""" MarianMTModel.from_pretrained(A_ ) @slow @require_torch_gpu def a__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = { '$MAX_LEN': 64, '$BS': 64, '$GAS': 1, '$ENRO_DIR': self.data_dir, 'facebook/mbart-large-cc25': MARIAN_MODEL, # "val_check_interval=0.25": "val_check_interval=1.0", '--learning_rate=3e-5': '--learning_rate 3e-4', '--num_train_epochs 6': '--num_train_epochs 1', } # Clean up bash script lowerCamelCase_ = (self.test_file_dir / 'train_mbart_cc25_enro.sh').open().read().split('finetune.py' )[1].strip() lowerCamelCase_ = bash_script.replace('\\\n' , '' ).strip().replace('"$@"' , '' ) for k, v in env_vars_to_replace.items(): lowerCamelCase_ = bash_script.replace(A_ , str(A_ ) ) lowerCamelCase_ = self.get_auto_remove_tmp_dir() # bash_script = bash_script.replace("--fp16 ", "") lowerCamelCase_ = f""" --output_dir {output_dir} --tokenizer_name Helsinki-NLP/opus-mt-en-ro --sortish_sampler --do_predict --gpus 1 --freeze_encoder --n_train 40000 --n_val 500 --n_test 500 --fp16_opt_level O1 --num_sanity_val_steps 0 --eval_beams 2 """.split() # XXX: args.gpus > 1 : handle multi_gpu in the future lowerCamelCase_ = ['finetune.py'] + bash_script.split() + args with patch.object(A_ , 'argv' , A_ ): lowerCamelCase_ = argparse.ArgumentParser() lowerCamelCase_ = pl.Trainer.add_argparse_args(A_ ) lowerCamelCase_ = SummarizationModule.add_model_specific_args(A_ , os.getcwd() ) lowerCamelCase_ = parser.parse_args() lowerCamelCase_ = main(A_ ) # Check metrics lowerCamelCase_ = load_json(model.metrics_save_path ) lowerCamelCase_ = metrics['val'][0] lowerCamelCase_ = metrics['val'][-1] self.assertEqual(len(metrics['val'] ) , (args.max_epochs / args.val_check_interval) ) assert isinstance(last_step_stats[f"""val_avg_{model.val_metric}"""] , A_ ) self.assertGreater(last_step_stats['val_avg_gen_time'] , 0.01 ) # model hanging on generate. Maybe bad config was saved. (XXX: old comment/assert?) self.assertLessEqual(last_step_stats['val_avg_gen_time'] , 1.0 ) # test learning requirements: # 1. BLEU improves over the course of training by more than 2 pts self.assertGreater(last_step_stats['val_avg_bleu'] - first_step_stats['val_avg_bleu'] , 2 ) # 2. BLEU finishes above 17 self.assertGreater(last_step_stats['val_avg_bleu'] , 17 ) # 3. test BLEU and val BLEU within ~1.1 pt. self.assertLess(abs(metrics['val'][-1]['val_avg_bleu'] - metrics['test'][-1]['test_avg_bleu'] ) , 1.1 ) # check lightning ckpt can be loaded and has a reasonable statedict lowerCamelCase_ = os.listdir(A_ ) lowerCamelCase_ = [x for x in contents if x.endswith('.ckpt' )][0] lowerCamelCase_ = os.path.join(args.output_dir , A_ ) lowerCamelCase_ = torch.load(A_ , map_location='cpu' ) lowerCamelCase_ = 'model.model.decoder.layers.0.encoder_attn_layer_norm.weight' assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowerCamelCase_ = {os.path.basename(A_ ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['test'] ) == 1 class A( UpperCamelCase ): '''simple docstring''' @timeout_decorator.timeout(600 ) @slow @require_torch_gpu def a__ ( self : List[str] ) -> Any: """simple docstring""" lowerCamelCase_ = f"""{self.test_file_dir_str}/test_data/wmt_en_ro""" lowerCamelCase_ = { '--fp16_opt_level=O1': '', '$MAX_LEN': 128, '$BS': 16, '$GAS': 1, '$ENRO_DIR': data_dir, '$m': 'sshleifer/student_marian_en_ro_6_1', 'val_check_interval=0.25': 'val_check_interval=1.0', } # Clean up bash script lowerCamelCase_ = ( (self.test_file_dir / 'distil_marian_no_teacher.sh').open().read().split('distillation.py' )[1].strip() ) lowerCamelCase_ = bash_script.replace('\\\n' , '' ).strip().replace('"$@"' , '' ) lowerCamelCase_ = bash_script.replace('--fp16 ' , ' ' ) for k, v in env_vars_to_replace.items(): lowerCamelCase_ = bash_script.replace(A_ , str(A_ ) ) lowerCamelCase_ = self.get_auto_remove_tmp_dir() lowerCamelCase_ = bash_script.replace('--fp16' , '' ) lowerCamelCase_ = 6 lowerCamelCase_ = ( ['distillation.py'] + bash_script.split() + [ f"""--output_dir={output_dir}""", '--gpus=1', '--learning_rate=1e-3', f"""--num_train_epochs={epochs}""", '--warmup_steps=10', '--val_check_interval=1.0', '--do_predict', ] ) with patch.object(A_ , 'argv' , A_ ): lowerCamelCase_ = argparse.ArgumentParser() lowerCamelCase_ = pl.Trainer.add_argparse_args(A_ ) lowerCamelCase_ = SummarizationDistiller.add_model_specific_args(A_ , os.getcwd() ) lowerCamelCase_ = parser.parse_args() # assert args.gpus == gpus THIS BREAKS for multi_gpu lowerCamelCase_ = distill_main(A_ ) # Check metrics lowerCamelCase_ = load_json(model.metrics_save_path ) lowerCamelCase_ = metrics['val'][0] lowerCamelCase_ = metrics['val'][-1] assert len(metrics['val'] ) >= (args.max_epochs / args.val_check_interval) # +1 accounts for val_sanity_check assert last_step_stats["val_avg_gen_time"] >= 0.01 assert first_step_stats["val_avg_bleu"] < last_step_stats["val_avg_bleu"] # model learned nothing assert 1.0 >= last_step_stats["val_avg_gen_time"] # model hanging on generate. Maybe bad config was saved. assert isinstance(last_step_stats[f"""val_avg_{model.val_metric}"""] , A_ ) # check lightning ckpt can be loaded and has a reasonable statedict lowerCamelCase_ = os.listdir(A_ ) lowerCamelCase_ = [x for x in contents if x.endswith('.ckpt' )][0] lowerCamelCase_ = os.path.join(args.output_dir , A_ ) lowerCamelCase_ = torch.load(A_ , map_location='cpu' ) lowerCamelCase_ = 'model.model.decoder.layers.0.encoder_attn_layer_norm.weight' assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowerCamelCase_ = {os.path.basename(A_ ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['test'] ) == 1
208
0
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class lowercase__ ( unittest.TestCase ): '''simple docstring''' def UpperCAmelCase_ ( self ): # For consistency across different places the DisjunctiveConstraint is called, # dc.token_ids is a list of integers. It is also initialized only by integers. _SCREAMING_SNAKE_CASE : List[str] = [[1, 2, 4], [1, 2, 3, 4]] _SCREAMING_SNAKE_CASE : Dict = DisjunctiveConstraint(__snake_case ) self.assertTrue(isinstance(dc.token_ids , __snake_case ) ) with self.assertRaises(__snake_case ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(__snake_case ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def UpperCAmelCase_ ( self ): # We can't have constraints that are complete subsets of another. This leads to a preverse # interpretation of "constraint fulfillment": does generating [1,2,3] fulfill the constraint? # It would mean that it generated [1,2] which fulfills it, but it's in the middle of potentially # fulfilling [1,2,3,4]. If we believe that [1,2,3] does fulfill the constraint, then the algorithm # will necessarily never reach [1,2,3,4], giving users a false sense of control (better to just not allow it). _SCREAMING_SNAKE_CASE : Union[str, Any] = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(__snake_case ): DisjunctiveConstraint(__snake_case ) # fails here def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : List[Any] = [[1, 2, 3], [1, 2, 4]] _SCREAMING_SNAKE_CASE : Tuple = DisjunctiveConstraint(__snake_case ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : List[Any] = dc.update(1 ) _SCREAMING_SNAKE_CASE : Optional[Any] = stepped is True and completed is False and reset is False self.assertTrue(__snake_case ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : str = dc.update(2 ) _SCREAMING_SNAKE_CASE : List[str] = stepped is True and completed is False and reset is False self.assertTrue(__snake_case ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Union[str, Any] = dc.update(3 ) _SCREAMING_SNAKE_CASE : Tuple = stepped is True and completed is True and reset is False self.assertTrue(__snake_case ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : Any = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] _SCREAMING_SNAKE_CASE : List[str] = DisjunctiveConstraint(__snake_case ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Optional[int] = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : List[str] = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : List[Any] = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Optional[Any] = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Tuple = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Optional[int] = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Optional[int] = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
200
'''simple docstring''' from __future__ import annotations import os from typing import Any import requests UpperCAmelCase_ : List[Any] = 'https://api.github.com' # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user UpperCAmelCase_ : Optional[Any] = BASE_URL + '/user' # https://github.com/settings/tokens UpperCAmelCase_ : str = os.environ.get('USER_TOKEN', '') def snake_case_ ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _SCREAMING_SNAKE_CASE : Union[str, Any] = { """Authorization""": f"""token {auth_token}""", """Accept""": """application/vnd.github.v3+json""", } return requests.get(SCREAMING_SNAKE_CASE__ , headers=SCREAMING_SNAKE_CASE__ ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F"{key}: {value}") else: raise ValueError('\'USER_TOKEN\' field cannot be empty.')
200
1
'''simple docstring''' from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowerCamelCase = { """configuration_vivit""": ["""VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """VivitConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ["""VivitImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ """VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """VivitModel""", """VivitPreTrainedModel""", """VivitForVideoClassification""", ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
351
'''simple docstring''' import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing the experiment tracking capability, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _lowerCamelCase = 16 _lowerCamelCase = 32 def a__ ( _SCREAMING_SNAKE_CASE : Accelerator , _SCREAMING_SNAKE_CASE : int = 16 ) -> List[str]: """simple docstring""" UpperCAmelCase_ : Union[str, Any] = AutoTokenizer.from_pretrained("bert-base-cased" ) UpperCAmelCase_ : Optional[Any] = load_dataset("glue" , "mrpc" ) def tokenize_function(_SCREAMING_SNAKE_CASE : int ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase_ : str = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): UpperCAmelCase_ : Optional[Any] = datasets.map( _SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE , remove_columns=["idx", "sentence1", "sentence2"] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library UpperCAmelCase_ : Any = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(_SCREAMING_SNAKE_CASE : Any ): # On TPU it's best to pad everything to the same length or training will be very slow. UpperCAmelCase_ : Optional[Any] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": UpperCAmelCase_ : Optional[Any] = 16 elif accelerator.mixed_precision != "no": UpperCAmelCase_ : Union[str, Any] = 8 else: UpperCAmelCase_ : List[str] = None return tokenizer.pad( _SCREAMING_SNAKE_CASE , padding="longest" , max_length=_SCREAMING_SNAKE_CASE , pad_to_multiple_of=_SCREAMING_SNAKE_CASE , return_tensors="pt" , ) # Instantiate dataloaders. UpperCAmelCase_ : Union[str, Any] = DataLoader( tokenized_datasets["train"] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = DataLoader( tokenized_datasets["validation"] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders _lowerCamelCase = mocked_dataloaders # noqa: F811 def a__ ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Tuple ) -> str: """simple docstring""" if os.environ.get("TESTING_MOCKED_DATALOADERS" , _SCREAMING_SNAKE_CASE ) == "1": UpperCAmelCase_ : Tuple = 2 # Initialize Accelerator # New Code # # We pass in "all" to `log_with` to grab all available trackers in the environment # Note: If using a custom `Tracker` class, should be passed in here such as: # >>> log_with = ["all", MyCustomTrackerClassInstance()] if args.with_tracking: UpperCAmelCase_ : Optional[Any] = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with="all" , project_dir=args.project_dir ) else: UpperCAmelCase_ : List[Any] = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCAmelCase_ : Optional[Any] = config["lr"] UpperCAmelCase_ : Union[str, Any] = int(config["num_epochs"] ) UpperCAmelCase_ : str = int(config["seed"] ) UpperCAmelCase_ : Tuple = int(config["batch_size"] ) set_seed(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = get_dataloaders(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = evaluate.load("glue" , "mrpc" ) # If the batch size is too big we use gradient accumulation UpperCAmelCase_ : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: UpperCAmelCase_ : Tuple = batch_size // MAX_GPU_BATCH_SIZE UpperCAmelCase_ : Tuple = MAX_GPU_BATCH_SIZE # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCAmelCase_ : Tuple = AutoModelForSequenceClassification.from_pretrained("bert-base-cased" , return_dict=_SCREAMING_SNAKE_CASE ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). UpperCAmelCase_ : Union[str, Any] = model.to(accelerator.device ) # Instantiate optimizer UpperCAmelCase_ : int = AdamW(params=model.parameters() , lr=_SCREAMING_SNAKE_CASE ) # Instantiate scheduler UpperCAmelCase_ : Optional[int] = get_linear_schedule_with_warmup( optimizer=_SCREAMING_SNAKE_CASE , num_warmup_steps=1_00 , num_training_steps=(len(_SCREAMING_SNAKE_CASE ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : int = accelerator.prepare( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # New Code # # We need to initialize the trackers we use. Overall configurations can also be stored if args.with_tracking: UpperCAmelCase_ : List[str] = os.path.split(_SCREAMING_SNAKE_CASE )[-1].split("." )[0] accelerator.init_trackers(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Now we train the model for epoch in range(_SCREAMING_SNAKE_CASE ): model.train() # New Code # # For our tracking example, we will log the total loss of each epoch if args.with_tracking: UpperCAmelCase_ : Dict = 0 for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) UpperCAmelCase_ : Union[str, Any] = model(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = outputs.loss # New Code # if args.with_tracking: total_loss += loss.detach().float() UpperCAmelCase_ : List[str] = loss / gradient_accumulation_steps accelerator.backward(_SCREAMING_SNAKE_CASE ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True` (the default). batch.to(accelerator.device ) with torch.no_grad(): UpperCAmelCase_ : Optional[int] = model(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Tuple = outputs.logits.argmax(dim=-1 ) UpperCAmelCase_ , UpperCAmelCase_ : List[str] = accelerator.gather_for_metrics((predictions, batch["labels"]) ) metric.add_batch( predictions=_SCREAMING_SNAKE_CASE , references=_SCREAMING_SNAKE_CASE , ) UpperCAmelCase_ : Dict = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''' , _SCREAMING_SNAKE_CASE ) # New Code # # To actually log, we call `Accelerator.log` # The values passed can be of `str`, `int`, `float` or `dict` of `str` to `float`/`int` if args.with_tracking: accelerator.log( { "accuracy": eval_metric["accuracy"], "f1": eval_metric["f1"], "train_loss": total_loss.item() / len(_SCREAMING_SNAKE_CASE ), "epoch": epoch, } , step=_SCREAMING_SNAKE_CASE , ) # New Code # # When a run is finished, you should call `accelerator.end_training()` # to close all of the open trackers if args.with_tracking: accelerator.end_training() def a__ ( ) -> List[str]: """simple docstring""" UpperCAmelCase_ : int = argparse.ArgumentParser(description="Simple example of training script." ) parser.add_argument( "--mixed_precision" , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , choices=["no", "fp16", "bf16", "fp8"] , help="Whether to use mixed precision. Choose" "between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10." "and an Nvidia Ampere GPU." , ) parser.add_argument("--cpu" , action="store_true" , help="If passed, will train on the CPU." ) parser.add_argument( "--with_tracking" , action="store_true" , help="Whether to load in all available experiment trackers from the environment and use them for logging." , ) parser.add_argument( "--project_dir" , type=_SCREAMING_SNAKE_CASE , default="logs" , help="Location on where to store experiment tracking logs` and relevent project information" , ) UpperCAmelCase_ : List[Any] = parser.parse_args() UpperCAmelCase_ : Dict = {"lr": 2E-5, "num_epochs": 3, "seed": 42, "batch_size": 16} training_function(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
67
0
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0 ) -> list: lowerCAmelCase__ : Optional[int] = length or len(_UpperCamelCase ) lowerCAmelCase__ : Optional[Any] = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: lowerCAmelCase__ : Union[str, Any] = list_data[i + 1], list_data[i] lowerCAmelCase__ : str = True return list_data if not swapped else bubble_sort(_UpperCamelCase , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
212
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase : Union[str, Any] = {'configuration_mra': ['MRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MraConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : int = [ 'MRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MraForMaskedLM', 'MraForMultipleChoice', 'MraForQuestionAnswering', 'MraForSequenceClassification', 'MraForTokenClassification', 'MraLayer', 'MraModel', 'MraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys UpperCAmelCase : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure)
115
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() a__ = logging.get_logger(__name__) def UpperCamelCase__( UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : str=False )->List[Any]: A__ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"blocks.{i}.norm1.weight", f"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((f"blocks.{i}.norm1.bias", f"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append((f"blocks.{i}.attn.proj.weight", f"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((f"blocks.{i}.attn.proj.bias", f"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((f"blocks.{i}.norm2.weight", f"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((f"blocks.{i}.norm2.bias", f"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((f"blocks.{i}.mlp.fc1.weight", f"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((f"blocks.{i}.mlp.fc1.bias", f"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((f"blocks.{i}.mlp.fc2.weight", f"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((f"blocks.{i}.mlp.fc2.bias", f"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" A__ = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def UpperCamelCase__( UpperCamelCase__ : int , UpperCamelCase__ : Dict , UpperCamelCase__ : Optional[Any]=False )->Optional[Any]: for i in range(config.num_hidden_layers ): if base_model: A__ = """""" else: A__ = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ = state_dict.pop(f"blocks.{i}.attn.qkv.weight" ) A__ = state_dict.pop(f"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict A__ = in_proj_weight[ : config.hidden_size, : ] A__ = in_proj_bias[: config.hidden_size] A__ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ = in_proj_weight[ -config.hidden_size :, : ] A__ = in_proj_bias[-config.hidden_size :] def UpperCamelCase__( UpperCamelCase__ : Any )->Any: A__ = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(lowercase__ , lowercase__ ) def UpperCamelCase__( UpperCamelCase__ : List[str] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Optional[Any] )->List[str]: A__ = dct.pop(lowercase__ ) A__ = val def UpperCamelCase__( )->str: A__ = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ = Image.open(requests.get(lowercase__ , stream=lowercase__ ).raw ) return im @torch.no_grad() def UpperCamelCase__( UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Any=True )->Optional[int]: A__ = ViTConfig() # patch_size if model_name[-1] == "8": A__ = 8 # set labels if required if not base_model: A__ = 10_00 A__ = """huggingface/label-files""" A__ = """imagenet-1k-id2label.json""" A__ = json.load(open(hf_hub_download(lowercase__ , lowercase__ , repo_type='''dataset''' ) , '''r''' ) ) A__ = {int(lowercase__ ): v for k, v in idalabel.items()} A__ = idalabel A__ = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: A__ = 3_84 A__ = 15_36 A__ = 12 A__ = 6 # load original model from torch hub A__ = torch.hub.load('''facebookresearch/dino:main''' , lowercase__ ) original_model.eval() # load state_dict of original model, remove and rename some keys A__ = original_model.state_dict() if base_model: remove_classification_head_(lowercase__ ) A__ = create_rename_keys(lowercase__ , base_model=lowercase__ ) for src, dest in rename_keys: rename_key(lowercase__ , lowercase__ , lowercase__ ) read_in_q_k_v(lowercase__ , lowercase__ , lowercase__ ) # load HuggingFace model if base_model: A__ = ViTModel(lowercase__ , add_pooling_layer=lowercase__ ).eval() else: A__ = ViTForImageClassification(lowercase__ ).eval() model.load_state_dict(lowercase__ ) # Check outputs on an image, prepared by ViTImageProcessor A__ = ViTImageProcessor() A__ = image_processor(images=prepare_img() , return_tensors='''pt''' ) A__ = encoding["""pixel_values"""] A__ = model(lowercase__ ) if base_model: A__ = original_model(lowercase__ ) assert torch.allclose(lowercase__ , outputs.last_hidden_state[:, 0, :] , atol=1e-1 ) else: A__ = original_model(lowercase__ ) assert logits.shape == outputs.logits.shape assert torch.allclose(lowercase__ , outputs.logits , atol=1e-3 ) Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) print(f"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(lowercase__ ) print(f"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(lowercase__ ) if __name__ == "__main__": a__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='dino_vitb16', type=str, help='Name of the model trained with DINO you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--base_model', action='store_true', help='Whether to only convert the base model (no projection head weights).', ) parser.set_defaults(base_model=True) a__ = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
356
a__: dict[str, float] = { "km/h": 1.0, "m/s": 3.6, "mph": 1.60_93_44, "knot": 1.8_52, } a__: dict[str, float] = { "km/h": 1.0, "m/s": 0.2_77_77_77_78, "mph": 0.6_21_37_11_92, "knot": 0.5_39_95_68_03, } def UpperCamelCase__( UpperCamelCase__ : float , UpperCamelCase__ : str , UpperCamelCase__ : str )->float: if unit_to not in speed_chart or unit_from not in speed_chart_inverse: A__ = ( f"Incorrect 'from_type' or 'to_type' value: {unit_from!r}, {unit_to!r}\n" f"Valid values are: {', '.join(UpperCamelCase__ )}" ) raise ValueError(UpperCamelCase__ ) return round(speed * speed_chart[unit_from] * speed_chart_inverse[unit_to] , 3 ) if __name__ == "__main__": import doctest doctest.testmod()
39
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _snake_case ( a__ , a__ , unittest.TestCase ): snake_case__ = StableDiffusionXLImgaImgPipeline snake_case__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} snake_case__ = PipelineTesterMixin.required_optional_params - {"""latents"""} snake_case__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS snake_case__ = IMAGE_TO_IMAGE_IMAGE_PARAMS snake_case__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase__ ( self : Dict ): torch.manual_seed(0 ) __lowerCamelCase : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , attention_head_dim=(2, 4) , use_linear_projection=__UpperCAmelCase , addition_embed_type="text_time" , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __lowerCamelCase : str = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule="scaled_linear" , timestep_spacing="leading" , ) torch.manual_seed(0 ) __lowerCamelCase : str = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __lowerCamelCase : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=32 , ) __lowerCamelCase : int = CLIPTextModel(__UpperCAmelCase ) __lowerCamelCase : Tuple = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" , local_files_only=__UpperCAmelCase ) __lowerCamelCase : Any = CLIPTextModelWithProjection(__UpperCAmelCase ) __lowerCamelCase : Optional[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" , local_files_only=__UpperCAmelCase ) __lowerCamelCase : str = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'text_encoder_2': text_encoder_a, 'tokenizer_2': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def lowerCamelCase__ ( self : Union[str, Any] , UpperCAmelCase : Dict , UpperCAmelCase : Dict=0 ): __lowerCamelCase : Dict = floats_tensor((1, 3, 32, 32) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) __lowerCamelCase : Optional[int] = image / 2 + 0.5 if str(__UpperCAmelCase ).startswith("mps" ): __lowerCamelCase : Optional[int] = torch.manual_seed(__UpperCAmelCase ) else: __lowerCamelCase : Optional[Any] = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase : Dict = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 5.0, 'output_type': 'numpy', 'strength': 0.7_5, } return inputs def lowerCamelCase__ ( self : Tuple ): __lowerCamelCase : List[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : int = self.get_dummy_components() __lowerCamelCase : List[str] = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase : str = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase : str = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase : int = sd_pipe(**__UpperCAmelCase ).images __lowerCamelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase : List[str] = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self : Optional[Any] ): super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def lowerCamelCase__ ( self : Any ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowerCamelCase__ ( self : Any ): pass def lowerCamelCase__ ( self : Optional[Any] ): __lowerCamelCase : Tuple = self.get_dummy_components() __lowerCamelCase : str = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase : str = sd_pipe.to(__UpperCAmelCase ) __lowerCamelCase : List[str] = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) # forward without prompt embeds __lowerCamelCase : int = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase : Optional[int] = 3 * ['this is a negative prompt'] __lowerCamelCase : Tuple = negative_prompt __lowerCamelCase : str = 3 * [inputs['prompt']] __lowerCamelCase : Optional[Any] = sd_pipe(**__UpperCAmelCase ) __lowerCamelCase : List[Any] = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCamelCase : Optional[Any] = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase : Tuple = 3 * ['this is a negative prompt'] __lowerCamelCase : str = 3 * [inputs.pop("prompt" )] ( __lowerCamelCase ) : List[str] = sd_pipe.encode_prompt(__UpperCAmelCase , negative_prompt=__UpperCAmelCase ) __lowerCamelCase : str = sd_pipe( **__UpperCAmelCase , prompt_embeds=__UpperCAmelCase , negative_prompt_embeds=__UpperCAmelCase , pooled_prompt_embeds=__UpperCAmelCase , negative_pooled_prompt_embeds=__UpperCAmelCase , ) __lowerCamelCase : Optional[Any] = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class _snake_case ( unittest.TestCase ): def lowerCamelCase__ ( self : List[Any] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase__ ( self : List[Any] , UpperCAmelCase : List[Any] , UpperCAmelCase : str="cpu" , UpperCAmelCase : str=torch.floataa , UpperCAmelCase : List[Any]=0 ): __lowerCamelCase : Any = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase : Dict = np.random.RandomState(__UpperCAmelCase ).standard_normal((1, 4, 64, 64) ) __lowerCamelCase : Optional[int] = torch.from_numpy(__UpperCAmelCase ).to(device=__UpperCAmelCase , dtype=__UpperCAmelCase ) __lowerCamelCase : int = { 'prompt': 'a photograph of an astronaut riding a horse', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def lowerCamelCase__ ( self : List[str] ): __lowerCamelCase : List[Any] = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-2-base" ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase : Tuple = self.get_inputs(__UpperCAmelCase ) __lowerCamelCase : int = pipe(**__UpperCAmelCase ).images __lowerCamelCase : Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __lowerCamelCase : List[str] = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
135
"""simple docstring""" def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: """simple docstring""" if discount_rate < 0: raise ValueError('Discount rate cannot be negative' ) if not cash_flows: raise ValueError('Cash flows list cannot be empty' ) lowerCAmelCase__ :Union[str, Any] = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(_SCREAMING_SNAKE_CASE ) ) return round(_SCREAMING_SNAKE_CASE , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
293
0
"""simple docstring""" import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class snake_case ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): a_ : Optional[Any] = ConsistencyModelPipeline a_ : int = UNCONDITIONAL_IMAGE_GENERATION_PARAMS a_ : str = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt a_ : Union[str, Any] = frozenset( [ """num_inference_steps""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) @property def UpperCAmelCase__ ( self ) ->Union[str, Any]: a_ = UNetaDModel.from_pretrained( "diffusers/consistency-models-test" , subfolder="test_unet" , ) return unet @property def UpperCAmelCase__ ( self ) ->str: a_ = UNetaDModel.from_pretrained( "diffusers/consistency-models-test" , subfolder="test_unet_class_cond" , ) return unet def UpperCAmelCase__ ( self , __UpperCAmelCase=False ) ->List[Any]: if class_cond: a_ = self.dummy_cond_unet else: a_ = self.dummy_uncond_unet # Default to CM multistep sampler a_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a_ = { "unet": unet, "scheduler": scheduler, } return components def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase=0 ) ->Union[str, Any]: if str(__UpperCAmelCase ).startswith("mps" ): a_ = torch.manual_seed(__UpperCAmelCase ) else: a_ = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) a_ = { "batch_size": 1, "num_inference_steps": None, "timesteps": [22, 0], "generator": generator, "output_type": "np", } return inputs def UpperCAmelCase__ ( self ) ->Any: a_ = "cpu" # ensure determinism for the device-dependent torch.Generator a_ = self.get_dummy_components() a_ = ConsistencyModelPipeline(**__UpperCAmelCase ) a_ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a_ = self.get_dummy_inputs(__UpperCAmelCase ) a_ = pipe(**__UpperCAmelCase ).images assert image.shape == (1, 32, 32, 3) a_ = image[0, -3:, -3:, -1] a_ = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def UpperCAmelCase__ ( self ) ->List[str]: a_ = "cpu" # ensure determinism for the device-dependent torch.Generator a_ = self.get_dummy_components(class_cond=__UpperCAmelCase ) a_ = ConsistencyModelPipeline(**__UpperCAmelCase ) a_ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a_ = self.get_dummy_inputs(__UpperCAmelCase ) a_ = 0 a_ = pipe(**__UpperCAmelCase ).images assert image.shape == (1, 32, 32, 3) a_ = image[0, -3:, -3:, -1] a_ = np.array([0.3_572, 0.6_273, 0.4_031, 0.3_961, 0.4_321, 0.5_730, 0.5_266, 0.4_780, 0.5_004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def UpperCAmelCase__ ( self ) ->str: a_ = "cpu" # ensure determinism for the device-dependent torch.Generator a_ = self.get_dummy_components() a_ = ConsistencyModelPipeline(**__UpperCAmelCase ) a_ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a_ = self.get_dummy_inputs(__UpperCAmelCase ) a_ = 1 a_ = None a_ = pipe(**__UpperCAmelCase ).images assert image.shape == (1, 32, 32, 3) a_ = image[0, -3:, -3:, -1] a_ = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def UpperCAmelCase__ ( self ) ->int: a_ = "cpu" # ensure determinism for the device-dependent torch.Generator a_ = self.get_dummy_components(class_cond=__UpperCAmelCase ) a_ = ConsistencyModelPipeline(**__UpperCAmelCase ) a_ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a_ = self.get_dummy_inputs(__UpperCAmelCase ) a_ = 1 a_ = None a_ = 0 a_ = pipe(**__UpperCAmelCase ).images assert image.shape == (1, 32, 32, 3) a_ = image[0, -3:, -3:, -1] a_ = np.array([0.5_004, 0.5_004, 0.4_994, 0.5_008, 0.4_976, 0.5_018, 0.4_990, 0.4_982, 0.4_987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) ->Any: super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self , __UpperCAmelCase=0 , __UpperCAmelCase=False , __UpperCAmelCase="cpu" , __UpperCAmelCase=torch.floataa , __UpperCAmelCase=(1, 3, 64, 64) ) ->List[str]: a_ = torch.manual_seed(__UpperCAmelCase ) a_ = { "num_inference_steps": None, "timesteps": [22, 0], "class_labels": 0, "generator": generator, "output_type": "np", } if get_fixed_latents: a_ = self.get_fixed_latents(seed=__UpperCAmelCase , device=__UpperCAmelCase , dtype=__UpperCAmelCase , shape=__UpperCAmelCase ) a_ = latents return inputs def UpperCAmelCase__ ( self , __UpperCAmelCase=0 , __UpperCAmelCase="cpu" , __UpperCAmelCase=torch.floataa , __UpperCAmelCase=(1, 3, 64, 64) ) ->List[Any]: if type(__UpperCAmelCase ) == str: a_ = torch.device(__UpperCAmelCase ) a_ = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) a_ = randn_tensor(__UpperCAmelCase , generator=__UpperCAmelCase , device=__UpperCAmelCase , dtype=__UpperCAmelCase ) return latents def UpperCAmelCase__ ( self ) ->Any: a_ = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) a_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a_ = ConsistencyModelPipeline(unet=__UpperCAmelCase , scheduler=__UpperCAmelCase ) pipe.to(torch_device=__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a_ = self.get_inputs() a_ = pipe(**__UpperCAmelCase ).images assert image.shape == (1, 64, 64, 3) a_ = image[0, -3:, -3:, -1] a_ = np.array([0.0_888, 0.0_881, 0.0_666, 0.0_479, 0.0_292, 0.0_195, 0.0_201, 0.0_163, 0.0_254] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def UpperCAmelCase__ ( self ) ->List[Any]: a_ = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) a_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a_ = ConsistencyModelPipeline(unet=__UpperCAmelCase , scheduler=__UpperCAmelCase ) pipe.to(torch_device=__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a_ = self.get_inputs() a_ = 1 a_ = None a_ = pipe(**__UpperCAmelCase ).images assert image.shape == (1, 64, 64, 3) a_ = image[0, -3:, -3:, -1] a_ = np.array([0.0_340, 0.0_152, 0.0_063, 0.0_267, 0.0_221, 0.0_107, 0.0_416, 0.0_186, 0.0_217] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def UpperCAmelCase__ ( self ) ->Dict: a_ = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) a_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a_ = ConsistencyModelPipeline(unet=__UpperCAmelCase , scheduler=__UpperCAmelCase ) pipe.to(torch_device=__UpperCAmelCase , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a_ = self.get_inputs(get_fixed_latents=__UpperCAmelCase , device=__UpperCAmelCase ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=__UpperCAmelCase , enable_math=__UpperCAmelCase , enable_mem_efficient=__UpperCAmelCase ): a_ = pipe(**__UpperCAmelCase ).images assert image.shape == (1, 64, 64, 3) a_ = image[0, -3:, -3:, -1] a_ = np.array([0.1_875, 0.1_428, 0.1_289, 0.2_151, 0.2_092, 0.1_477, 0.1_877, 0.1_641, 0.1_353] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def UpperCAmelCase__ ( self ) ->int: a_ = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) a_ = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a_ = ConsistencyModelPipeline(unet=__UpperCAmelCase , scheduler=__UpperCAmelCase ) pipe.to(torch_device=__UpperCAmelCase , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a_ = self.get_inputs(get_fixed_latents=__UpperCAmelCase , device=__UpperCAmelCase ) a_ = 1 a_ = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=__UpperCAmelCase , enable_math=__UpperCAmelCase , enable_mem_efficient=__UpperCAmelCase ): a_ = pipe(**__UpperCAmelCase ).images assert image.shape == (1, 64, 64, 3) a_ = image[0, -3:, -3:, -1] a_ = np.array([0.1_663, 0.1_948, 0.2_275, 0.1_680, 0.1_204, 0.1_245, 0.1_858, 0.1_338, 0.2_095] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
354
"""simple docstring""" from heapq import heappop, heappush import numpy as np def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , ) ->tuple[float | int, list[tuple[int, int]]]: """simple docstring""" a_ , a_ = grid.shape a_ = [-1, 1, 0, 0] a_ = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] a_ , a_ = [(0, source)], set() a_ = np.full((rows, cols) , np.inf ) a_ = 0 a_ = np.empty((rows, cols) , dtype=UpperCAmelCase ) a_ = None while queue: ((a_) , (a_)) = heappop(UpperCAmelCase ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: a_ = [] while (x, y) != source: path.append((x, y) ) a_ , a_ = predecessors[x, y] path.append(UpperCAmelCase ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(UpperCAmelCase ) ): a_ , a_ = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: a_ = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(UpperCAmelCase , (dist + 1, (nx, ny)) ) a_ = dist + 1 a_ = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
303
0
"""simple docstring""" import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class snake_case ( snake_case_ , unittest.TestCase ): '''simple docstring''' A_ : Tuple = "hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline" def _SCREAMING_SNAKE_CASE ( self : Any, _lowerCamelCase : Union[str, Any]=0 ): '''simple docstring''' __A = floats_tensor((1, 3, 1_28, 1_28), rng=random.Random(_A ) ) __A = np.random.RandomState(_A ) __A = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'strength': 0.75, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def _SCREAMING_SNAKE_CASE ( self : List[Any] ): '''simple docstring''' __A = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint, provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=_A ) __A = self.get_dummy_inputs() __A = pipe(**_A ).images __A = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 1_28, 1_28, 3) __A = np.array([0.6_96_43, 0.5_84_84, 0.5_03_14, 0.5_87_60, 0.5_53_68, 0.5_96_43, 0.5_15_29, 0.4_12_17, 0.4_90_87] ) assert np.abs(image_slice - expected_slice ).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self : Dict ): '''simple docstring''' __A = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint, provider='''CPUExecutionProvider''' ) __A = PNDMScheduler.from_config(pipe.scheduler.config, skip_prk_steps=_A ) pipe.set_progress_bar_config(disable=_A ) __A = self.get_dummy_inputs() __A = pipe(**_A ).images __A = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __A = np.array([0.6_17_37, 0.5_46_42, 0.5_31_83, 0.5_44_65, 0.5_27_42, 0.6_05_25, 0.4_99_69, 0.4_06_55, 0.4_81_54] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self : List[str] ): '''simple docstring''' __A = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint, provider='''CPUExecutionProvider''' ) __A = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_A ) # warmup pass to apply optimizations __A = pipe(**self.get_dummy_inputs() ) __A = self.get_dummy_inputs() __A = pipe(**_A ).images __A = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __A = np.array([0.5_27_61, 0.5_99_77, 0.4_90_33, 0.4_96_19, 0.5_42_82, 0.5_03_11, 0.4_76_00, 0.4_09_18, 0.4_52_03] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' __A = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint, provider='''CPUExecutionProvider''' ) __A = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_A ) __A = self.get_dummy_inputs() __A = pipe(**_A ).images __A = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __A = np.array([0.5_29_11, 0.6_00_04, 0.4_92_29, 0.4_98_05, 0.5_45_02, 0.5_06_80, 0.4_77_77, 0.4_10_28, 0.4_53_04] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self : List[str] ): '''simple docstring''' __A = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint, provider='''CPUExecutionProvider''' ) __A = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_A ) __A = self.get_dummy_inputs() __A = pipe(**_A ).images __A = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __A = np.array([0.5_29_11, 0.6_00_04, 0.4_92_29, 0.4_98_05, 0.5_45_02, 0.5_06_80, 0.4_77_77, 0.4_10_28, 0.4_53_04] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _SCREAMING_SNAKE_CASE ( self : Tuple ): '''simple docstring''' __A = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint, provider='''CPUExecutionProvider''' ) __A = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_A ) __A = self.get_dummy_inputs() __A = pipe(**_A ).images __A = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) __A = np.array([0.6_53_31, 0.5_82_77, 0.4_82_04, 0.5_60_59, 0.5_36_65, 0.5_62_35, 0.5_09_69, 0.4_00_09, 0.4_65_52] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class snake_case ( unittest.TestCase ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self : Dict ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _SCREAMING_SNAKE_CASE ( self : List[str] ): '''simple docstring''' __A = ort.SessionOptions() __A = False return options def _SCREAMING_SNAKE_CASE ( self : List[Any] ): '''simple docstring''' __A = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) __A = init_image.resize((7_68, 5_12) ) # using the PNDM scheduler by default __A = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''', revision='''onnx''', safety_checker=_A, feature_extractor=_A, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=_A ) __A = 'A fantasy landscape, trending on artstation' __A = np.random.RandomState(0 ) __A = pipe( prompt=_A, image=_A, strength=0.75, guidance_scale=7.5, num_inference_steps=10, generator=_A, output_type='''np''', ) __A = output.images __A = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) __A = np.array([0.49_09, 0.50_59, 0.53_72, 0.46_23, 0.48_76, 0.50_49, 0.48_20, 0.49_56, 0.50_19] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def _SCREAMING_SNAKE_CASE ( self : Optional[int] ): '''simple docstring''' __A = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) __A = init_image.resize((7_68, 5_12) ) __A = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''', subfolder='''scheduler''', revision='''onnx''' ) __A = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''', revision='''onnx''', scheduler=_A, safety_checker=_A, feature_extractor=_A, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=_A ) __A = 'A fantasy landscape, trending on artstation' __A = np.random.RandomState(0 ) __A = pipe( prompt=_A, image=_A, strength=0.75, guidance_scale=7.5, num_inference_steps=20, generator=_A, output_type='''np''', ) __A = output.images __A = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) __A = np.array([0.80_43, 0.9_26, 0.95_81, 0.81_19, 0.89_54, 0.9_13, 0.72_09, 0.74_63, 0.74_31] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2
266
def SCREAMING_SNAKE_CASE__ ( __a , __a ): if density <= 0: raise ValueError('Impossible fluid density' ) if bulk_modulus <= 0: raise ValueError('Impossible bulk modulus' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
327
0
def __UpperCamelCase ( lowercase__ : str , lowercase__ : str ) -> list: '''simple docstring''' lowerCAmelCase_ : str = len(lowercase__ ) lowerCAmelCase_ : Dict = [] for i in range(len(lowercase__ ) - pat_len + 1 ): lowerCAmelCase_ : str = True for j in range(lowercase__ ): if s[i + j] != pattern[j]: lowerCAmelCase_ : str = False break if match_found: position.append(lowercase__ ) return position if __name__ == "__main__": assert naive_pattern_search('ABCDEFG', 'DE') == [3] print(naive_pattern_search('ABAAABCDBBABCDDEBCABC', 'ABC'))
28
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor __UpperCAmelCase = logging.get_logger(__name__) class __a ( __UpperCamelCase ): def __init__( self : Union[str, Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : Dict ): warnings.warn( """The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use GLPNImageProcessor instead.""" , UpperCAmelCase , ) super().__init__(*UpperCAmelCase , **UpperCAmelCase )
28
1
"""simple docstring""" import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import evaluate import numpy as np from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/text-classification/requirements.txt') _a : Tuple = logging.getLogger(__name__) @dataclass class __A : _UpperCamelCase : Optional[int] = field( default=128 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) _UpperCamelCase : bool = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Overwrite the cached preprocessed datasets or not."} ) _UpperCamelCase : bool = field( default=SCREAMING_SNAKE_CASE_ , metadata={ "help": ( "Whether to pad all samples to `max_seq_length`. " "If False, will pad the samples dynamically when batching to the maximum length in the batch." ) } , ) _UpperCamelCase : Optional[int] = field( default=SCREAMING_SNAKE_CASE_ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) _UpperCamelCase : Optional[int] = field( default=SCREAMING_SNAKE_CASE_ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) _UpperCamelCase : Optional[int] = field( default=SCREAMING_SNAKE_CASE_ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of prediction examples to this " "value if set." ) } , ) @dataclass class __A : _UpperCamelCase : str = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) _UpperCamelCase : str = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Evaluation language. Also train language if `train_language` is set to None."} ) _UpperCamelCase : Optional[str] = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Train language if it is different from the evaluation language."} ) _UpperCamelCase : Optional[str] = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) _UpperCamelCase : Optional[str] = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) _UpperCamelCase : Optional[str] = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) _UpperCamelCase : Optional[bool] = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "arg to indicate if tokenizer should do lower case in AutoTokenizer.from_pretrained()"} , ) _UpperCamelCase : bool = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."} , ) _UpperCamelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) _UpperCamelCase : bool = field( default=SCREAMING_SNAKE_CASE_ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) _UpperCamelCase : bool = field( default=SCREAMING_SNAKE_CASE_ , metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} , ) def SCREAMING_SNAKE_CASE ( ) -> Union[str, Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowerCAmelCase : str = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Tuple = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_xnli""" ,_lowerCamelCase ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" ,datefmt="""%m/%d/%Y %H:%M:%S""" ,handlers=[logging.StreamHandler(sys.stdout )] ,) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _lowerCAmelCase : int = training_args.get_process_log_level() logger.setLevel(_lowerCamelCase ) datasets.utils.logging.set_verbosity(_lowerCamelCase ) transformers.utils.logging.set_verbosity(_lowerCamelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}" + f"distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}" ) logger.info(f"Training/evaluation parameters {training_args}" ) # Detecting last checkpoint. _lowerCAmelCase : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowerCAmelCase : int = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"Output directory ({training_args.output_dir}) already exists and is not empty. " """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None: logger.info( f"Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change " """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Set seed before initializing model. set_seed(training_args.seed ) # In distributed training, the load_dataset function guarantees that only one local process can concurrently # download the dataset. # Downloading and loading xnli dataset from the hub. if training_args.do_train: if model_args.train_language is None: _lowerCAmelCase : Any = load_dataset( """xnli""" ,model_args.language ,split="""train""" ,cache_dir=model_args.cache_dir ,use_auth_token=True if model_args.use_auth_token else None ,) else: _lowerCAmelCase : Dict = load_dataset( """xnli""" ,model_args.train_language ,split="""train""" ,cache_dir=model_args.cache_dir ,use_auth_token=True if model_args.use_auth_token else None ,) _lowerCAmelCase : Tuple = train_dataset.features["""label"""].names if training_args.do_eval: _lowerCAmelCase : List[str] = load_dataset( """xnli""" ,model_args.language ,split="""validation""" ,cache_dir=model_args.cache_dir ,use_auth_token=True if model_args.use_auth_token else None ,) _lowerCAmelCase : List[Any] = eval_dataset.features["""label"""].names if training_args.do_predict: _lowerCAmelCase : Union[str, Any] = load_dataset( """xnli""" ,model_args.language ,split="""test""" ,cache_dir=model_args.cache_dir ,use_auth_token=True if model_args.use_auth_token else None ,) _lowerCAmelCase : Union[str, Any] = predict_dataset.features["""label"""].names # Labels _lowerCAmelCase : Optional[Any] = len(_lowerCamelCase ) # Load pretrained model and tokenizer # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowerCAmelCase : Optional[int] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path ,num_labels=_lowerCamelCase ,idalabel={str(_lowerCamelCase ): label for i, label in enumerate(_lowerCamelCase )} ,labelaid={label: i for i, label in enumerate(_lowerCamelCase )} ,finetuning_task="""xnli""" ,cache_dir=model_args.cache_dir ,revision=model_args.model_revision ,use_auth_token=True if model_args.use_auth_token else None ,) _lowerCAmelCase : List[Any] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path ,do_lower_case=model_args.do_lower_case ,cache_dir=model_args.cache_dir ,use_fast=model_args.use_fast_tokenizer ,revision=model_args.model_revision ,use_auth_token=True if model_args.use_auth_token else None ,) _lowerCAmelCase : Any = AutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path ,from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) ,config=_lowerCamelCase ,cache_dir=model_args.cache_dir ,revision=model_args.model_revision ,use_auth_token=True if model_args.use_auth_token else None ,ignore_mismatched_sizes=model_args.ignore_mismatched_sizes ,) # Preprocessing the datasets # Padding strategy if data_args.pad_to_max_length: _lowerCAmelCase : Optional[int] = """max_length""" else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch _lowerCAmelCase : List[str] = False def preprocess_function(_lowerCamelCase : Dict ): # Tokenize the texts return tokenizer( examples["""premise"""] ,examples["""hypothesis"""] ,padding=_lowerCamelCase ,max_length=data_args.max_seq_length ,truncation=_lowerCamelCase ,) if training_args.do_train: if data_args.max_train_samples is not None: _lowerCAmelCase : Any = min(len(_lowerCamelCase ) ,data_args.max_train_samples ) _lowerCAmelCase : str = train_dataset.select(range(_lowerCamelCase ) ) with training_args.main_process_first(desc="""train dataset map pre-processing""" ): _lowerCAmelCase : Optional[int] = train_dataset.map( _lowerCamelCase ,batched=_lowerCamelCase ,load_from_cache_file=not data_args.overwrite_cache ,desc="""Running tokenizer on train dataset""" ,) # Log a few random samples from the training set: for index in random.sample(range(len(_lowerCamelCase ) ) ,3 ): logger.info(f"Sample {index} of the training set: {train_dataset[index]}." ) if training_args.do_eval: if data_args.max_eval_samples is not None: _lowerCAmelCase : Optional[Any] = min(len(_lowerCamelCase ) ,data_args.max_eval_samples ) _lowerCAmelCase : str = eval_dataset.select(range(_lowerCamelCase ) ) with training_args.main_process_first(desc="""validation dataset map pre-processing""" ): _lowerCAmelCase : List[str] = eval_dataset.map( _lowerCamelCase ,batched=_lowerCamelCase ,load_from_cache_file=not data_args.overwrite_cache ,desc="""Running tokenizer on validation dataset""" ,) if training_args.do_predict: if data_args.max_predict_samples is not None: _lowerCAmelCase : Optional[Any] = min(len(_lowerCamelCase ) ,data_args.max_predict_samples ) _lowerCAmelCase : int = predict_dataset.select(range(_lowerCamelCase ) ) with training_args.main_process_first(desc="""prediction dataset map pre-processing""" ): _lowerCAmelCase : Tuple = predict_dataset.map( _lowerCamelCase ,batched=_lowerCamelCase ,load_from_cache_file=not data_args.overwrite_cache ,desc="""Running tokenizer on prediction dataset""" ,) # Get the metric function _lowerCAmelCase : Optional[Any] = evaluate.load("""xnli""" ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(_lowerCamelCase : EvalPrediction ): _lowerCAmelCase : str = p.predictions[0] if isinstance(p.predictions ,_lowerCamelCase ) else p.predictions _lowerCAmelCase : str = np.argmax(_lowerCamelCase ,axis=1 ) return metric.compute(predictions=_lowerCamelCase ,references=p.label_ids ) # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: _lowerCAmelCase : List[str] = default_data_collator elif training_args.fpaa: _lowerCAmelCase : List[str] = DataCollatorWithPadding(_lowerCamelCase ,pad_to_multiple_of=8 ) else: _lowerCAmelCase : List[Any] = None # Initialize our Trainer _lowerCAmelCase : Union[str, Any] = Trainer( model=_lowerCamelCase ,args=_lowerCamelCase ,train_dataset=train_dataset if training_args.do_train else None ,eval_dataset=eval_dataset if training_args.do_eval else None ,compute_metrics=_lowerCamelCase ,tokenizer=_lowerCamelCase ,data_collator=_lowerCamelCase ,) # Training if training_args.do_train: _lowerCAmelCase : Union[str, Any] = None if training_args.resume_from_checkpoint is not None: _lowerCAmelCase : Union[str, Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowerCAmelCase : int = last_checkpoint _lowerCAmelCase : List[str] = trainer.train(resume_from_checkpoint=_lowerCamelCase ) _lowerCAmelCase : Dict = train_result.metrics _lowerCAmelCase : Dict = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(_lowerCamelCase ) ) _lowerCAmelCase : List[str] = min(_lowerCamelCase ,len(_lowerCamelCase ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics("""train""" ,_lowerCamelCase ) trainer.save_metrics("""train""" ,_lowerCamelCase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info("""*** Evaluate ***""" ) _lowerCAmelCase : Tuple = trainer.evaluate(eval_dataset=_lowerCamelCase ) _lowerCAmelCase : Optional[Any] = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(_lowerCamelCase ) _lowerCAmelCase : Any = min(_lowerCamelCase ,len(_lowerCamelCase ) ) trainer.log_metrics("""eval""" ,_lowerCamelCase ) trainer.save_metrics("""eval""" ,_lowerCamelCase ) # Prediction if training_args.do_predict: logger.info("""*** Predict ***""" ) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = trainer.predict(_lowerCamelCase ,metric_key_prefix="""predict""" ) _lowerCAmelCase : Optional[int] = ( data_args.max_predict_samples if data_args.max_predict_samples is not None else len(_lowerCamelCase ) ) _lowerCAmelCase : Dict = min(_lowerCamelCase ,len(_lowerCamelCase ) ) trainer.log_metrics("""predict""" ,_lowerCamelCase ) trainer.save_metrics("""predict""" ,_lowerCamelCase ) _lowerCAmelCase : Optional[int] = np.argmax(_lowerCamelCase ,axis=1 ) _lowerCAmelCase : Optional[Any] = os.path.join(training_args.output_dir ,"""predictions.txt""" ) if trainer.is_world_process_zero(): with open(_lowerCamelCase ,"""w""" ) as writer: writer.write("""index\tprediction\n""" ) for index, item in enumerate(_lowerCamelCase ): _lowerCAmelCase : Tuple = label_list[item] writer.write(f"{index}\t{item}\n" ) if __name__ == "__main__": main()
44
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : Optional[Any] = "naver-clova-ix/donut-base-finetuned-docvqa" _UpperCamelCase : Dict = ( "This is a tool that answers a question about an document (pdf). It takes an input named `document` which " "should be the document containing the information, as well as a `question` that is the question about the " "document. It returns a text that contains the answer to the question." ) _UpperCamelCase : Optional[int] = "document_qa" _UpperCamelCase : Any = AutoProcessor _UpperCamelCase : Union[str, Any] = VisionEncoderDecoderModel _UpperCamelCase : Union[str, Any] = ["image", "text"] _UpperCamelCase : List[str] = ["text"] def __init__( self , *a__ , **a__ ): if not is_vision_available(): raise ValueError("""Pillow must be installed to use the DocumentQuestionAnsweringTool.""" ) super().__init__(*a__ , **a__ ) def __A ( self , a__ , a__ ): _lowerCAmelCase : Optional[int] = """<s_docvqa><s_question>{user_input}</s_question><s_answer>""" _lowerCAmelCase : Dict = task_prompt.replace("""{user_input}""" , a__ ) _lowerCAmelCase : str = self.pre_processor.tokenizer( a__ , add_special_tokens=a__ , return_tensors="""pt""" ).input_ids _lowerCAmelCase : Dict = self.pre_processor(a__ , return_tensors="""pt""" ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def __A ( self , a__ ): return self.model.generate( inputs["""pixel_values"""].to(self.device ) , decoder_input_ids=inputs["""decoder_input_ids"""].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=a__ , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=a__ , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=a__ , ).sequences def __A ( self , a__ ): _lowerCAmelCase : Tuple = self.pre_processor.batch_decode(a__ )[0] _lowerCAmelCase : int = sequence.replace(self.pre_processor.tokenizer.eos_token , """""" ) _lowerCAmelCase : List[str] = sequence.replace(self.pre_processor.tokenizer.pad_token , """""" ) _lowerCAmelCase : List[str] = re.sub(r"""<.*?>""" , """""" , a__ , count=1 ).strip() # remove first task start token _lowerCAmelCase : List[str] = self.pre_processor.tokenajson(a__ ) return sequence["answer"]
44
1
'''simple docstring''' import numpy class lowerCAmelCase_: '''simple docstring''' def __init__( self ,__UpperCAmelCase ,__UpperCAmelCase ) -> None: lowerCAmelCase__ : List[Any] = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. lowerCAmelCase__ : Any = numpy.random.rand( self.input_array.shape[1] ,4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. lowerCAmelCase__ : List[str] = numpy.random.rand( 4 ,3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. lowerCAmelCase__ : Optional[int] = numpy.random.rand(3 ,1 ) # Real output values provided. lowerCAmelCase__ : Union[str, Any] = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. lowerCAmelCase__ : Tuple = numpy.zeros(output_array.shape ) def UpperCAmelCase_ ( self ) -> numpy.ndarray: lowerCAmelCase__ : Any = sigmoid( numpy.dot(self.input_array ,self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer ,self.first_hidden_layer_and_second_hidden_layer_weights ,) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. lowerCAmelCase__ : List[str] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer ,self.second_hidden_layer_and_output_layer_weights ,) ) return self.layer_between_second_hidden_layer_and_output def UpperCAmelCase_ ( self ) -> None: lowerCAmelCase__ : List[Any] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T ,2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) ,) lowerCAmelCase__ : Dict = numpy.dot( self.layer_between_input_and_first_hidden_layer.T ,numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) ,self.second_hidden_layer_and_output_layer_weights.T ,) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) ,) lowerCAmelCase__ : List[Any] = numpy.dot( self.input_array.T ,numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) ,self.second_hidden_layer_and_output_layer_weights.T ,) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) ,self.first_hidden_layer_and_second_hidden_layer_weights.T ,) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) ,) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> None: for iteration in range(1 ,iterations + 1 ): lowerCAmelCase__ : Dict = self.feedforward() self.back_propagation() if give_loss: lowerCAmelCase__ : Optional[int] = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> int: lowerCAmelCase__ : str = input_arr lowerCAmelCase__ : Union[str, Any] = sigmoid( numpy.dot(self.array ,self.input_layer_and_first_hidden_layer_weights ) ) lowerCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer ,self.first_hidden_layer_and_second_hidden_layer_weights ,) ) lowerCAmelCase__ : Dict = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer ,self.second_hidden_layer_and_output_layer_weights ,) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" return 1 / (1 + numpy.exp(-value )) def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" return (value) * (1 - (value)) def _SCREAMING_SNAKE_CASE ( ): """simple docstring""" lowerCAmelCase__ : Any = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. lowerCAmelCase__ : Union[str, Any] = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. lowerCAmelCase__ : Dict = TwoHiddenLayerNeuralNetwork( input_array=UpperCamelCase , output_array=UpperCamelCase ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=UpperCamelCase , iterations=10 , give_loss=UpperCamelCase ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
361
'''simple docstring''' from typing import Any class lowerCAmelCase_: '''simple docstring''' def __init__( self ,__UpperCAmelCase ) -> Any: lowerCAmelCase__ : Optional[Any] = data lowerCAmelCase__ : str = None def __repr__( self ) -> str: return F"""Node({self.data})""" class lowerCAmelCase_: '''simple docstring''' def __init__( self ) -> List[Any]: lowerCAmelCase__ : List[str] = None def __iter__( self ) -> Any: lowerCAmelCase__ : int = self.head while node: yield node.data lowerCAmelCase__ : Optional[Any] = node.next def __len__( self ) -> int: return sum(1 for _ in self ) def __repr__( self ) -> str: return "->".join([str(__UpperCAmelCase ) for item in self] ) def __getitem__( self ,__UpperCAmelCase ) -> Any: if not 0 <= index < len(self ): raise ValueError("""list index out of range.""" ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self ,__UpperCAmelCase ,__UpperCAmelCase ) -> None: if not 0 <= index < len(self ): raise ValueError("""list index out of range.""" ) lowerCAmelCase__ : List[Any] = self.head for _ in range(__UpperCAmelCase ): lowerCAmelCase__ : List[str] = current.next lowerCAmelCase__ : Dict = data def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> None: self.insert_nth(len(self ) ,__UpperCAmelCase ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> None: self.insert_nth(0 ,__UpperCAmelCase ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ) -> None: if not 0 <= index <= len(self ): raise IndexError("""list index out of range""" ) lowerCAmelCase__ : str = Node(__UpperCAmelCase ) if self.head is None: lowerCAmelCase__ : List[Any] = new_node elif index == 0: lowerCAmelCase__ : Dict = self.head # link new_node to head lowerCAmelCase__ : int = new_node else: lowerCAmelCase__ : Any = self.head for _ in range(index - 1 ): lowerCAmelCase__ : Tuple = temp.next lowerCAmelCase__ : str = temp.next lowerCAmelCase__ : Optional[int] = new_node def UpperCAmelCase_ ( self ) -> None: # print every node data print(self ) def UpperCAmelCase_ ( self ) -> Any: return self.delete_nth(0 ) def UpperCAmelCase_ ( self ) -> Any: # delete from tail return self.delete_nth(len(self ) - 1 ) def UpperCAmelCase_ ( self ,__UpperCAmelCase = 0 ) -> Any: if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError("""List index out of range.""" ) lowerCAmelCase__ : List[Any] = self.head # default first node if index == 0: lowerCAmelCase__ : str = self.head.next else: lowerCAmelCase__ : int = self.head for _ in range(index - 1 ): lowerCAmelCase__ : Union[str, Any] = temp.next lowerCAmelCase__ : Optional[Any] = temp.next lowerCAmelCase__ : Any = temp.next.next return delete_node.data def UpperCAmelCase_ ( self ) -> bool: return self.head is None def UpperCAmelCase_ ( self ) -> None: lowerCAmelCase__ : Optional[int] = None lowerCAmelCase__ : str = self.head while current: # Store the current node's next node. lowerCAmelCase__ : Optional[Any] = current.next # Make the current node's next point backwards lowerCAmelCase__ : Optional[int] = prev # Make the previous node be the current node lowerCAmelCase__ : Optional[Any] = current # Make the current node the next node (to progress iteration) lowerCAmelCase__ : Optional[int] = next_node # Return prev in order to put the head at the end lowerCAmelCase__ : List[str] = prev def _SCREAMING_SNAKE_CASE ( ): """simple docstring""" lowerCAmelCase__ : Dict = LinkedList() assert linked_list.is_empty() is True assert str(UpperCamelCase ) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10 ): assert len(UpperCamelCase ) == i linked_list.insert_nth(UpperCamelCase , i + 1 ) assert str(UpperCamelCase ) == "->".join(str(UpperCamelCase ) for i in range(1 , 11 ) ) linked_list.insert_head(0 ) linked_list.insert_tail(11 ) assert str(UpperCamelCase ) == "->".join(str(UpperCamelCase ) for i in range(0 , 12 ) ) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9 ) == 10 assert linked_list.delete_tail() == 11 assert len(UpperCamelCase ) == 9 assert str(UpperCamelCase ) == "->".join(str(UpperCamelCase ) for i in range(1 , 10 ) ) assert all(linked_list[i] == i + 1 for i in range(0 , 9 ) ) is True for i in range(0 , 9 ): lowerCAmelCase__ : str = -i assert all(linked_list[i] == -i for i in range(0 , 9 ) ) is True linked_list.reverse() assert str(UpperCamelCase ) == "->".join(str(UpperCamelCase ) for i in range(-8 , 1 ) ) def _SCREAMING_SNAKE_CASE ( ): """simple docstring""" lowerCAmelCase__ : Any = [ -9, 100, Node(77345112 ), """dlrow olleH""", 7, 5555, 0, -192.5_5555, """Hello, world!""", 77.9, Node(10 ), None, None, 12.20, ] lowerCAmelCase__ : List[Any] = LinkedList() for i in test_input: linked_list.insert_tail(UpperCamelCase ) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(UpperCamelCase ) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head lowerCAmelCase__ : List[str] = linked_list.delete_head() assert result == -9 assert ( str(UpperCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail lowerCAmelCase__ : Union[str, Any] = linked_list.delete_tail() assert result == 12.2 assert ( str(UpperCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list lowerCAmelCase__ : str = linked_list.delete_nth(10 ) assert result is None assert ( str(UpperCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node("""Hello again, world!""" ) ) assert ( str(UpperCamelCase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(UpperCamelCase ) assert ( str(UpperCamelCase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(UpperCamelCase ) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def _SCREAMING_SNAKE_CASE ( ): """simple docstring""" from doctest import testmod testmod() lowerCAmelCase__ : str = LinkedList() linked_list.insert_head(input("""Inserting 1st at head """ ).strip() ) linked_list.insert_head(input("""Inserting 2nd at head """ ).strip() ) print("""\nPrint list:""" ) linked_list.print_list() linked_list.insert_tail(input("""\nInserting 1st at tail """ ).strip() ) linked_list.insert_tail(input("""Inserting 2nd at tail """ ).strip() ) print("""\nPrint list:""" ) linked_list.print_list() print("""\nDelete head""" ) linked_list.delete_head() print("""Delete tail""" ) linked_list.delete_tail() print("""\nPrint list:""" ) linked_list.print_list() print("""\nReverse linked list""" ) linked_list.reverse() print("""\nPrint list:""" ) linked_list.print_list() print("""\nString representation of linked list:""" ) print(UpperCamelCase ) print("""\nReading/changing Node data using indexing:""" ) print(f"""Element at Position 1: {linked_list[1]}""" ) lowerCAmelCase__ : Dict = input("""Enter New Value: """ ).strip() print("""New list:""" ) print(UpperCamelCase ) print(f"""length of linked_list is : {len(UpperCamelCase )}""" ) if __name__ == "__main__": main()
184
0
import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCAmelCase_ ( a__ ): def snake_case_ ( self ) -> Any: UpperCamelCase : Optional[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(SCREAMING_SNAKE_CASE_, 'embed_dim' ) ) self.parent.assertTrue(hasattr(SCREAMING_SNAKE_CASE_, 'num_heads' ) ) class lowerCAmelCase_ : def __init__( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=13, SCREAMING_SNAKE_CASE_=64, SCREAMING_SNAKE_CASE_=3, SCREAMING_SNAKE_CASE_=[16, 48, 96], SCREAMING_SNAKE_CASE_=[1, 3, 6], SCREAMING_SNAKE_CASE_=[1, 2, 10], SCREAMING_SNAKE_CASE_=[7, 3, 3], SCREAMING_SNAKE_CASE_=[4, 2, 2], SCREAMING_SNAKE_CASE_=[2, 1, 1], SCREAMING_SNAKE_CASE_=[2, 2, 2], SCREAMING_SNAKE_CASE_=[False, False, True], SCREAMING_SNAKE_CASE_=[0.0, 0.0, 0.0], SCREAMING_SNAKE_CASE_=0.02, SCREAMING_SNAKE_CASE_=1e-12, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=2, ) -> Optional[int]: UpperCamelCase : Optional[Any] = parent UpperCamelCase : Tuple = batch_size UpperCamelCase : List[Any] = image_size UpperCamelCase : List[str] = patch_sizes UpperCamelCase : str = patch_stride UpperCamelCase : Tuple = patch_padding UpperCamelCase : Dict = is_training UpperCamelCase : Any = use_labels UpperCamelCase : Union[str, Any] = num_labels UpperCamelCase : Optional[int] = num_channels UpperCamelCase : Tuple = embed_dim UpperCamelCase : List[Any] = num_heads UpperCamelCase : Any = stride_kv UpperCamelCase : Optional[Any] = depth UpperCamelCase : Optional[Any] = cls_token UpperCamelCase : Any = attention_drop_rate UpperCamelCase : int = initializer_range UpperCamelCase : Optional[Any] = layer_norm_eps def snake_case_ ( self ) -> Optional[Any]: UpperCamelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase : Union[str, Any] = None if self.use_labels: UpperCamelCase : int = ids_tensor([self.batch_size], self.num_labels ) UpperCamelCase : Any = self.get_config() return config, pixel_values, labels def snake_case_ ( self ) -> Any: return CvtConfig( image_size=self.image_size, num_labels=self.num_labels, num_channels=self.num_channels, embed_dim=self.embed_dim, num_heads=self.num_heads, patch_sizes=self.patch_sizes, patch_padding=self.patch_padding, patch_stride=self.patch_stride, stride_kv=self.stride_kv, depth=self.depth, cls_token=self.cls_token, attention_drop_rate=self.attention_drop_rate, initializer_range=self.initializer_range, ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> int: UpperCamelCase : str = CvtModel(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCamelCase : Optional[Any] = model(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Dict = (self.image_size, self.image_size) UpperCamelCase , UpperCamelCase : Optional[int] = image_size[0], image_size[1] for i in range(len(self.depth ) ): UpperCamelCase : str = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) UpperCamelCase : Optional[int] = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.embed_dim[-1], height, width) ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> str: UpperCamelCase : str = self.num_labels UpperCamelCase : Optional[int] = CvtForImageClassification(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCamelCase : List[str] = model(SCREAMING_SNAKE_CASE_, labels=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def snake_case_ ( self ) -> Optional[int]: UpperCamelCase : List[str] = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase : List[str] = config_and_inputs UpperCamelCase : Optional[int] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase_ ( a__ , a__ , unittest.TestCase ): UpperCAmelCase__ : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () UpperCAmelCase__ : str = ( {"feature-extraction": CvtModel, "image-classification": CvtForImageClassification} if is_torch_available() else {} ) UpperCAmelCase__ : str = False UpperCAmelCase__ : Any = False UpperCAmelCase__ : Any = False UpperCAmelCase__ : Dict = False UpperCAmelCase__ : Dict = False def snake_case_ ( self ) -> Union[str, Any]: UpperCamelCase : str = CvtModelTester(self ) UpperCamelCase : List[str] = ConfigTester(self, config_class=SCREAMING_SNAKE_CASE_, has_text_modality=SCREAMING_SNAKE_CASE_, hidden_size=37 ) def snake_case_ ( self ) -> Union[str, Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case_ ( self ) -> Tuple: return @unittest.skip(reason='Cvt does not output attentions' ) def snake_case_ ( self ) -> Dict: pass @unittest.skip(reason='Cvt does not use inputs_embeds' ) def snake_case_ ( self ) -> int: pass @unittest.skip(reason='Cvt does not support input and output embeddings' ) def snake_case_ ( self ) -> Tuple: pass def snake_case_ ( self ) -> List[Any]: UpperCamelCase , UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : int = model_class(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase : int = [*signature.parameters.keys()] UpperCamelCase : str = ['pixel_values'] self.assertListEqual(arg_names[:1], SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self ) -> int: UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self ) -> int: def check_hidden_states_output(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): UpperCamelCase : int = model_class(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() with torch.no_grad(): UpperCamelCase : List[str] = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) ) UpperCamelCase : str = outputs.hidden_states UpperCamelCase : Dict = len(self.model_tester.depth ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ), SCREAMING_SNAKE_CASE_ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ), [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ], ) UpperCamelCase , UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Optional[Any] = True check_hidden_states_output(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase : int = True check_hidden_states_output(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self ) -> Optional[Any]: UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE_ ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def snake_case_ ( self ) -> Dict: pass @slow def snake_case_ ( self ) -> Union[str, Any]: for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase : Optional[Any] = CvtModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) def UpperCamelCase ( ) -> Tuple: UpperCamelCase : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @cached_property def snake_case_ ( self ) -> int: return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def snake_case_ ( self ) -> Union[str, Any]: UpperCamelCase : Tuple = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : List[str] = self.default_image_processor UpperCamelCase : int = prepare_img() UpperCamelCase : Optional[int] = image_processor(images=SCREAMING_SNAKE_CASE_, return_tensors='pt' ).to(SCREAMING_SNAKE_CASE_ ) # forward pass with torch.no_grad(): UpperCamelCase : str = model(**SCREAMING_SNAKE_CASE_ ) # verify the logits UpperCamelCase : List[str] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape, SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Dict = torch.tensor([0.92_85, 0.90_15, -0.31_50] ).to(SCREAMING_SNAKE_CASE_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3], SCREAMING_SNAKE_CASE_, atol=1e-4 ) )
119
def UpperCamelCase ( snake_case__ : int , snake_case__ : int ) -> int: return int(input_a == input_a == 0 ) def UpperCamelCase ( ) -> None: print('Truth Table of NOR Gate:' ) print('| Input 1 | Input 2 | Output |' ) print(F"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) print(F"""| 0 | 1 | {nor_gate(0 , 1 )} |""" ) print(F"""| 1 | 0 | {nor_gate(1 , 0 )} |""" ) print(F"""| 1 | 1 | {nor_gate(1 , 1 )} |""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
119
1
'''simple docstring''' def a__ ( _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : int ) -> float: """simple docstring""" if digit_amount > 0: return round(number - int(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) return number - int(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.3_45, 1)) print(decimal_isolate(35.3_45, 2)) print(decimal_isolate(35.3_45, 3)) print(decimal_isolate(-14.7_89, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.1_23, 1)) print(decimal_isolate(-14.1_23, 2)) print(decimal_isolate(-14.1_23, 3))
67
'''simple docstring''' from __future__ import annotations from collections.abc import Sequence from typing import Literal def a__ ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : str ) -> str | Literal[False]: """simple docstring""" UpperCAmelCase_ : Tuple = list(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Dict = list(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = 0 for i in range(len(_SCREAMING_SNAKE_CASE ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase_ : List[str] = "_" if count > 1: return False else: return "".join(_SCREAMING_SNAKE_CASE ) def a__ ( _SCREAMING_SNAKE_CASE : list[str] ) -> list[str]: """simple docstring""" UpperCAmelCase_ : List[str] = [] while True: UpperCAmelCase_ : Any = ["$"] * len(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = [] for i in range(len(_SCREAMING_SNAKE_CASE ) ): for j in range(i + 1 , len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : Optional[int] = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase_ : Union[str, Any] = "*" UpperCAmelCase_ : List[Any] = "*" temp.append("X" ) for i in range(len(_SCREAMING_SNAKE_CASE ) ): if checka[i] == "$": pi.append(binary[i] ) if len(_SCREAMING_SNAKE_CASE ) == 0: return pi UpperCAmelCase_ : str = list(set(_SCREAMING_SNAKE_CASE ) ) def a__ ( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Sequence[float] ) -> list[str]: """simple docstring""" UpperCAmelCase_ : Optional[Any] = [] for minterm in minterms: UpperCAmelCase_ : Optional[Any] = "" for _ in range(_SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : Dict = str(minterm % 2 ) + string minterm //= 2 temp.append(_SCREAMING_SNAKE_CASE ) return temp def a__ ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : int ) -> bool: """simple docstring""" UpperCAmelCase_ : Tuple = list(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = list(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[Any] = 0 for i in range(len(_SCREAMING_SNAKE_CASE ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def a__ ( _SCREAMING_SNAKE_CASE : list[list[int]] , _SCREAMING_SNAKE_CASE : list[str] ) -> list[str]: """simple docstring""" UpperCAmelCase_ : Optional[int] = [] UpperCAmelCase_ : str = [0] * len(_SCREAMING_SNAKE_CASE ) for i in range(len(chart[0] ) ): UpperCAmelCase_ : Optional[Any] = 0 UpperCAmelCase_ : Optional[int] = -1 for j in range(len(_SCREAMING_SNAKE_CASE ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase_ : Any = j if count == 1: UpperCAmelCase_ : Any = 1 for i in range(len(_SCREAMING_SNAKE_CASE ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : int = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ : List[str] = -1 UpperCAmelCase_ : int = 0 for i in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : str = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase_ : List[str] = count_n UpperCAmelCase_ : Dict = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : int = 0 def a__ ( _SCREAMING_SNAKE_CASE : list[str] , _SCREAMING_SNAKE_CASE : list[str] ) -> list[list[int]]: """simple docstring""" UpperCAmelCase_ : Dict = [[0 for x in range(len(_SCREAMING_SNAKE_CASE ) )] for x in range(len(_SCREAMING_SNAKE_CASE ) )] for i in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCAmelCase_ : Optional[int] = prime_implicants[i].count("_" ) for j in range(len(_SCREAMING_SNAKE_CASE ) ): if is_for_table(prime_implicants[i] , binary[j] , _SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : str = 1 return chart def a__ ( ) -> None: """simple docstring""" UpperCAmelCase_ : List[Any] = int(input("Enter the no. of variables\n" ) ) UpperCAmelCase_ : Tuple = [ float(_SCREAMING_SNAKE_CASE ) for x in input( "Enter the decimal representation of Minterms 'Spaces Separated'\n" ).split() ] UpperCAmelCase_ : int = decimal_to_binary(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[str] = check(_SCREAMING_SNAKE_CASE ) print("Prime Implicants are:" ) print(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = prime_implicant_chart(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[str] = selection(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print("Essential Prime Implicants are:" ) print(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() main()
67
1
"""simple docstring""" import re def snake_case_ ( A_ : str ): '''simple docstring''' _lowerCamelCase : int = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A_, A_ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
72
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __snake_case = logging.get_logger(__name__) __snake_case = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class lowercase__ ( _UpperCAmelCase ): def __init__( self : str , UpperCAmelCase_ : int=None , UpperCAmelCase_ : List[str]=None , *UpperCAmelCase_ : Union[str, Any] , **UpperCAmelCase_ : Optional[Any] ): super().__init__(*UpperCAmelCase_ , **UpperCAmelCase_ ) if config is None: assert isinstance(self.model , UpperCAmelCase_ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F' {self.model.__class__}' ) SCREAMING_SNAKE_CASE__ = self.model.config else: SCREAMING_SNAKE_CASE__ = config SCREAMING_SNAKE_CASE__ = data_args SCREAMING_SNAKE_CASE__ = self.config.tgt_vocab_size if isinstance(self.config , UpperCAmelCase_ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F'The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for' ' padding..' ) if self.args.label_smoothing == 0: SCREAMING_SNAKE_CASE__ = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss SCREAMING_SNAKE_CASE__ = label_smoothed_nll_loss def A_ ( self : Tuple , UpperCAmelCase_ : int ): if self.optimizer is None: SCREAMING_SNAKE_CASE__ = ['bias', 'LayerNorm.weight'] SCREAMING_SNAKE_CASE__ = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] SCREAMING_SNAKE_CASE__ = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: SCREAMING_SNAKE_CASE__ = Adafactor SCREAMING_SNAKE_CASE__ = {'scale_parameter': False, 'relative_step': False} else: SCREAMING_SNAKE_CASE__ = AdamW SCREAMING_SNAKE_CASE__ = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } SCREAMING_SNAKE_CASE__ = self.args.learning_rate if self.sharded_ddp: SCREAMING_SNAKE_CASE__ = OSS( params=UpperCAmelCase_ , optim=UpperCAmelCase_ , **UpperCAmelCase_ , ) else: SCREAMING_SNAKE_CASE__ = optimizer_cls(UpperCAmelCase_ , **UpperCAmelCase_ ) if self.lr_scheduler is None: SCREAMING_SNAKE_CASE__ = self._get_lr_scheduler(UpperCAmelCase_ ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def A_ ( self : str , UpperCAmelCase_ : List[Any] ): SCREAMING_SNAKE_CASE__ = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": SCREAMING_SNAKE_CASE__ = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": SCREAMING_SNAKE_CASE__ = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: SCREAMING_SNAKE_CASE__ = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=UpperCAmelCase_ ) return scheduler def A_ ( self : List[str] ): if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def A_ ( self : Any , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Dict ): if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token SCREAMING_SNAKE_CASE__ = model(**UpperCAmelCase_ , use_cache=UpperCAmelCase_ )[0] SCREAMING_SNAKE_CASE__ = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = model(**UpperCAmelCase_ , labels=UpperCAmelCase_ , use_cache=UpperCAmelCase_ )[:2] else: # compute label smoothed loss SCREAMING_SNAKE_CASE__ = model(**UpperCAmelCase_ , use_cache=UpperCAmelCase_ )[0] SCREAMING_SNAKE_CASE__ = torch.nn.functional.log_softmax(UpperCAmelCase_ , dim=-1 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.loss_fn(UpperCAmelCase_ , UpperCAmelCase_ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def A_ ( self : Union[str, Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : List[Any] ): SCREAMING_SNAKE_CASE__ = inputs.pop('labels' ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self._compute_loss(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) return loss def A_ ( self : List[str] , UpperCAmelCase_ : nn.Module , UpperCAmelCase_ : Dict[str, Union[torch.Tensor, Any]] , UpperCAmelCase_ : bool , UpperCAmelCase_ : Optional[List[str]] = None , ): SCREAMING_SNAKE_CASE__ = self._prepare_inputs(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: SCREAMING_SNAKE_CASE__ = self.model.generate( inputs['input_ids'] , attention_mask=inputs['attention_mask'] , **UpperCAmelCase_ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: SCREAMING_SNAKE_CASE__ = self._pad_tensors_to_max_len(UpperCAmelCase_ , gen_kwargs['max_length'] ) SCREAMING_SNAKE_CASE__ = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self._compute_loss(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) SCREAMING_SNAKE_CASE__ = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: SCREAMING_SNAKE_CASE__ = self._pad_tensors_to_max_len(UpperCAmelCase_ , gen_kwargs['max_length'] ) return (loss, logits, labels) def A_ ( self : Union[str, Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[str] ): # If PAD token is not defined at least EOS token has to be defined SCREAMING_SNAKE_CASE__ = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F' padded to `max_length`={max_length}' ) SCREAMING_SNAKE_CASE__ = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) SCREAMING_SNAKE_CASE__ = tensor return padded_tensor
176
0
'''simple docstring''' import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder __snake_case : Optional[int] = '__DUMMY_TRANSFORMERS_USER__' __snake_case : List[Any] = 'Dummy User' __snake_case : List[str] = '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' __snake_case : List[str] = 'https://hub-ci.huggingface.co' __snake_case : Optional[int] = CI_HUB_ENDPOINT + '/datasets/{repo_id}/resolve/{revision}/{path}' __snake_case : Any = CI_HUB_ENDPOINT + '/{repo_id}/resolve/{revision}/{filename}' __snake_case : str = Path('~/.huggingface/hub_ci_token').expanduser() @pytest.fixture def __lowerCamelCase ( __snake_case : Any ) -> Union[str, Any]: monkeypatch.setattr( """huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE""", __snake_case ) @pytest.fixture def __lowerCamelCase ( __snake_case : str ) -> str: monkeypatch.setattr("""datasets.config.HF_ENDPOINT""", __snake_case ) monkeypatch.setattr("""datasets.config.HUB_DATASETS_URL""", __snake_case ) @pytest.fixture def __lowerCamelCase ( __snake_case : Any ) -> Optional[int]: monkeypatch.setattr("""huggingface_hub.hf_api.HfFolder.path_token""", __snake_case ) @pytest.fixture def __lowerCamelCase ( __snake_case : Union[str, Any], __snake_case : Tuple ) -> Optional[Any]: HfFolder.save_token(__snake_case ) yield HfFolder.delete_token() @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( ) -> List[str]: return HfApi(endpoint=__snake_case ) @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( __snake_case : HfApi ) -> List[Any]: A__ : Optional[int] =HfFolder.get_token() HfFolder.save_token(__snake_case ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(__snake_case ) @pytest.fixture def __lowerCamelCase ( __snake_case : Any ) -> Optional[Any]: def _cleanup_repo(__snake_case : Optional[int] ): hf_api.delete_repo(__snake_case, token=__snake_case, repo_type="""dataset""" ) return _cleanup_repo @pytest.fixture def __lowerCamelCase ( __snake_case : List[str] ) -> str: @contextmanager def _temporary_repo(__snake_case : int ): try: yield repo_id finally: cleanup_repo(__snake_case ) return _temporary_repo @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( __snake_case : HfApi, __snake_case : int, __snake_case : Optional[int] ) -> Union[str, Any]: A__ : Optional[int] =f"repo_txt_data-{int(time.time() * 10E3 )}" A__ : Tuple =f"{CI_HUB_USER}/{repo_name}" hf_api.create_repo(__snake_case, token=__snake_case, repo_type="""dataset""", private=__snake_case ) hf_api.upload_file( token=__snake_case, path_or_fileobj=str(__snake_case ), path_in_repo="""data/text_data.txt""", repo_id=__snake_case, repo_type="""dataset""", ) yield repo_id try: hf_api.delete_repo(__snake_case, token=__snake_case, repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __lowerCamelCase ( __snake_case : Optional[Any], __snake_case : List[Any], __snake_case : Dict ) -> Optional[int]: return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( __snake_case : HfApi, __snake_case : Any, __snake_case : str ) -> Union[str, Any]: A__ : Union[str, Any] =f"repo_zipped_txt_data-{int(time.time() * 10E3 )}" A__ : Dict =f"{CI_HUB_USER}/{repo_name}" hf_api.create_repo(__snake_case, token=__snake_case, repo_type="""dataset""", private=__snake_case ) hf_api.upload_file( token=__snake_case, path_or_fileobj=str(__snake_case ), path_in_repo="""data.zip""", repo_id=__snake_case, repo_type="""dataset""", ) yield repo_id try: hf_api.delete_repo(__snake_case, token=__snake_case, repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __lowerCamelCase ( __snake_case : Optional[int], __snake_case : Dict, __snake_case : str ) -> List[Any]: return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="""session""" ) def __lowerCamelCase ( __snake_case : HfApi, __snake_case : str, __snake_case : Union[str, Any] ) -> str: A__ : Tuple =f"repo_zipped_img_data-{int(time.time() * 10E3 )}" A__ : Optional[Any] =f"{CI_HUB_USER}/{repo_name}" hf_api.create_repo(__snake_case, token=__snake_case, repo_type="""dataset""", private=__snake_case ) hf_api.upload_file( token=__snake_case, path_or_fileobj=str(__snake_case ), path_in_repo="""data.zip""", repo_id=__snake_case, repo_type="""dataset""", ) yield repo_id try: hf_api.delete_repo(__snake_case, token=__snake_case, repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __lowerCamelCase ( __snake_case : Optional[int], __snake_case : str, __snake_case : List[str] ) -> Union[str, Any]: return hf_private_dataset_repo_zipped_img_data_
366
'''simple docstring''' import torch from torch import nn class lowerCamelCase ( nn.Module ): '''simple docstring''' def __init__( self : Any , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[int]=1 , lowerCAmelCase_ : str=False ) -> List[str]: '''simple docstring''' super().__init__() A__ : Any =n_token A__ : int =d_embed A__ : Any =d_proj A__ : Tuple =cutoffs + [n_token] A__ : Optional[Any] =[0] + self.cutoffs A__ : Dict =div_val A__ : str =self.cutoffs[0] A__ : Optional[Any] =len(self.cutoffs ) - 1 A__ : List[Any] =self.shortlist_size + self.n_clusters if self.n_clusters > 0: A__ : Any =nn.Parameter(torch.zeros(self.n_clusters , self.d_embed ) ) A__ : str =nn.Parameter(torch.zeros(self.n_clusters ) ) A__ : Union[str, Any] =nn.ModuleList() A__ : Optional[int] =nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(lowerCAmelCase_ , lowerCAmelCase_ ) ) ) else: self.out_projs.append(lowerCAmelCase_ ) self.out_layers.append(nn.Linear(lowerCAmelCase_ , lowerCAmelCase_ ) ) else: for i in range(len(self.cutoffs ) ): A__ , A__ : Optional[int] =self.cutoff_ends[i], self.cutoff_ends[i + 1] A__ : Tuple =d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(lowerCAmelCase_ , lowerCAmelCase_ ) ) ) self.out_layers.append(nn.Linear(lowerCAmelCase_ , r_idx - l_idx ) ) A__ : Optional[int] =keep_order def lowercase__ ( self : Dict , lowerCAmelCase_ : int , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : Any ) -> Union[str, Any]: '''simple docstring''' if proj is None: A__ : Optional[int] =nn.functional.linear(lowerCAmelCase_ , lowerCAmelCase_ , bias=lowerCAmelCase_ ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: A__ : Optional[int] =nn.functional.linear(lowerCAmelCase_ , proj.t().contiguous() ) A__ : Union[str, Any] =nn.functional.linear(lowerCAmelCase_ , lowerCAmelCase_ , bias=lowerCAmelCase_ ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def lowercase__ ( self : Tuple , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[Any]=None , lowerCAmelCase_ : Dict=False ) -> Optional[int]: '''simple docstring''' if labels is not None: # Shift so that tokens < n predict n A__ : Optional[Any] =hidden[..., :-1, :].contiguous() A__ : List[Any] =labels[..., 1:].contiguous() A__ : Optional[int] =hidden.view(-1 , hidden.size(-1 ) ) A__ : str =labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError("""Input and labels should have the same size in the batch dimension.""" ) else: A__ : Optional[int] =hidden.view(-1 , hidden.size(-1 ) ) if self.n_clusters == 0: A__ : Optional[Any] =self._compute_logit(lowerCAmelCase_ , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) if labels is not None: A__ : Tuple =labels != -1_00 A__ : int =torch.zeros_like(lowerCAmelCase_ , dtype=hidden.dtype , device=hidden.device ) A__ : Union[str, Any] =( -nn.functional.log_softmax(lowerCAmelCase_ , dim=-1 )[mask].gather(1 , labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: A__ : List[Any] =nn.functional.log_softmax(lowerCAmelCase_ , dim=-1 ) else: # construct weights and biases A__ , A__ : Any =[], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: A__ , A__ : Optional[int] =self.cutoff_ends[i], self.cutoff_ends[i + 1] A__ : int =self.out_layers[0].weight[l_idx:r_idx] A__ : List[str] =self.out_layers[0].bias[l_idx:r_idx] else: A__ : List[str] =self.out_layers[i].weight A__ : Union[str, Any] =self.out_layers[i].bias if i == 0: A__ : Tuple =torch.cat([weight_i, self.cluster_weight] , dim=0 ) A__ : List[str] =torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(lowerCAmelCase_ ) biases.append(lowerCAmelCase_ ) A__ , A__ , A__ : Tuple =weights[0], biases[0], self.out_projs[0] A__ : List[Any] =self._compute_logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) A__ : int =nn.functional.log_softmax(lowerCAmelCase_ , dim=1 ) if labels is None: A__ : Union[str, Any] =hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: A__ : Union[str, Any] =torch.zeros_like(lowerCAmelCase_ , dtype=hidden.dtype , device=hidden.device ) A__ : Any =0 A__ : Tuple =[0] + self.cutoffs for i in range(len(lowerCAmelCase_ ) - 1 ): A__ , A__ : Tuple =cutoff_values[i], cutoff_values[i + 1] if labels is not None: A__ : Tuple =(labels >= l_idx) & (labels < r_idx) A__ : Any =mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue A__ : int =labels.index_select(0 , lowerCAmelCase_ ) - l_idx A__ : List[str] =head_logprob.index_select(0 , lowerCAmelCase_ ) A__ : str =hidden.index_select(0 , lowerCAmelCase_ ) else: A__ : Optional[Any] =hidden if i == 0: if labels is not None: A__ : Optional[Any] =head_logprob_i.gather(1 , target_i[:, None] ).squeeze(1 ) else: A__ : Union[str, Any] =head_logprob[:, : self.cutoffs[0]] else: A__ , A__ , A__ : Dict =weights[i], biases[i], self.out_projs[i] A__ : List[Any] =self._compute_logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) A__ : List[Any] =nn.functional.log_softmax(lowerCAmelCase_ , dim=1 ) A__ : Optional[Any] =self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: A__ : Union[str, Any] =head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None] ).squeeze(1 ) else: A__ : List[str] =head_logprob[:, cluster_prob_idx, None] + tail_logprob_i A__ : Tuple =logprob_i if labels is not None: if (hasattr(self , """keep_order""" ) and self.keep_order) or keep_order: out.index_copy_(0 , lowerCAmelCase_ , -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def lowercase__ ( self : List[str] , lowerCAmelCase_ : Optional[Any] ) -> Any: '''simple docstring''' if self.n_clusters == 0: A__ : List[str] =self._compute_logit(lowerCAmelCase_ , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) return nn.functional.log_softmax(lowerCAmelCase_ , dim=-1 ) else: # construct weights and biases A__ , A__ : List[str] =[], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: A__ , A__ : int =self.cutoff_ends[i], self.cutoff_ends[i + 1] A__ : List[str] =self.out_layers[0].weight[l_idx:r_idx] A__ : List[Any] =self.out_layers[0].bias[l_idx:r_idx] else: A__ : Dict =self.out_layers[i].weight A__ : Any =self.out_layers[i].bias if i == 0: A__ : List[str] =torch.cat([weight_i, self.cluster_weight] , dim=0 ) A__ : Tuple =torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(lowerCAmelCase_ ) biases.append(lowerCAmelCase_ ) A__ , A__ , A__ : Optional[int] =weights[0], biases[0], self.out_projs[0] A__ : Any =self._compute_logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) A__ : Dict =hidden.new_empty((head_logit.size(0 ), self.n_token) ) A__ : Dict =nn.functional.log_softmax(lowerCAmelCase_ , dim=1 ) A__ : Tuple =[0] + self.cutoffs for i in range(len(lowerCAmelCase_ ) - 1 ): A__ , A__ : List[Any] =cutoff_values[i], cutoff_values[i + 1] if i == 0: A__ : Tuple =head_logprob[:, : self.cutoffs[0]] else: A__ , A__ , A__ : Any =weights[i], biases[i], self.out_projs[i] A__ : Dict =self._compute_logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) A__ : str =nn.functional.log_softmax(lowerCAmelCase_ , dim=1 ) A__ : str =head_logprob[:, -i] + tail_logprob_i A__ : List[Any] =logprob_i return out
136
0
'''simple docstring''' def lowercase_ ( lowerCAmelCase__ : int ): """simple docstring""" __UpperCAmelCase : str = 0 while len(_lowerCAmelCase ) > 1: __UpperCAmelCase : Union[str, Any] = 0 # Consider two files with minimum cost to be merged for _ in range(2 ): __UpperCAmelCase : Any = files.index(min(_lowerCAmelCase ) ) temp += files[min_index] files.pop(_lowerCAmelCase ) files.append(_lowerCAmelCase ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
254
'''simple docstring''' # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class lowerCamelCase_ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" a_ =StableDiffusionControlNetImgaImgPipeline a_ =TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} a_ =TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a_ =IMAGE_TO_IMAGE_IMAGE_PARAMS.union({"""control_image"""} ) a_ =IMAGE_TO_IMAGE_IMAGE_PARAMS def _lowercase ( self : Union[str, Any] ) -> List[str]: torch.manual_seed(0 ) __lowerCamelCase : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) torch.manual_seed(0 ) __lowerCamelCase : Any = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) __lowerCamelCase : Any = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=_a , set_alpha_to_one=_a , ) torch.manual_seed(0 ) __lowerCamelCase : Dict = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCamelCase : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) __lowerCamelCase : Union[str, Any] = CLIPTextModel(_a ) __lowerCamelCase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCamelCase : int = { 'unet': unet, 'controlnet': controlnet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def _lowercase ( self : Union[str, Any] , _a : Any , _a : List[str]=0 ) -> List[str]: if str(_a ).startswith('mps' ): __lowerCamelCase : Any = torch.manual_seed(_a ) else: __lowerCamelCase : Any = torch.Generator(device=_a ).manual_seed(_a ) __lowerCamelCase : Union[str, Any] = 2 __lowerCamelCase : Any = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=_a , device=torch.device(_a ) , ) __lowerCamelCase : Optional[Any] = floats_tensor(control_image.shape , rng=random.Random(_a ) ).to(_a ) __lowerCamelCase : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCamelCase : int = Image.fromarray(np.uinta(_a ) ).convert('RGB' ).resize((64, 64) ) __lowerCamelCase : int = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', 'image': image, 'control_image': control_image, } return inputs def _lowercase ( self : List[str] ) -> Optional[Any]: return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3 ) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def _lowercase ( self : str ) -> List[str]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3 ) def _lowercase ( self : Any ) -> List[str]: self._test_inference_batch_single_identical(expected_max_diff=2e-3 ) class lowerCamelCase_ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" a_ =StableDiffusionControlNetImgaImgPipeline a_ =TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} a_ =TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a_ =frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def _lowercase ( self : List[Any] ) -> List[str]: torch.manual_seed(0 ) __lowerCamelCase : Any = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(_a : str ): if isinstance(_a , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) __lowerCamelCase : Any = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(_a ) torch.manual_seed(0 ) __lowerCamelCase : Any = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(_a ) torch.manual_seed(0 ) __lowerCamelCase : List[Any] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=_a , set_alpha_to_one=_a , ) torch.manual_seed(0 ) __lowerCamelCase : str = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCamelCase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) __lowerCamelCase : Dict = CLIPTextModel(_a ) __lowerCamelCase : Optional[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCamelCase : List[str] = MultiControlNetModel([controlneta, controlneta] ) __lowerCamelCase : str = { 'unet': unet, 'controlnet': controlnet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def _lowercase ( self : Dict , _a : int , _a : Union[str, Any]=0 ) -> Union[str, Any]: if str(_a ).startswith('mps' ): __lowerCamelCase : Optional[Any] = torch.manual_seed(_a ) else: __lowerCamelCase : Union[str, Any] = torch.Generator(device=_a ).manual_seed(_a ) __lowerCamelCase : Any = 2 __lowerCamelCase : Any = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=_a , device=torch.device(_a ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=_a , device=torch.device(_a ) , ), ] __lowerCamelCase : List[Any] = floats_tensor(control_image[0].shape , rng=random.Random(_a ) ).to(_a ) __lowerCamelCase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCamelCase : Dict = Image.fromarray(np.uinta(_a ) ).convert('RGB' ).resize((64, 64) ) __lowerCamelCase : Dict = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', 'image': image, 'control_image': control_image, } return inputs def _lowercase ( self : Tuple ) -> Union[str, Any]: __lowerCamelCase : Any = self.get_dummy_components() __lowerCamelCase : Any = self.pipeline_class(**_a ) pipe.to(_a ) __lowerCamelCase : Optional[int] = 10.0 __lowerCamelCase : Tuple = 4 __lowerCamelCase : str = self.get_dummy_inputs(_a ) __lowerCamelCase : int = steps __lowerCamelCase : List[Any] = scale __lowerCamelCase : Optional[Any] = pipe(**_a )[0] __lowerCamelCase : Dict = self.get_dummy_inputs(_a ) __lowerCamelCase : List[str] = steps __lowerCamelCase : List[Any] = scale __lowerCamelCase : Dict = pipe(**_a , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] __lowerCamelCase : Any = self.get_dummy_inputs(_a ) __lowerCamelCase : Union[str, Any] = steps __lowerCamelCase : Any = scale __lowerCamelCase : Optional[int] = pipe(**_a , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] __lowerCamelCase : Any = self.get_dummy_inputs(_a ) __lowerCamelCase : Tuple = steps __lowerCamelCase : List[Any] = scale __lowerCamelCase : List[str] = pipe(**_a , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 def _lowercase ( self : Optional[Any] ) -> List[str]: return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3 ) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def _lowercase ( self : Dict ) -> Tuple: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3 ) def _lowercase ( self : str ) -> int: self._test_inference_batch_single_identical(expected_max_diff=2e-3 ) def _lowercase ( self : List[Any] ) -> str: __lowerCamelCase : int = self.get_dummy_components() __lowerCamelCase : List[Any] = self.pipeline_class(**_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(_a ) except NotImplementedError: pass @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : List[str] ) -> List[str]: super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: __lowerCamelCase : Optional[Any] = ControlNetModel.from_pretrained('lllyasviel/sd-controlnet-canny' ) __lowerCamelCase : List[str] = StableDiffusionControlNetImgaImgPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , safety_checker=_a , controlnet=_a ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=_a ) __lowerCamelCase : Optional[int] = torch.Generator(device='cpu' ).manual_seed(0 ) __lowerCamelCase : List[Any] = 'evil space-punk bird' __lowerCamelCase : List[str] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ).resize((512, 512) ) __lowerCamelCase : Any = load_image( 'https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png' ).resize((512, 512) ) __lowerCamelCase : Dict = pipe( _a , _a , control_image=_a , generator=_a , output_type='np' , num_inference_steps=50 , strength=0.6 , ) __lowerCamelCase : Optional[Any] = output.images[0] assert image.shape == (512, 512, 3) __lowerCamelCase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy' ) assert np.abs(expected_image - image ).max() < 9e-2
208
0
"""simple docstring""" import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def __SCREAMING_SNAKE_CASE ( *__UpperCAmelCase ): if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): _lowercase : Tuple = list(__UpperCAmelCase ) for i in range(len(__UpperCAmelCase ) ): _lowercase : Tuple = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : List[Any] = [ """CUDA out of memory.""", # CUDA OOM """cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.""", # CUDNN SNAFU """DefaultCPUAllocator: can't allocate memory""", # CPU OOM ] if isinstance(__UpperCAmelCase , __UpperCAmelCase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase = None , __UpperCAmelCase = 128 ): if function is None: return functools.partial(__UpperCAmelCase , starting_batch_size=__UpperCAmelCase ) _lowercase : Any = starting_batch_size def decorator(*__UpperCAmelCase , **__UpperCAmelCase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() _lowercase : Tuple = list(inspect.signature(__UpperCAmelCase ).parameters.keys() ) # Guard against user error if len(__UpperCAmelCase ) < (len(__UpperCAmelCase ) + 1): _lowercase : Union[str, Any] = """, """.join([F"""{arg}={value}""" for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F"""Batch size was passed into `{function.__name__}` as the first argument when called.""" F"""Remove this as the decorator already does so: `{function.__name__}({arg_str})`""" ) while True: if batch_size == 0: raise RuntimeError("""No executable batch size found, reached zero.""" ) try: return function(__UpperCAmelCase , *__UpperCAmelCase , **__UpperCAmelCase ) except Exception as e: if should_reduce_batch_size(__UpperCAmelCase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
336
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase: Optional[int] = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class UpperCamelCase ( snake_case , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = XLNetTokenizer SCREAMING_SNAKE_CASE_ : Dict = XLNetTokenizerFast SCREAMING_SNAKE_CASE_ : int = True SCREAMING_SNAKE_CASE_ : Union[str, Any] = True def lowerCamelCase__ ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowercase : Dict = XLNetTokenizer(UpperCAmelCase_ ,keep_accents=UpperCAmelCase_ ) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase__ ( self ): _lowercase : Union[str, Any] = """<s>""" _lowercase : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase_ ) ,UpperCAmelCase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase_ ) ,UpperCAmelCase_ ) def lowerCamelCase__ ( self ): _lowercase : str = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""<unk>""" ) self.assertEqual(vocab_keys[1] ,"""<s>""" ) self.assertEqual(vocab_keys[-1] ,"""<eod>""" ) self.assertEqual(len(UpperCAmelCase_ ) ,10_06 ) def lowerCamelCase__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size ,10_00 ) def lowerCamelCase__ ( self ): _lowercase : int = XLNetTokenizer(UpperCAmelCase_ ,keep_accents=UpperCAmelCase_ ) _lowercase : Union[str, Any] = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(UpperCAmelCase_ ,["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) ,[2_85, 46, 10, 1_70, 3_82] ) _lowercase : Any = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCAmelCase_ ,[ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] ,) _lowercase : List[Any] = tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ ,[8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] ) _lowercase : Optional[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase_ ) self.assertListEqual( UpperCAmelCase_ ,[ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] ,) def lowerCamelCase__ ( self ): _lowercase : Dict = XLNetTokenizer(UpperCAmelCase_ ,do_lower_case=UpperCAmelCase_ ) _lowercase : List[str] = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCAmelCase_ ,[ SPIECE_UNDERLINE + """""", """i""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """se""", """.""", ] ,) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) ,["""▁he""", """ll""", """o"""] ) def lowerCamelCase__ ( self ): _lowercase : int = XLNetTokenizer(UpperCAmelCase_ ,do_lower_case=UpperCAmelCase_ ) _lowercase : List[str] = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCAmelCase_ ,[ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """se""", """.""", ] ,) @slow def lowerCamelCase__ ( self ): _lowercase : Dict = XLNetTokenizer.from_pretrained("""xlnet-base-cased""" ) _lowercase : int = tokenizer.encode("""sequence builders""" ,add_special_tokens=UpperCAmelCase_ ) _lowercase : List[str] = tokenizer.encode("""multi-sequence build""" ,add_special_tokens=UpperCAmelCase_ ) _lowercase : Optional[int] = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ) _lowercase : Dict = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase_ ,UpperCAmelCase_ ) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def lowerCamelCase__ ( self ): # fmt: off _lowercase : Union[str, Any] = {"""input_ids""": [[17, 2_14_42, 2_70, 17, 10, 1_46_45, 3_18, 34, 17, 45_46, 31_45, 7_87, 13, 77_52, 2_20_18, 23, 21, 17, 45_46, 31_45, 7_87, 13, 33_52, 1_44_31, 13, 55_00, 11, 11_76, 5_80, 13, 1_68_19, 47_97, 23, 17, 10, 1_71_35, 6_58, 19, 4_57, 79_32, 13, 1_84, 19, 31_54, 1_71_35, 64_68, 19, 14_04, 1_22_69, 19, 42_29, 53_56, 1_62_64, 46, 19, 17, 2_05_45, 1_03_95, 9, 9, 9, 11, 28, 64_21, 95_31, 2_07_29, 17, 10, 3_53, 1_70_22, 11, 21, 64_21, 95_31, 1_69_49, 17, 10, 1_15_09, 7_53, 11, 33, 95, 24_21, 73_85, 9_56, 1_44_31, 26_26, 25, 8_42, 73_85, 48_36, 21, 14_29, 22_72, 98_55, 31_20, 1_61, 2_47_38, 19, 1_32_03, 6_58, 2_18, 7_87, 21, 4_30, 1_84_82, 8_47, 26_37, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_22, 2_21_78, 27, 10_64, 22, 9_56, 13, 1_11_01, 14_29, 58_54, 2_43_13, 1_89_53, 40, 4_22, 2_43_66, 68, 17_58, 37, 1_04_83, 1_42_57, 31, 2_07, 2_63, 21, 2_03, 37_73, 25, 71, 97_35, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 32, 20_49, 34_42, 17, 1_38_94, 33_80, 23, 95, 18, 1_76_34, 22_88, 9, 4, 3]], """token_type_ids""": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase_ ,model_name="""xlnet-base-cased""" ,revision="""c841166438c31ec7ca9a106dee7bb312b73ae511""" ,)
336
1
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
182
'''simple docstring''' import logging import os from .state import PartialState class a__ ( logging.LoggerAdapter ): @staticmethod def SCREAMING_SNAKE_CASE__ ( a : Optional[Any] ): """simple docstring""" __lowerCamelCase = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def SCREAMING_SNAKE_CASE__ ( self : int , a : Optional[int] , a : str , *a : Optional[int] , **a : List[Any] ): """simple docstring""" if PartialState._shared_state == {}: raise RuntimeError( '''You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility.''' ) __lowerCamelCase = kwargs.pop('''main_process_only''' , a ) __lowerCamelCase = kwargs.pop('''in_order''' , a ) if self.isEnabledFor(a ): if self._should_log(a ): __lowerCamelCase , __lowerCamelCase = self.process(a , a ) self.logger.log(a , a , *a , **a ) elif in_order: __lowerCamelCase = PartialState() for i in range(state.num_processes ): if i == state.process_index: __lowerCamelCase , __lowerCamelCase = self.process(a , a ) self.logger.log(a , a , *a , **a ) state.wait_for_everyone() def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ = None ) -> Optional[int]: if log_level is None: __lowerCamelCase = os.environ.get('''ACCELERATE_LOG_LEVEL''' , UpperCamelCase__ ) __lowerCamelCase = logging.getLogger(UpperCamelCase__ ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(UpperCamelCase__ , {} )
67
0
import argparse import struct import unittest class __lowercase : """simple docstring""" def __init__( self , A ) -> str: snake_case : Dict = data # Initialize hash values snake_case : Dict = [ 0X6a09_e667, 0Xbb67_ae85, 0X3c6e_f372, 0Xa54f_f53a, 0X510e_527f, 0X9b05_688c, 0X1f83_d9ab, 0X5be0_cd19, ] # Initialize round constants snake_case : int = [ 0X428a_2f98, 0X7137_4491, 0Xb5c0_fbcf, 0Xe9b5_dba5, 0X3956_c25b, 0X59f1_11f1, 0X923f_82a4, 0Xab1c_5ed5, 0Xd807_aa98, 0X1283_5b01, 0X2431_85be, 0X550c_7dc3, 0X72be_5d74, 0X80de_b1fe, 0X9bdc_06a7, 0Xc19b_f174, 0Xe49b_69c1, 0Xefbe_4786, 0X0fc1_9dc6, 0X240c_a1cc, 0X2de9_2c6f, 0X4a74_84aa, 0X5cb0_a9dc, 0X76f9_88da, 0X983e_5152, 0Xa831_c66d, 0Xb003_27c8, 0Xbf59_7fc7, 0Xc6e0_0bf3, 0Xd5a7_9147, 0X06ca_6351, 0X1429_2967, 0X27b7_0a85, 0X2e1b_2138, 0X4d2c_6dfc, 0X5338_0d13, 0X650a_7354, 0X766a_0abb, 0X81c2_c92e, 0X9272_2c85, 0Xa2bf_e8a1, 0Xa81a_664b, 0Xc24b_8b70, 0Xc76c_51a3, 0Xd192_e819, 0Xd699_0624, 0Xf40e_3585, 0X106a_a070, 0X19a4_c116, 0X1e37_6c08, 0X2748_774c, 0X34b0_bcb5, 0X391c_0cb3, 0X4ed8_aa4a, 0X5b9c_ca4f, 0X682e_6ff3, 0X748f_82ee, 0X78a5_636f, 0X84c8_7814, 0X8cc7_0208, 0X90be_fffa, 0Xa450_6ceb, 0Xbef9_a3f7, 0Xc671_78f2, ] snake_case : Union[str, Any] = self.preprocessing(self.data ) self.final_hash() @staticmethod def UpperCAmelCase ( A ) -> Tuple: snake_case : Dict = B"""\x80""" + (B"""\x00""" * (6_3 - (len(__lowerCAmelCase ) + 8) % 6_4)) snake_case : Dict = struct.pack(""">Q""" , (len(__lowerCAmelCase ) * 8) ) return data + padding + big_endian_integer def UpperCAmelCase ( self ) -> Dict: snake_case : List[Any] = [ self.preprocessed_data[x : x + 6_4] for x in range(0 , len(self.preprocessed_data ) , 6_4 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers snake_case : Optional[Any] = list(struct.unpack(""">16L""" , __lowerCAmelCase ) ) # add 48 0-ed integers words += [0] * 4_8 snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case : Union[str, Any] = self.hashes for index in range(0 , 6_4 ): if index > 1_5: # modify the zero-ed indexes at the end of the array snake_case : List[str] = ( self.ror(words[index - 1_5] , 7 ) ^ self.ror(words[index - 1_5] , 1_8 ) ^ (words[index - 1_5] >> 3) ) snake_case : List[Any] = ( self.ror(words[index - 2] , 1_7 ) ^ self.ror(words[index - 2] , 1_9 ) ^ (words[index - 2] >> 1_0) ) snake_case : Dict = ( words[index - 1_6] + sa + words[index - 7] + sa ) % 0X1_0000_0000 # Compression snake_case : Optional[Any] = self.ror(__lowerCAmelCase , 6 ) ^ self.ror(__lowerCAmelCase , 1_1 ) ^ self.ror(__lowerCAmelCase , 2_5 ) snake_case : int = (e & f) ^ ((~e & 0Xffff_ffff) & g) snake_case : int = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X1_0000_0000 snake_case : Union[str, Any] = self.ror(__lowerCAmelCase , 2 ) ^ self.ror(__lowerCAmelCase , 1_3 ) ^ self.ror(__lowerCAmelCase , 2_2 ) snake_case : Dict = (a & b) ^ (a & c) ^ (b & c) snake_case : Optional[int] = (sa + maj) % 0X1_0000_0000 snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case : Optional[Any] = ( g, f, e, ((d + tempa) % 0X1_0000_0000), c, b, a, ((tempa + tempa) % 0X1_0000_0000), ) snake_case : Optional[int] = [a, b, c, d, e, f, g, h] # Modify final values snake_case : List[Any] = [ ((element + mutated_hash_values[index]) % 0X1_0000_0000) for index, element in enumerate(self.hashes ) ] snake_case : str = """""".join([hex(__lowerCAmelCase )[2:].zfill(8 ) for value in self.hashes] ) def UpperCAmelCase ( self , A , A ) -> int: return 0Xffff_ffff & (value << (3_2 - rotations)) | (value >> rotations) class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Tuple: import hashlib snake_case : str = bytes("""Test String""" , """utf-8""" ) self.assertEqual(SHAaaa(__lowerCAmelCase ).hash , hashlib.shaaaa(__lowerCAmelCase ).hexdigest() ) def SCREAMING_SNAKE_CASE__ ( ) -> None: import doctest doctest.testmod() snake_case : int = argparse.ArgumentParser() parser.add_argument( """-s""" ,"""--string""" ,dest="""input_string""" ,default="""Hello World!! Welcome to Cryptography""" ,help="""Hash the string""" ,) parser.add_argument( """-f""" ,"""--file""" ,dest="""input_file""" ,help="""Hash contents of a file""" ) snake_case : Union[str, Any] = parser.parse_args() snake_case : Optional[int] = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file ,"""rb""" ) as f: snake_case : Tuple = f.read() else: snake_case : Tuple = bytes(lowerCAmelCase__ ,"""utf-8""" ) print(SHAaaa(lowerCAmelCase__ ).hash ) if __name__ == "__main__": main()
360
from io import BytesIO from typing import List, Union import requests from ..utils import add_end_docstrings, is_decord_available, is_torch_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_decord_available(): import numpy as np from decord import VideoReader if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING lowerCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(UpperCamelCase__ ) class __lowercase (UpperCamelCase__ ): """simple docstring""" def __init__( self , *A , **A ) -> Optional[int]: super().__init__(*A , **A ) requires_backends(self , """decord""" ) self.check_model_type(A ) def UpperCAmelCase ( self , A=None , A=None , A=None ) -> int: snake_case : Any = {} if frame_sampling_rate is not None: snake_case : int = frame_sampling_rate if num_frames is not None: snake_case : Union[str, Any] = num_frames snake_case : str = {} if top_k is not None: snake_case : Optional[int] = top_k return preprocess_params, {}, postprocess_params def __call__( self , A , **A ) -> Dict: return super().__call__(A , **A ) def UpperCAmelCase ( self , A , A=None , A=1 ) -> Tuple: if num_frames is None: snake_case : Tuple = self.model.config.num_frames if video.startswith("""http://""" ) or video.startswith("""https://""" ): snake_case : Optional[int] = BytesIO(requests.get(A ).content ) snake_case : Optional[Any] = VideoReader(A ) videoreader.seek(0 ) snake_case : Optional[Any] = 0 snake_case : Optional[Any] = num_frames * frame_sampling_rate - 1 snake_case : Any = np.linspace(A , A , num=A , dtype=np.intaa ) snake_case : int = videoreader.get_batch(A ).asnumpy() snake_case : List[Any] = list(A ) snake_case : Optional[Any] = self.image_processor(A , return_tensors=self.framework ) return model_inputs def UpperCAmelCase ( self , A ) -> List[str]: snake_case : Dict = self.model(**A ) return model_outputs def UpperCAmelCase ( self , A , A=5 ) -> int: if top_k > self.model.config.num_labels: snake_case : str = self.model.config.num_labels if self.framework == "pt": snake_case : List[Any] = model_outputs.logits.softmax(-1 )[0] snake_case , snake_case : Tuple = probs.topk(A ) else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) snake_case : List[Any] = scores.tolist() snake_case : str = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(A , A )]
176
0
'''simple docstring''' import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.local_sgd import LocalSGD ######################################################################## # This is a fully working simple example to use Accelerate # with LocalSGD, which is a method to synchronize model # parameters every K batches. It is different, but complementary # to gradient accumulation. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __lowercase : Optional[int] = 16 __lowercase : List[str] = 32 def lowerCamelCase (_SCREAMING_SNAKE_CASE : Accelerator , _SCREAMING_SNAKE_CASE : int = 16 ): __a : Union[str, Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) __a : Optional[Any] = load_dataset('glue' , 'mrpc' ) def tokenize_function(_SCREAMING_SNAKE_CASE : Optional[int] ): # max_length=None => use the model max length (it's actually the default) __a : str = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): __a : List[Any] = datasets.map( _SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __a : List[str] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(_SCREAMING_SNAKE_CASE : Dict ): # On TPU it's best to pad everything to the same length or training will be very slow. __a : int = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": __a : List[Any] = 16 elif accelerator.mixed_precision != "no": __a : Any = 8 else: __a : Optional[int] = None return tokenizer.pad( _SCREAMING_SNAKE_CASE , padding='longest' , max_length=_SCREAMING_SNAKE_CASE , pad_to_multiple_of=_SCREAMING_SNAKE_CASE , return_tensors='pt' , ) # Instantiate dataloaders. __a : Union[str, Any] = DataLoader( tokenized_datasets['train'] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) __a : List[str] = DataLoader( tokenized_datasets['validation'] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders __lowercase : Tuple = mocked_dataloaders # noqa: F811 def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Dict ): # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS' , _SCREAMING_SNAKE_CASE ) == "1": __a : Optional[int] = 2 # New Code # __a : Union[str, Any] = int(args.gradient_accumulation_steps ) __a : Dict = int(args.local_sgd_steps ) # Initialize accelerator __a : Tuple = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=_SCREAMING_SNAKE_CASE ) if accelerator.distributed_type not in [DistributedType.NO, DistributedType.MULTI_CPU, DistributedType.MULTI_GPU]: raise NotImplementedError('LocalSGD is supported only for CPUs and GPUs (no DeepSpeed or MegatronLM)' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __a : Dict = config['lr'] __a : Any = int(config['num_epochs'] ) __a : int = int(config['seed'] ) __a : List[str] = int(config['batch_size'] ) __a : Any = evaluate.load('glue' , 'mrpc' ) set_seed(_SCREAMING_SNAKE_CASE ) __a , __a : int = get_dataloaders(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __a : List[str] = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=_SCREAMING_SNAKE_CASE ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). __a : Dict = model.to(accelerator.device ) # Instantiate optimizer __a : int = AdamW(params=model.parameters() , lr=_SCREAMING_SNAKE_CASE ) # Instantiate scheduler __a : Any = get_linear_schedule_with_warmup( optimizer=_SCREAMING_SNAKE_CASE , num_warmup_steps=100 , num_training_steps=(len(_SCREAMING_SNAKE_CASE ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __a , __a , __a , __a , __a : Any = accelerator.prepare( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Now we train the model for epoch in range(_SCREAMING_SNAKE_CASE ): model.train() with LocalSGD( accelerator=_SCREAMING_SNAKE_CASE , model=_SCREAMING_SNAKE_CASE , local_sgd_steps=_SCREAMING_SNAKE_CASE , enabled=local_sgd_steps is not None ) as local_sgd: for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(_SCREAMING_SNAKE_CASE ): __a : List[Any] = model(**_SCREAMING_SNAKE_CASE ) __a : int = output.loss accelerator.backward(_SCREAMING_SNAKE_CASE ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() # LocalSGD-specific line local_sgd.step() model.eval() for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __a : Dict = model(**_SCREAMING_SNAKE_CASE ) __a : Tuple = outputs.logits.argmax(dim=-1 ) __a , __a : Union[str, Any] = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=_SCREAMING_SNAKE_CASE , references=_SCREAMING_SNAKE_CASE , ) __a : int = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , _SCREAMING_SNAKE_CASE ) def lowerCamelCase (): __a : Optional[Any] = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) # New Code # parser.add_argument( '--gradient_accumulation_steps' , type=_SCREAMING_SNAKE_CASE , default=1 , help='The number of minibatches to be ran before gradients are accumulated.' , ) parser.add_argument( '--local_sgd_steps' , type=_SCREAMING_SNAKE_CASE , default=8 , help='Number of local SGD steps or None to disable local SGD' ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) __a : Optional[Any] = parser.parse_args() __a : str = {'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
27
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _a = logging.get_logger(__name__) class __lowerCamelCase ( snake_case__): """simple docstring""" UpperCamelCase__ = "AutoTokenizer" UpperCamelCase__ = ["tokenizer"] UpperCamelCase__ = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self , UpperCAmelCase , UpperCAmelCase=None ): """simple docstring""" super().__init__(UpperCAmelCase ) _UpperCAmelCase = speaker_embeddings @classmethod def UpperCamelCase ( cls , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , **UpperCAmelCase ): """simple docstring""" if speaker_embeddings_dict_path is not None: _UpperCAmelCase = get_file_from_repo( UpperCAmelCase , UpperCAmelCase , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(UpperCAmelCase , UpperCAmelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) _UpperCAmelCase = None else: with open(UpperCAmelCase ) as speaker_embeddings_json: _UpperCAmelCase = json.load(UpperCAmelCase ) else: _UpperCAmelCase = None _UpperCAmelCase = AutoTokenizer.from_pretrained(UpperCAmelCase , **UpperCAmelCase ) return cls(tokenizer=UpperCAmelCase , speaker_embeddings=UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase="speaker_embeddings_path.json" , UpperCAmelCase="speaker_embeddings" , UpperCAmelCase = False , **UpperCAmelCase , ): """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(UpperCAmelCase , UpperCAmelCase , 'v2' ) , exist_ok=UpperCAmelCase ) _UpperCAmelCase = {} _UpperCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) _UpperCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] , UpperCAmelCase , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=UpperCAmelCase , ) _UpperCAmelCase = os.path.join(UpperCAmelCase , F"""{prompt_key}_{key}.npy""" ) _UpperCAmelCase = tmp_dict with open(os.path.join(UpperCAmelCase , UpperCAmelCase ) , 'w' ) as fp: json.dump(UpperCAmelCase , UpperCAmelCase ) super().save_pretrained(UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase = None , **UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = self.speaker_embeddings[voice_preset] _UpperCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) _UpperCAmelCase = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] , subfolder=kwargs.pop('subfolder' , UpperCAmelCase ) , cache_dir=kwargs.pop('cache_dir' , UpperCAmelCase ) , force_download=kwargs.pop('force_download' , UpperCAmelCase ) , proxies=kwargs.pop('proxies' , UpperCAmelCase ) , resume_download=kwargs.pop('resume_download' , UpperCAmelCase ) , local_files_only=kwargs.pop('local_files_only' , UpperCAmelCase ) , use_auth_token=kwargs.pop('use_auth_token' , UpperCAmelCase ) , revision=kwargs.pop('revision' , UpperCAmelCase ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) _UpperCAmelCase = np.load(UpperCAmelCase ) return voice_preset_dict def UpperCamelCase ( self , UpperCAmelCase = None ): """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase="pt" , UpperCAmelCase=256 , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=False , **UpperCAmelCase , ): """simple docstring""" if voice_preset is not None and not isinstance(UpperCAmelCase , UpperCAmelCase ): if ( isinstance(UpperCAmelCase , UpperCAmelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _UpperCAmelCase = self._load_voice_preset(UpperCAmelCase ) else: if isinstance(UpperCAmelCase , UpperCAmelCase ) and not voice_preset.endswith('.npz' ): _UpperCAmelCase = voice_preset + '.npz' _UpperCAmelCase = np.load(UpperCAmelCase ) if voice_preset is not None: self._validate_voice_preset_dict(UpperCAmelCase , **UpperCAmelCase ) _UpperCAmelCase = BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase ) _UpperCAmelCase = self.tokenizer( UpperCAmelCase , return_tensors=UpperCAmelCase , padding='max_length' , max_length=UpperCAmelCase , return_attention_mask=UpperCAmelCase , return_token_type_ids=UpperCAmelCase , add_special_tokens=UpperCAmelCase , **UpperCAmelCase , ) if voice_preset is not None: _UpperCAmelCase = voice_preset return encoded_text
39
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __UpperCamelCase : int = logging.get_logger(__name__) __UpperCamelCase : Optional[int] = '▁' __UpperCamelCase : int = {'vocab_file': 'sentencepiece.bpe.model', 'monolingual_vocab_file': 'dict.txt'} __UpperCamelCase : List[Any] = { 'vocab_file': { 'vinai/bartpho-syllable': 'https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model', }, 'monolingual_vocab_file': { 'vinai/bartpho-syllable': 'https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt', }, } __UpperCamelCase : Optional[Any] = {'vinai/bartpho-syllable': 1024} class lowercase__ ( UpperCamelCase_): UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = ["""input_ids""", """attention_mask"""] def __init__( self : Union[str, Any] , UpperCamelCase__ : str , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str="<s>" , UpperCamelCase__ : Union[str, Any]="</s>" , UpperCamelCase__ : Optional[Any]="</s>" , UpperCamelCase__ : str="<s>" , UpperCamelCase__ : Any="<unk>" , UpperCamelCase__ : Optional[Any]="<pad>" , UpperCamelCase__ : List[str]="<mask>" , UpperCamelCase__ : Optional[Dict[str, Any]] = None , **UpperCamelCase__ : Union[str, Any] , ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token SCREAMING_SNAKE_CASE : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Optional[Any] = vocab_file SCREAMING_SNAKE_CASE : Tuple = monolingual_vocab_file SCREAMING_SNAKE_CASE : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(UpperCamelCase__ ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility SCREAMING_SNAKE_CASE : int = {} SCREAMING_SNAKE_CASE : Optional[int] = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(UpperCamelCase__ ) not in self.fairseq_tokens_to_ids: SCREAMING_SNAKE_CASE : Optional[int] = cnt cnt += 1 with open(UpperCamelCase__ , '''r''' , encoding='''utf-8''' ) as f: for line in f.readlines(): SCREAMING_SNAKE_CASE : Dict = line.strip().split()[0] SCREAMING_SNAKE_CASE : Union[str, Any] = len(self.fairseq_tokens_to_ids ) if str(UpperCamelCase__ ) not in self.fairseq_tokens_to_ids: SCREAMING_SNAKE_CASE : Dict = len(self.fairseq_tokens_to_ids ) SCREAMING_SNAKE_CASE : List[Any] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : str ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = self.__dict__.copy() SCREAMING_SNAKE_CASE : str = None SCREAMING_SNAKE_CASE : Any = self.sp_model.serialized_model_proto() return state def __setstate__( self : Optional[int] , UpperCamelCase__ : Any ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): SCREAMING_SNAKE_CASE : Optional[int] = {} SCREAMING_SNAKE_CASE : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __A ( self : Optional[Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] SCREAMING_SNAKE_CASE : Any = [self.cls_token_id] SCREAMING_SNAKE_CASE : Any = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __A ( self : Union[str, Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(UpperCamelCase__ )) + [1] return [1] + ([0] * len(UpperCamelCase__ )) + [1, 1] + ([0] * len(UpperCamelCase__ )) + [1] def __A ( self : str , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = [self.sep_token_id] SCREAMING_SNAKE_CASE : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __A ( self : Optional[int] ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def __A ( self : Optional[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = {self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __A ( self : List[Any] , UpperCamelCase__ : str ): '''simple docstring''' return self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) def __A ( self : Tuple , UpperCamelCase__ : Dict ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __A ( self : int , UpperCamelCase__ : int ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def __A ( self : List[Any] , UpperCamelCase__ : Union[str, Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = ''''''.join(UpperCamelCase__ ).replace(UpperCamelCase__ , ''' ''' ).strip() return out_string def __A ( self : Optional[int] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(UpperCamelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return SCREAMING_SNAKE_CASE : Optional[Any] = os.path.join( UpperCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) SCREAMING_SNAKE_CASE : Optional[Any] = os.path.join( UpperCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''monolingual_vocab_file'''] , ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , '''wb''' ) as fi: SCREAMING_SNAKE_CASE : List[Any] = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( UpperCamelCase__ ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(UpperCamelCase__ , '''w''' , encoding='''utf-8''' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(f"""{str(UpperCamelCase__ )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
258
import os import posixpath import uuid from dataclasses import dataclass from typing import TYPE_CHECKING, Iterable, List, Optional, Tuple, Union import numpy as np import pyarrow as pa import datasets from datasets.arrow_writer import ArrowWriter, ParquetWriter from datasets.config import MAX_SHARD_SIZE from datasets.filesystems import ( is_remote_filesystem, rename, ) from datasets.iterable_dataset import _BaseExamplesIterable from datasets.utils.py_utils import convert_file_size_to_int __UpperCamelCase : Union[str, Any] = datasets.utils.logging.get_logger(__name__) if TYPE_CHECKING: import pyspark @dataclass class lowercase__ ( datasets.BuilderConfig): UpperCamelCase_ = None def A ( _lowercase , _lowercase , ): import pyspark def generate_fn(): SCREAMING_SNAKE_CASE : str = df.select('''*''' , pyspark.sql.functions.spark_partition_id().alias('''part_id''' ) ) for partition_id in partition_order: SCREAMING_SNAKE_CASE : Any = df_with_partition_id.select('''*''' ).where(f"""part_id = {partition_id}""" ).drop('''part_id''' ) SCREAMING_SNAKE_CASE : Tuple = partition_df.collect() SCREAMING_SNAKE_CASE : str = 0 for row in rows: yield f"""{partition_id}_{row_id}""", row.asDict() row_id += 1 return generate_fn class lowercase__ ( _BaseExamplesIterable): def __init__( self : Optional[Any] , UpperCamelCase__ : "pyspark.sql.DataFrame" , UpperCamelCase__ : Union[str, Any]=None , ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = df SCREAMING_SNAKE_CASE : List[str] = partition_order or range(self.df.rdd.getNumPartitions() ) SCREAMING_SNAKE_CASE : Dict = _generate_iterable_examples(self.df , self.partition_order ) def __iter__( self : List[Any] ): '''simple docstring''' yield from self.generate_examples_fn() def __A ( self : Tuple , UpperCamelCase__ : np.random.Generator ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = list(range(self.df.rdd.getNumPartitions() ) ) generator.shuffle(UpperCamelCase__ ) return SparkExamplesIterable(self.df , partition_order=UpperCamelCase__ ) def __A ( self : List[str] , UpperCamelCase__ : int , UpperCamelCase__ : int ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = self.split_shard_indices_by_worker(UpperCamelCase__ , UpperCamelCase__ ) return SparkExamplesIterable(self.df , partition_order=UpperCamelCase__ ) @property def __A ( self : Tuple ): '''simple docstring''' return len(self.partition_order ) class lowercase__ ( datasets.DatasetBuilder): UpperCamelCase_ = SparkConfig def __init__( self : Union[str, Any] , UpperCamelCase__ : "pyspark.sql.DataFrame" , UpperCamelCase__ : str = None , UpperCamelCase__ : str = None , **UpperCamelCase__ : Union[str, Any] , ): '''simple docstring''' import pyspark SCREAMING_SNAKE_CASE : str = pyspark.sql.SparkSession.builder.getOrCreate() SCREAMING_SNAKE_CASE : List[Any] = df SCREAMING_SNAKE_CASE : Tuple = working_dir super().__init__( cache_dir=UpperCamelCase__ , config_name=str(self.df.semanticHash() ) , **UpperCamelCase__ , ) def __A ( self : Tuple ): '''simple docstring''' def create_cache_and_write_probe(UpperCamelCase__ : str ): # makedirs with exist_ok will recursively create the directory. It will not throw an error if directories # already exist. os.makedirs(self._cache_dir , exist_ok=UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Any = os.path.join(self._cache_dir , '''fs_test''' + uuid.uuida().hex ) # Opening the file in append mode will create a new file unless it already exists, in which case it will not # change the file contents. open(UpperCamelCase__ , '''a''' ) return [probe_file] if self._spark.conf.get('''spark.master''' , '''''' ).startswith('''local''' ): return # If the cluster is multi-node, make sure that the user provided a cache_dir and that it is on an NFS # accessible to the driver. # TODO: Stream batches to the driver using ArrowCollectSerializer instead of throwing an error. if self._cache_dir: SCREAMING_SNAKE_CASE : Dict = ( self._spark.sparkContext.parallelize(range(1 ) , 1 ).mapPartitions(UpperCamelCase__ ).collect() ) if os.path.isfile(probe[0] ): return raise ValueError( '''When using Dataset.from_spark on a multi-node cluster, the driver and all workers should be able to access cache_dir''' ) def __A ( self : Any ): '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def __A ( self : str , UpperCamelCase__ : datasets.download.download_manager.DownloadManager ): '''simple docstring''' return [datasets.SplitGenerator(name=datasets.Split.TRAIN )] def __A ( self : int , UpperCamelCase__ : List[Any] ): '''simple docstring''' import pyspark def get_arrow_batch_size(UpperCamelCase__ : Tuple ): for batch in it: yield pa.RecordBatch.from_pydict({'''batch_bytes''': [batch.nbytes]} ) SCREAMING_SNAKE_CASE : int = self.df.count() SCREAMING_SNAKE_CASE : Tuple = df_num_rows if df_num_rows <= 100 else 100 # Approximate the size of each row (in Arrow format) by averaging over a max-100-row sample. SCREAMING_SNAKE_CASE : Optional[Any] = ( self.df.limit(UpperCamelCase__ ) .repartition(1 ) .mapInArrow(UpperCamelCase__ , '''batch_bytes: long''' ) .agg(pyspark.sql.functions.sum('''batch_bytes''' ).alias('''sample_bytes''' ) ) .collect()[0] .sample_bytes / sample_num_rows ) SCREAMING_SNAKE_CASE : Optional[Any] = approx_bytes_per_row * df_num_rows if approx_total_size > max_shard_size: # Make sure there is at least one row per partition. SCREAMING_SNAKE_CASE : List[str] = min(UpperCamelCase__ , int(approx_total_size / max_shard_size ) ) SCREAMING_SNAKE_CASE : Optional[int] = self.df.repartition(UpperCamelCase__ ) def __A ( self : str , UpperCamelCase__ : str , UpperCamelCase__ : str , UpperCamelCase__ : int , ): '''simple docstring''' import pyspark SCREAMING_SNAKE_CASE : int = ParquetWriter if file_format == '''parquet''' else ArrowWriter SCREAMING_SNAKE_CASE : Optional[Any] = os.path.join(self._working_dir , os.path.basename(UpperCamelCase__ ) ) if self._working_dir else fpath SCREAMING_SNAKE_CASE : Optional[int] = file_format == '''parquet''' # Define these so that we don't reference self in write_arrow, which will result in a pickling error due to # pickling the SparkContext. SCREAMING_SNAKE_CASE : str = self.config.features SCREAMING_SNAKE_CASE : Optional[int] = self._writer_batch_size SCREAMING_SNAKE_CASE : Optional[int] = self._fs.storage_options def write_arrow(UpperCamelCase__ : Optional[Any] ): # Within the same SparkContext, no two task attempts will share the same attempt ID. SCREAMING_SNAKE_CASE : int = pyspark.TaskContext().taskAttemptId() SCREAMING_SNAKE_CASE : str = next(UpperCamelCase__ , UpperCamelCase__ ) if first_batch is None: # Some partitions might not receive any data. return pa.RecordBatch.from_arrays( [[task_id], [0], [0]] , names=['''task_id''', '''num_examples''', '''num_bytes'''] , ) SCREAMING_SNAKE_CASE : List[Any] = 0 SCREAMING_SNAKE_CASE : Optional[int] = writer_class( features=UpperCamelCase__ , path=working_fpath.replace('''SSSSS''' , f"""{shard_id:05d}""" ).replace('''TTTTT''' , f"""{task_id:05d}""" ) , writer_batch_size=UpperCamelCase__ , storage_options=UpperCamelCase__ , embed_local_files=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Tuple = pa.Table.from_batches([first_batch] ) writer.write_table(UpperCamelCase__ ) for batch in it: if max_shard_size is not None and writer._num_bytes >= max_shard_size: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[Any] = writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]] , names=['''task_id''', '''num_examples''', '''num_bytes'''] , ) shard_id += 1 SCREAMING_SNAKE_CASE : Optional[int] = writer_class( features=writer._features , path=working_fpath.replace('''SSSSS''' , f"""{shard_id:05d}""" ).replace('''TTTTT''' , f"""{task_id:05d}""" ) , writer_batch_size=UpperCamelCase__ , storage_options=UpperCamelCase__ , embed_local_files=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : List[str] = pa.Table.from_batches([batch] ) writer.write_table(UpperCamelCase__ ) if writer._num_bytes > 0: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[Any] = writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]] , names=['''task_id''', '''num_examples''', '''num_bytes'''] , ) if working_fpath != fpath: for file in os.listdir(os.path.dirname(UpperCamelCase__ ) ): SCREAMING_SNAKE_CASE : int = os.path.join(os.path.dirname(UpperCamelCase__ ) , os.path.basename(UpperCamelCase__ ) ) shutil.move(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = ( self.df.mapInArrow(UpperCamelCase__ , '''task_id: long, num_examples: long, num_bytes: long''' ) .groupBy('''task_id''' ) .agg( pyspark.sql.functions.sum('''num_examples''' ).alias('''total_num_examples''' ) , pyspark.sql.functions.sum('''num_bytes''' ).alias('''total_num_bytes''' ) , pyspark.sql.functions.count('''num_bytes''' ).alias('''num_shards''' ) , pyspark.sql.functions.collect_list('''num_examples''' ).alias('''shard_lengths''' ) , ) .collect() ) for row in stats: yield row.task_id, (row.total_num_examples, row.total_num_bytes, row.num_shards, row.shard_lengths) def __A ( self : Dict , UpperCamelCase__ : "datasets.SplitGenerator" , UpperCamelCase__ : str = "arrow" , UpperCamelCase__ : Optional[Union[str, int]] = None , UpperCamelCase__ : Optional[int] = None , **UpperCamelCase__ : str , ): '''simple docstring''' self._validate_cache_dir() SCREAMING_SNAKE_CASE : List[str] = convert_file_size_to_int(max_shard_size or MAX_SHARD_SIZE ) self._repartition_df_if_needed(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : List[str] = not is_remote_filesystem(self._fs ) SCREAMING_SNAKE_CASE : Optional[Any] = os.path.join if is_local else posixpath.join SCREAMING_SNAKE_CASE : List[Any] = '''-TTTTT-SSSSS-of-NNNNN''' SCREAMING_SNAKE_CASE : List[str] = f"""{self.name}-{split_generator.name}{SUFFIX}.{file_format}""" SCREAMING_SNAKE_CASE : Dict = path_join(self._output_dir , UpperCamelCase__ ) SCREAMING_SNAKE_CASE : List[str] = 0 SCREAMING_SNAKE_CASE : Dict = 0 SCREAMING_SNAKE_CASE : Tuple = 0 SCREAMING_SNAKE_CASE : Dict = [] SCREAMING_SNAKE_CASE : Dict = [] for task_id, content in self._prepare_split_single(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): ( ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ( SCREAMING_SNAKE_CASE ) , ) : int = content if num_bytes > 0: total_num_examples += num_examples total_num_bytes += num_bytes total_shards += num_shards task_id_and_num_shards.append((task_id, num_shards) ) all_shard_lengths.extend(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = total_num_examples SCREAMING_SNAKE_CASE : Dict = total_num_bytes # should rename everything at the end logger.debug(f"""Renaming {total_shards} shards.""" ) if total_shards > 1: SCREAMING_SNAKE_CASE : Tuple = all_shard_lengths # Define fs outside of _rename_shard so that we don't reference self in the function, which will result in a # pickling error due to pickling the SparkContext. SCREAMING_SNAKE_CASE : str = self._fs # use the -SSSSS-of-NNNNN pattern def _rename_shard( UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int , ): rename( UpperCamelCase__ , fpath.replace('''SSSSS''' , f"""{shard_id:05d}""" ).replace('''TTTTT''' , f"""{task_id:05d}""" ) , fpath.replace('''TTTTT-SSSSS''' , f"""{global_shard_id:05d}""" ).replace('''NNNNN''' , f"""{total_shards:05d}""" ) , ) SCREAMING_SNAKE_CASE : Optional[Any] = [] SCREAMING_SNAKE_CASE : Optional[Any] = 0 for i in range(len(UpperCamelCase__ ) ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Any = task_id_and_num_shards[i] for shard_id in range(UpperCamelCase__ ): args.append([task_id, shard_id, global_shard_id] ) global_shard_id += 1 self._spark.sparkContext.parallelize(UpperCamelCase__ , len(UpperCamelCase__ ) ).map(lambda UpperCamelCase__ : _rename_shard(*UpperCamelCase__ ) ).collect() else: # don't use any pattern SCREAMING_SNAKE_CASE : Optional[int] = 0 SCREAMING_SNAKE_CASE : Union[str, Any] = task_id_and_num_shards[0][0] self._rename( fpath.replace('''SSSSS''' , f"""{shard_id:05d}""" ).replace('''TTTTT''' , f"""{task_id:05d}""" ) , fpath.replace(UpperCamelCase__ , '''''' ) , ) def __A ( self : int , UpperCamelCase__ : "datasets.SplitGenerator" , ): '''simple docstring''' return SparkExamplesIterable(self.df )
258
1