code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
from math import factorial def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if successes > trials: raise ValueError("""successes must be lower or equal to trials""" ) if trials < 0 or successes < 0: raise ValueError("""the function is defined for non-negative integers""" ) if not isinstance(lowerCamelCase, lowerCamelCase ) or not isinstance(lowerCamelCase, lowerCamelCase ): raise ValueError("""the function is defined for non-negative integers""" ) if not 0 < prob < 1: raise ValueError("""prob has to be in range of 1 - 0""" ) lowerCamelCase : List[str] = (prob**successes) * ((1 - prob) ** (trials - successes)) # Calculate the binomial coefficient: n! / k!(n-k)! lowerCamelCase : Optional[Any] = float(factorial(lowerCamelCase ) ) coefficient /= factorial(lowerCamelCase ) * factorial(trials - successes ) return probability * coefficient if __name__ == "__main__": from doctest import testmod testmod() print("""Probability of 2 successes out of 4 trails""") print("""with probability of 0.75 is:""", end=""" """) print(binomial_distribution(2, 4, 0.75))
287
# Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from packaging import version from .. import __version__ from .constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD from .doc import ( add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, copy_func, replace_return_docstrings, ) from .generic import ( ContextManagers, ExplicitEnum, ModelOutput, PaddingStrategy, TensorType, add_model_info_to_auto_map, cached_property, can_return_loss, expand_dims, find_labels, flatten_dict, infer_framework, is_jax_tensor, is_numpy_array, is_tensor, is_tf_symbolic_tensor, is_tf_tensor, is_torch_device, is_torch_dtype, is_torch_tensor, reshape, squeeze, strtobool, tensor_size, to_numpy, to_py_obj, transpose, working_or_temp_dir, ) from .hub import ( CLOUDFRONT_DISTRIB_PREFIX, DISABLE_TELEMETRY, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, EntryNotFoundError, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, cached_file, default_cache_path, define_sagemaker_information, download_url, extract_commit_hash, get_cached_models, get_file_from_repo, get_full_repo_name, has_file, http_user_agent, is_offline_mode, is_remote_url, move_cache, send_example_telemetry, try_to_load_from_cache, ) from .import_utils import ( ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, TORCH_FX_REQUIRED_VERSION, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, ccl_version, direct_transformers_import, get_torch_version, is_accelerate_available, is_apex_available, is_bitsandbytes_available, is_bsa_available, is_coloredlogs_available, is_cython_available, is_datasets_available, is_decord_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_jieba_available, is_jumanpp_available, is_kenlm_available, is_keras_nlp_available, is_librosa_available, is_natten_available, is_ninja_available, is_onnx_available, is_openai_available, is_optimum_available, is_pandas_available, is_peft_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytest_available, is_pytorch_quantization_available, is_rjieba_available, is_sacremoses_available, is_safetensors_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_sudachi_available, is_tensorflow_probability_available, is_tensorflow_text_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_bfaa_cpu_available, is_torch_bfaa_gpu_available, is_torch_compile_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_neuroncore_available, is_torch_tensorrt_fx_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_torchdistx_available, is_torchdynamo_available, is_torchvision_available, is_training_run_on_sagemaker, is_vision_available, requires_backends, torch_only_method, ) _lowerCamelCase ="""pytorch_model.bin""" _lowerCamelCase ="""pytorch_model.bin.index.json""" _lowerCamelCase ="""adapter_config.json""" _lowerCamelCase ="""adapter_model.bin""" _lowerCamelCase ="""adapter_model.safetensors""" _lowerCamelCase ="""tf_model.h5""" _lowerCamelCase ="""tf_model.h5.index.json""" _lowerCamelCase ="""model.ckpt""" _lowerCamelCase ="""flax_model.msgpack""" _lowerCamelCase ="""flax_model.msgpack.index.json""" _lowerCamelCase ="""model.safetensors""" _lowerCamelCase ="""model.safetensors.index.json""" _lowerCamelCase ="""config.json""" _lowerCamelCase ="""preprocessor_config.json""" _lowerCamelCase =FEATURE_EXTRACTOR_NAME _lowerCamelCase ="""generation_config.json""" _lowerCamelCase ="""modelcard.json""" _lowerCamelCase ="""▁""" _lowerCamelCase =SENTENCEPIECE_UNDERLINE # Kept for backward compatibility _lowerCamelCase =[ [[0, 1, 0, 1], [1, 0, 0, 1]] ] * 2 # Needs to have 0s and 1s only since XLM uses it for langs too. _lowerCamelCase =[[7, 6, 0, 0, 1], [1, 2, 3, 0, 0], [0, 0, 0, 4, 5]] _lowerCamelCase =[[1, 1, 1, 1, 1], [1, 1, 1, 0, 0], [0, 0, 0, 1, 1]] def _a ( lowerCamelCase ): if version.parse(lowerCamelCase ) < version.parse(lowerCamelCase ): if "dev" in min_version: lowerCamelCase : Optional[int] = ( """This example requires a source install from HuggingFace Transformers (see """ """`https://huggingface.co/docs/transformers/installation#install-from-source`),""" ) else: lowerCamelCase : int = F'''This example requires a minimum version of {min_version},''' error_message += F''' but the version found is {__version__}.\n''' raise ImportError( error_message + """Check out https://github.com/huggingface/transformers/tree/main/examples#important-note for the examples corresponding to other """ """versions of HuggingFace Transformers.""" )
287
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """hustvl/yolos-small""": """https://huggingface.co/hustvl/yolos-small/resolve/main/config.json""", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = """yolos""" def __init__( self , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=[5_1_2, 8_6_4] , __magic_name__=1_6 , __magic_name__=3 , __magic_name__=True , __magic_name__=1_0_0 , __magic_name__=True , __magic_name__=False , __magic_name__=1 , __magic_name__=5 , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=0.1 , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Optional[Any] = hidden_size lowerCamelCase : str = num_hidden_layers lowerCamelCase : str = num_attention_heads lowerCamelCase : str = intermediate_size lowerCamelCase : List[Any] = hidden_act lowerCamelCase : Union[str, Any] = hidden_dropout_prob lowerCamelCase : List[Any] = attention_probs_dropout_prob lowerCamelCase : List[Any] = initializer_range lowerCamelCase : Any = layer_norm_eps lowerCamelCase : Tuple = image_size lowerCamelCase : Dict = patch_size lowerCamelCase : List[str] = num_channels lowerCamelCase : Union[str, Any] = qkv_bias lowerCamelCase : str = num_detection_tokens lowerCamelCase : List[str] = use_mid_position_embeddings lowerCamelCase : Union[str, Any] = auxiliary_loss # Hungarian matcher lowerCamelCase : Dict = class_cost lowerCamelCase : List[str] = bbox_cost lowerCamelCase : List[Any] = giou_cost # Loss coefficients lowerCamelCase : List[Any] = bbox_loss_coefficient lowerCamelCase : Optional[int] = giou_loss_coefficient lowerCamelCase : int = eos_coefficient class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-4 @property def UpperCamelCase__ ( self ): return 1_2
287
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = """camembert""" def __init__( self , __magic_name__=3_0_5_2_2 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , __magic_name__=None , **__magic_name__ , ): super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : int = num_hidden_layers lowerCamelCase : int = num_attention_heads lowerCamelCase : Optional[int] = hidden_act lowerCamelCase : List[Any] = intermediate_size lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Optional[int] = attention_probs_dropout_prob lowerCamelCase : Optional[int] = max_position_embeddings lowerCamelCase : str = type_vocab_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : int = layer_norm_eps lowerCamelCase : Any = position_embedding_type lowerCamelCase : Optional[int] = use_cache lowerCamelCase : Union[str, Any] = classifier_dropout class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): if self.task == "multiple-choice": lowerCamelCase : List[str] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase : List[str] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
287
1
import gc import unittest from diffusers import FlaxStableDiffusionInpaintPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) lowerCamelCase : Dict = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) lowerCamelCase : Optional[Any] = """xvjiarui/stable-diffusion-2-inpainting""" lowerCamelCase , lowerCamelCase : List[str] = FlaxStableDiffusionInpaintPipeline.from_pretrained(__magic_name__ , safety_checker=__magic_name__ ) lowerCamelCase : int = """Face of a yellow cat, high resolution, sitting on a park bench""" lowerCamelCase : Optional[Any] = jax.random.PRNGKey(0 ) lowerCamelCase : Tuple = 5_0 lowerCamelCase : Dict = jax.device_count() lowerCamelCase : Union[str, Any] = num_samples * [prompt] lowerCamelCase : Optional[Any] = num_samples * [init_image] lowerCamelCase : Dict = num_samples * [mask_image] lowerCamelCase , lowerCamelCase , lowerCamelCase : Dict = pipeline.prepare_inputs(__magic_name__ , __magic_name__ , __magic_name__ ) # shard inputs and rng lowerCamelCase : List[Any] = replicate(__magic_name__ ) lowerCamelCase : str = jax.random.split(__magic_name__ , jax.device_count() ) lowerCamelCase : Dict = shard(__magic_name__ ) lowerCamelCase : List[Any] = shard(__magic_name__ ) lowerCamelCase : int = shard(__magic_name__ ) lowerCamelCase : Optional[Any] = pipeline( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , jit=__magic_name__ ) lowerCamelCase : List[str] = output.images.reshape(__magic_name__ , 5_1_2 , 5_1_2 , 3 ) lowerCamelCase : Optional[Any] = images[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] lowerCamelCase : Union[str, Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCamelCase : Optional[Any] = jnp.array( [0.3_611_307, 0.37_649_736, 0.3_757_408, 0.38_213_953, 0.39_295_167, 0.3_841_631, 0.41_554_978, 0.4_137_475, 0.4_217_084] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
287
_lowerCamelCase ={ "joule": 1.0, "kilojoule": 1_0_0_0, "megajoule": 1_0_0_0_0_0_0, "gigajoule": 1_0_0_0_0_0_0_0_0_0, "wattsecond": 1.0, "watthour": 3_6_0_0, "kilowatthour": 3_6_0_0_0_0_0, "newtonmeter": 1.0, "calorie_nutr": 4_1_8_6.8, "kilocalorie_nutr": 4_1_8_6_8_0_0.0_0, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1_0_5_5.0_5_5_8_5, "footpound": 1.355818, } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowerCamelCase : Dict = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {", ".join(lowerCamelCase )}''' ) raise ValueError(lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
287
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCamelCase ={ """configuration_whisper""": ["""WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """WhisperConfig""", """WhisperOnnxConfig"""], """feature_extraction_whisper""": ["""WhisperFeatureExtractor"""], """processing_whisper""": ["""WhisperProcessor"""], """tokenization_whisper""": ["""WhisperTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =["""WhisperTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =[ """WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST""", """WhisperForConditionalGeneration""", """WhisperModel""", """WhisperPreTrainedModel""", """WhisperForAudioClassification""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =[ """TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFWhisperForConditionalGeneration""", """TFWhisperModel""", """TFWhisperPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =[ """FlaxWhisperForConditionalGeneration""", """FlaxWhisperModel""", """FlaxWhisperPreTrainedModel""", """FlaxWhisperForAudioClassification""", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _lowerCamelCase =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
287
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=__SCREAMING_SNAKE_CASE) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = field(default="""automatic-speech-recognition""" , metadata={"""include_in_asdict_even_if_is_default""": True}) _UpperCAmelCase : ClassVar[Features] = Features({"""audio""": Audio()}) _UpperCAmelCase : ClassVar[Features] = Features({"""transcription""": Value("""string""")}) _UpperCAmelCase : str = "audio" _UpperCAmelCase : str = "transcription" def UpperCamelCase__ ( self , __magic_name__ ): if self.audio_column not in features: raise ValueError(F'''Column {self.audio_column} is not present in features.''' ) if not isinstance(features[self.audio_column] , __magic_name__ ): raise ValueError(F'''Column {self.audio_column} is not an Audio type.''' ) lowerCamelCase : Optional[Any] = copy.deepcopy(self ) lowerCamelCase : List[Any] = self.input_schema.copy() lowerCamelCase : Tuple = features[self.audio_column] lowerCamelCase : int = input_schema return task_template @property def UpperCamelCase__ ( self ): return {self.audio_column: "audio", self.transcription_column: "transcription"}
287
1
import argparse import os import re _lowerCamelCase ="""src/diffusers""" # Pattern that looks at the indentation in a line. _lowerCamelCase =re.compile(R"""^(\s*)\S""") # Pattern that matches `"key":" and puts `key` in group 0. _lowerCamelCase =re.compile(R"""^\s*\"([^\"]+)\":""") # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. _lowerCamelCase =re.compile(R"""^\s*_import_structure\[\"([^\"]+)\"\]""") # Pattern that matches `"key",` and puts `key` in group 0. _lowerCamelCase =re.compile(R"""^\s*\"([^\"]+)\",\s*$""") # Pattern that matches any `[stuff]` and puts `stuff` in group 0. _lowerCamelCase =re.compile(R"""\[([^\]]+)\]""") def _a ( lowerCamelCase ): lowerCamelCase : Optional[Any] = _re_indent.search(lowerCamelCase ) return "" if search is None else search.groups()[0] def _a ( lowerCamelCase, lowerCamelCase="", lowerCamelCase=None, lowerCamelCase=None ): lowerCamelCase : Dict = 0 lowerCamelCase : List[Any] = code.split("""\n""" ) if start_prompt is not None: while not lines[index].startswith(lowerCamelCase ): index += 1 lowerCamelCase : List[str] = ["""\n""".join(lines[:index] )] else: lowerCamelCase : Optional[Any] = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). lowerCamelCase : Any = [lines[index]] index += 1 while index < len(lowerCamelCase ) and (end_prompt is None or not lines[index].startswith(lowerCamelCase )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(lowerCamelCase ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + """ """ ): current_block.append(lines[index] ) blocks.append("""\n""".join(lowerCamelCase ) ) if index < len(lowerCamelCase ) - 1: lowerCamelCase : Dict = [lines[index + 1]] index += 1 else: lowerCamelCase : List[Any] = [] else: blocks.append("""\n""".join(lowerCamelCase ) ) lowerCamelCase : Optional[Any] = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(lowerCamelCase ) > 0: blocks.append("""\n""".join(lowerCamelCase ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(lowerCamelCase ): blocks.append("""\n""".join(lines[index:] ) ) return blocks def _a ( lowerCamelCase ): def _inner(lowerCamelCase ): return key(lowerCamelCase ).lower().replace("""_""", """""" ) return _inner def _a ( lowerCamelCase, lowerCamelCase=None ): # If no key is provided, we use a noop. def noop(lowerCamelCase ): return x if key is None: lowerCamelCase : Tuple = noop # Constants are all uppercase, they go first. lowerCamelCase : List[Any] = [obj for obj in objects if key(lowerCamelCase ).isupper()] # Classes are not all uppercase but start with a capital, they go second. lowerCamelCase : List[Any] = [obj for obj in objects if key(lowerCamelCase )[0].isupper() and not key(lowerCamelCase ).isupper()] # Functions begin with a lowercase, they go last. lowerCamelCase : Optional[Any] = [obj for obj in objects if not key(lowerCamelCase )[0].isupper()] lowerCamelCase : List[Any] = ignore_underscore(lowerCamelCase ) return sorted(lowerCamelCase, key=lowerCamelCase ) + sorted(lowerCamelCase, key=lowerCamelCase ) + sorted(lowerCamelCase, key=lowerCamelCase ) def _a ( lowerCamelCase ): # This inner function sort imports between [ ]. def _replace(lowerCamelCase ): lowerCamelCase : Union[str, Any] = match.groups()[0] if "," not in imports: return F'''[{imports}]''' lowerCamelCase : Dict = [part.strip().replace("""\"""", """""" ) for part in imports.split(""",""" )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: lowerCamelCase : Any = keys[:-1] return "[" + ", ".join([F'''"{k}"''' for k in sort_objects(lowerCamelCase )] ) + "]" lowerCamelCase : int = import_statement.split("""\n""" ) if len(lowerCamelCase ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. lowerCamelCase : Any = 2 if lines[1].strip() == """[""" else 1 lowerCamelCase : List[Any] = [(i, _re_strip_line.search(lowerCamelCase ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] lowerCamelCase : str = sort_objects(lowerCamelCase, key=lambda lowerCamelCase : x[1] ) lowerCamelCase : List[str] = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(lowerCamelCase ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: lowerCamelCase : Optional[int] = _re_bracket_content.sub(_replace, lines[1] ) else: lowerCamelCase : int = [part.strip().replace("""\"""", """""" ) for part in lines[1].split(""",""" )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: lowerCamelCase : Dict = keys[:-1] lowerCamelCase : str = get_indent(lines[1] ) + """, """.join([F'''"{k}"''' for k in sort_objects(lowerCamelCase )] ) return "\n".join(lowerCamelCase ) else: # Finally we have to deal with imports fitting on one line lowerCamelCase : int = _re_bracket_content.sub(_replace, lowerCamelCase ) return import_statement def _a ( lowerCamelCase, lowerCamelCase=True ): with open(lowerCamelCase, """r""" ) as f: lowerCamelCase : str = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 lowerCamelCase : List[Any] = split_code_in_indented_blocks( lowerCamelCase, start_prompt="""_import_structure = {""", end_prompt="""if TYPE_CHECKING:""" ) # We ignore block 0 (everything until start_prompt) and the last block (everything after end_prompt). for block_idx in range(1, len(lowerCamelCase ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. lowerCamelCase : Dict = main_blocks[block_idx] lowerCamelCase : Union[str, Any] = block.split("""\n""" ) # Get to the start of the imports. lowerCamelCase : int = 0 while line_idx < len(lowerCamelCase ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: lowerCamelCase : int = len(lowerCamelCase ) else: line_idx += 1 if line_idx >= len(lowerCamelCase ): continue # Ignore beginning and last line: they don't contain anything. lowerCamelCase : Optional[int] = """\n""".join(block_lines[line_idx:-1] ) lowerCamelCase : Tuple = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. lowerCamelCase : str = split_code_in_indented_blocks(lowerCamelCase, indent_level=lowerCamelCase ) # We have two categories of import key: list or _import_structure[key].append/extend lowerCamelCase : List[str] = _re_direct_key if """_import_structure""" in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. lowerCamelCase : List[str] = [(pattern.search(lowerCamelCase ).groups()[0] if pattern.search(lowerCamelCase ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. lowerCamelCase : Union[str, Any] = [(i, key) for i, key in enumerate(lowerCamelCase ) if key is not None] lowerCamelCase : List[str] = [x[0] for x in sorted(lowerCamelCase, key=lambda lowerCamelCase : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. lowerCamelCase : Optional[Any] = 0 lowerCamelCase : Union[str, Any] = [] for i in range(len(lowerCamelCase ) ): if keys[i] is None: reordered_blocks.append(internal_blocks[i] ) else: lowerCamelCase : Optional[int] = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reordered_blocks.append(lowerCamelCase ) count += 1 # And we put our main block back together with its first and last line. lowerCamelCase : Optional[Any] = """\n""".join(block_lines[:line_idx] + reordered_blocks + [block_lines[-1]] ) if code != "\n".join(lowerCamelCase ): if check_only: return True else: print(F'''Overwriting {file}.''' ) with open(lowerCamelCase, """w""" ) as f: f.write("""\n""".join(lowerCamelCase ) ) def _a ( lowerCamelCase=True ): lowerCamelCase : Any = [] for root, _, files in os.walk(lowerCamelCase ): if "__init__.py" in files: lowerCamelCase : Tuple = sort_imports(os.path.join(lowerCamelCase, """__init__.py""" ), check_only=lowerCamelCase ) if result: lowerCamelCase : int = [os.path.join(lowerCamelCase, """__init__.py""" )] if len(lowerCamelCase ) > 0: raise ValueError(F'''Would overwrite {len(lowerCamelCase )} files, run `make style`.''' ) if __name__ == "__main__": _lowerCamelCase =argparse.ArgumentParser() parser.add_argument("""--check_only""", action="""store_true""", help="""Whether to only check or fix style.""") _lowerCamelCase =parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
287
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower_vision_model""" def __init__( self , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=3 , __magic_name__=1_6 , __magic_name__=2_8_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__=True , __magic_name__=False , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : str = num_hidden_layers lowerCamelCase : Optional[int] = num_channels lowerCamelCase : List[str] = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Any = initializer_factor lowerCamelCase : Tuple = layer_norm_eps lowerCamelCase : Tuple = stop_gradient lowerCamelCase : Optional[int] = share_layernorm lowerCamelCase : str = remove_last_layer @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : str = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """bridgetower_text_model""" def __init__( self , __magic_name__=5_0_2_6_5 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=1 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_4 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Any = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : Tuple = hidden_act lowerCamelCase : Optional[int] = initializer_factor lowerCamelCase : Any = intermediate_size lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : str = max_position_embeddings lowerCamelCase : Union[str, Any] = type_vocab_size lowerCamelCase : Optional[int] = layer_norm_eps lowerCamelCase : Optional[int] = position_embedding_type lowerCamelCase : List[str] = use_cache lowerCamelCase : List[str] = pad_token_id lowerCamelCase : List[str] = bos_token_id lowerCamelCase : Optional[int] = eos_token_id @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : Optional[int] = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower""" def __init__( self , __magic_name__=True , __magic_name__="gelu" , __magic_name__=7_6_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__="add" , __magic_name__=1_2 , __magic_name__=6 , __magic_name__=False , __magic_name__=False , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): # TODO: remove this once the Hub files are updated. lowerCamelCase : int = kwargs.pop("""text_config_dict""" , __magic_name__ ) lowerCamelCase : str = kwargs.pop("""vision_config_dict""" , __magic_name__ ) super().__init__(**__magic_name__ ) lowerCamelCase : str = share_cross_modal_transformer_layers lowerCamelCase : Union[str, Any] = hidden_act lowerCamelCase : str = hidden_size lowerCamelCase : Tuple = initializer_factor lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = share_link_tower_layers lowerCamelCase : List[Any] = link_tower_type lowerCamelCase : Tuple = num_attention_heads lowerCamelCase : int = num_hidden_layers lowerCamelCase : Union[str, Any] = tie_word_embeddings lowerCamelCase : Tuple = init_layernorm_from_vision_encoder if text_config is None: lowerCamelCase : Any = {} logger.info("""`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.""" ) if vision_config is None: lowerCamelCase : int = {} logger.info("""`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.""" ) lowerCamelCase : Any = BridgeTowerTextConfig(**__magic_name__ ) lowerCamelCase : Optional[Any] = BridgeTowerVisionConfig(**__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , **__magic_name__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = copy.deepcopy(self.__dict__ ) lowerCamelCase : int = self.text_config.to_dict() lowerCamelCase : Dict = self.vision_config.to_dict() lowerCamelCase : List[str] = self.__class__.model_type return output
287
1
from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class A__ ( yaml.SafeLoader): def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : List[Any] = [self.constructed_objects[key_node] for key_node, _ in node.value] lowerCamelCase : Optional[int] = [tuple(__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else key for key in keys] lowerCamelCase : Any = Counter(__magic_name__ ) lowerCamelCase : str = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F'''Got duplicate yaml keys: {duplicate_keys}''' ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__=False ): lowerCamelCase : Dict = super().construct_mapping(__magic_name__ , deep=__magic_name__ ) self._check_no_duplicates_on_constructed_node(__magic_name__ ) return mapping def _a ( lowerCamelCase ): lowerCamelCase : Dict = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: lowerCamelCase : Optional[Any] = full_content[1:].index("""---""" ) + 1 lowerCamelCase : Any = """\n""".join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(lowerCamelCase ) class A__ ( __SCREAMING_SNAKE_CASE): # class attributes _UpperCAmelCase : Tuple = {"""train_eval_index"""} # train-eval-index in the YAML metadata @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): with open(__magic_name__ , encoding="""utf-8""" ) as readme_file: lowerCamelCase , lowerCamelCase : Dict = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(__magic_name__ ) else: return cls() def UpperCamelCase__ ( self , __magic_name__ ): if path.exists(): with open(__magic_name__ , encoding="""utf-8""" ) as readme_file: lowerCamelCase : List[str] = readme_file.read() else: lowerCamelCase : Tuple = None lowerCamelCase : Optional[int] = self._to_readme(__magic_name__ ) with open(__magic_name__ , """w""" , encoding="""utf-8""" ) as readme_file: readme_file.write(__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ = None ): if readme_content is not None: lowerCamelCase , lowerCamelCase : List[str] = _split_yaml_from_readme(__magic_name__ ) lowerCamelCase : Any = """---\n""" + self.to_yaml_string() + """---\n""" + content else: lowerCamelCase : List[Any] = """---\n""" + self.to_yaml_string() + """---\n""" return full_content @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): lowerCamelCase : Optional[Any] = yaml.load(__magic_name__ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields lowerCamelCase : List[Any] = { (key.replace("""-""" , """_""" ) if key.replace("""-""" , """_""" ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**__magic_name__ ) def UpperCamelCase__ ( self ): return yaml.safe_dump( { (key.replace("""_""" , """-""" ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=__magic_name__ , allow_unicode=__magic_name__ , encoding="""utf-8""" , ).decode("""utf-8""" ) _lowerCamelCase ={ """image-classification""": [], """translation""": [], """image-segmentation""": [], """fill-mask""": [], """automatic-speech-recognition""": [], """token-classification""": [], """sentence-similarity""": [], """audio-classification""": [], """question-answering""": [], """summarization""": [], """zero-shot-classification""": [], """table-to-text""": [], """feature-extraction""": [], """other""": [], """multiple-choice""": [], """text-classification""": [], """text-to-image""": [], """text2text-generation""": [], """zero-shot-image-classification""": [], """tabular-classification""": [], """tabular-regression""": [], """image-to-image""": [], """tabular-to-text""": [], """unconditional-image-generation""": [], """text-retrieval""": [], """text-to-speech""": [], """object-detection""": [], """audio-to-audio""": [], """text-generation""": [], """conversational""": [], """table-question-answering""": [], """visual-question-answering""": [], """image-to-text""": [], """reinforcement-learning""": [], """voice-activity-detection""": [], """time-series-forecasting""": [], """document-question-answering""": [], } if __name__ == "__main__": from argparse import ArgumentParser _lowerCamelCase =ArgumentParser(usage="""Validate the yaml metadata block of a README.md file.""") ap.add_argument("""readme_filepath""") _lowerCamelCase =ap.parse_args() _lowerCamelCase =Path(args.readme_filepath) _lowerCamelCase =DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
287
def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = set(range(3, lowerCamelCase, 2 ) ) primes.add(2 ) for p in range(3, lowerCamelCase, 2 ): if p not in primes: continue primes.difference_update(set(range(p * p, lowerCamelCase, lowerCamelCase ) ) ) lowerCamelCase : Any = [float(lowerCamelCase ) for n in range(limit + 1 )] for p in primes: for n in range(lowerCamelCase, limit + 1, lowerCamelCase ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'''{solution() = }''')
287
1
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class A__ ( unittest.TestCase): def __init__( self , __magic_name__ , __magic_name__=1_3 , __magic_name__=7 , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=9_9 , __magic_name__=3_2 , __magic_name__=5 , __magic_name__=4 , __magic_name__=3_7 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=1_6 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=4 , ): lowerCamelCase : Any = parent lowerCamelCase : Optional[Any] = batch_size lowerCamelCase : str = seq_length lowerCamelCase : str = is_training lowerCamelCase : str = use_attention_mask lowerCamelCase : Any = use_token_type_ids lowerCamelCase : Dict = use_labels lowerCamelCase : str = vocab_size lowerCamelCase : Any = hidden_size lowerCamelCase : Union[str, Any] = num_hidden_layers lowerCamelCase : Optional[Any] = num_attention_heads lowerCamelCase : Tuple = intermediate_size lowerCamelCase : Any = hidden_act lowerCamelCase : List[Any] = hidden_dropout_prob lowerCamelCase : Any = attention_probs_dropout_prob lowerCamelCase : int = max_position_embeddings lowerCamelCase : str = type_vocab_size lowerCamelCase : Dict = type_sequence_label_size lowerCamelCase : Tuple = initializer_range lowerCamelCase : str = num_choices def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase : Optional[int] = None if self.use_attention_mask: lowerCamelCase : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase : str = None if self.use_token_type_ids: lowerCamelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase : int = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__magic_name__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCamelCase__ ( self ): lowerCamelCase : Dict = self.prepare_config_and_inputs() lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Union[str, Any] = config_and_inputs lowerCamelCase : List[str] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = self.prepare_config_and_inputs() lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : int = config_and_inputs lowerCamelCase : Tuple = True lowerCamelCase : Any = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowerCamelCase : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : int = True _UpperCAmelCase : Any = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = FlaxRobertaModelTester(self ) @slow def UpperCamelCase__ ( self ): for model_class_name in self.all_model_classes: lowerCamelCase : Tuple = model_class_name.from_pretrained("""roberta-base""" , from_pt=__magic_name__ ) lowerCamelCase : Optional[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(__magic_name__ )
287
from __future__ import annotations import collections import pprint from pathlib import Path def _a ( lowerCamelCase ): return "".join(sorted(lowerCamelCase ) ) def _a ( lowerCamelCase ): return word_by_signature[signature(lowerCamelCase )] _lowerCamelCase =Path(__file__).parent.joinpath("""words.txt""").read_text(encoding="""utf-8""") _lowerCamelCase =sorted({word.strip().lower() for word in data.splitlines()}) _lowerCamelCase =collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _lowerCamelCase ={word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open("""anagrams.txt""", """w""") as file: file.write("""all_anagrams = \n """) file.write(pprint.pformat(all_anagrams))
287
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """allenai/longformer-base-4096""": """https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json""", """allenai/longformer-large-4096""": """https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json""", """allenai/longformer-large-4096-finetuned-triviaqa""": ( """https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json""" ), """allenai/longformer-base-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json""" ), """allenai/longformer-large-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[Any] = """longformer""" def __init__( self , __magic_name__ = 5_1_2 , __magic_name__ = 2 , __magic_name__ = 1 , __magic_name__ = 0 , __magic_name__ = 2 , __magic_name__ = 3_0_5_2_2 , __magic_name__ = 7_6_8 , __magic_name__ = 1_2 , __magic_name__ = 1_2 , __magic_name__ = 3_0_7_2 , __magic_name__ = "gelu" , __magic_name__ = 0.1 , __magic_name__ = 0.1 , __magic_name__ = 5_1_2 , __magic_name__ = 2 , __magic_name__ = 0.02 , __magic_name__ = 1e-12 , __magic_name__ = False , **__magic_name__ , ): super().__init__(pad_token_id=__magic_name__ , **__magic_name__ ) lowerCamelCase : Dict = attention_window lowerCamelCase : Union[str, Any] = sep_token_id lowerCamelCase : Optional[int] = bos_token_id lowerCamelCase : int = eos_token_id lowerCamelCase : List[Any] = vocab_size lowerCamelCase : Dict = hidden_size lowerCamelCase : Optional[Any] = num_hidden_layers lowerCamelCase : Optional[int] = num_attention_heads lowerCamelCase : List[Any] = hidden_act lowerCamelCase : Any = intermediate_size lowerCamelCase : Any = hidden_dropout_prob lowerCamelCase : Tuple = attention_probs_dropout_prob lowerCamelCase : Any = max_position_embeddings lowerCamelCase : Dict = type_vocab_size lowerCamelCase : int = initializer_range lowerCamelCase : Optional[Any] = layer_norm_eps lowerCamelCase : Union[str, Any] = onnx_export class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , __magic_name__ , __magic_name__ = "default" , __magic_name__ = None ): super().__init__(__magic_name__ , __magic_name__ , __magic_name__ ) lowerCamelCase : Optional[int] = True @property def UpperCamelCase__ ( self ): if self.task == "multiple-choice": lowerCamelCase : Any = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""global_attention_mask""", dynamic_axis), ] ) @property def UpperCamelCase__ ( self ): lowerCamelCase : str = super().outputs if self.task == "default": lowerCamelCase : Any = {0: """batch"""} return outputs @property def UpperCamelCase__ ( self ): return 1e-4 @property def UpperCamelCase__ ( self ): # needs to be >= 14 to support tril operator return max(super().default_onnx_opset , 1_4 ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ): lowerCamelCase : List[Any] = super().generate_dummy_inputs( preprocessor=__magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly lowerCamelCase : Any = torch.zeros_like(inputs["""input_ids"""] ) # make every second token global lowerCamelCase : Union[str, Any] = 1 return inputs
287
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = ["""pixel_values"""] def __init__( self , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , __magic_name__ = PILImageResampling.BILINEAR , __magic_name__ = True , __magic_name__ = 1 / 2_5_5 , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = size if size is not None else {"""shortest_edge""": 3_8_4} lowerCamelCase : Tuple = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Dict = do_resize lowerCamelCase : List[Any] = size # Default value set here for backwards compatibility where the value in config is None lowerCamelCase : Any = crop_pct if crop_pct is not None else 2_2_4 / 2_5_6 lowerCamelCase : Union[str, Any] = resample lowerCamelCase : str = do_rescale lowerCamelCase : Union[str, Any] = rescale_factor lowerCamelCase : Tuple = do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCamelCase : int = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = PILImageResampling.BICUBIC , __magic_name__ = None , **__magic_name__ , ): lowerCamelCase : Union[str, Any] = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) lowerCamelCase : str = size["""shortest_edge"""] if shortest_edge < 3_8_4: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct lowerCamelCase : List[str] = int(shortest_edge / crop_pct ) lowerCamelCase : Optional[Any] = get_resize_output_image_size(__magic_name__ , size=__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Optional[int] = resize(image=__magic_name__ , size=__magic_name__ , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=__magic_name__ , size=(shortest_edge, shortest_edge) , data_format=__magic_name__ , **__magic_name__ ) else: # warping (no cropping) when evaluated at 384 or larger return resize( __magic_name__ , size=(shortest_edge, shortest_edge) , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return rescale(__magic_name__ , scale=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return normalize(__magic_name__ , mean=__magic_name__ , std=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = ChannelDimension.FIRST , **__magic_name__ , ): lowerCamelCase : str = do_resize if do_resize is not None else self.do_resize lowerCamelCase : Optional[Any] = crop_pct if crop_pct is not None else self.crop_pct lowerCamelCase : Optional[int] = resample if resample is not None else self.resample lowerCamelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else self.image_mean lowerCamelCase : Tuple = image_std if image_std is not None else self.image_std lowerCamelCase : Dict = size if size is not None else self.size lowerCamelCase : str = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : List[str] = make_list_of_images(__magic_name__ ) if not valid_images(__magic_name__ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_resize and size["shortest_edge"] < 3_8_4 and crop_pct is None: raise ValueError("""crop_pct must be specified if size < 384.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. lowerCamelCase : Optional[Any] = [to_numpy_array(__magic_name__ ) for image in images] if do_resize: lowerCamelCase : List[Any] = [self.resize(image=__magic_name__ , size=__magic_name__ , crop_pct=__magic_name__ , resample=__magic_name__ ) for image in images] if do_rescale: lowerCamelCase : Union[str, Any] = [self.rescale(image=__magic_name__ , scale=__magic_name__ ) for image in images] if do_normalize: lowerCamelCase : List[Any] = [self.normalize(image=__magic_name__ , mean=__magic_name__ , std=__magic_name__ ) for image in images] lowerCamelCase : Optional[int] = [to_channel_dimension_format(__magic_name__ , __magic_name__ ) for image in images] lowerCamelCase : List[str] = {"""pixel_values""": images} return BatchFeature(data=__magic_name__ , tensor_type=__magic_name__ )
287
1
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = """gpt_neo""" _UpperCAmelCase : Union[str, Any] = ["""past_key_values"""] _UpperCAmelCase : List[Any] = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , __magic_name__=5_0_2_5_7 , __magic_name__=2_0_4_8 , __magic_name__=2_0_4_8 , __magic_name__=2_4 , __magic_name__=[[["global", "local"], 1_2]] , __magic_name__=1_6 , __magic_name__=None , __magic_name__=2_5_6 , __magic_name__="gelu_new" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=1e-5 , __magic_name__=0.02 , __magic_name__=True , __magic_name__=5_0_2_5_6 , __magic_name__=5_0_2_5_6 , **__magic_name__ , ): lowerCamelCase : List[Any] = vocab_size lowerCamelCase : str = max_position_embeddings lowerCamelCase : str = hidden_size lowerCamelCase : Optional[int] = num_layers lowerCamelCase : str = num_heads lowerCamelCase : Optional[Any] = intermediate_size lowerCamelCase : List[Any] = window_size lowerCamelCase : int = activation_function lowerCamelCase : Union[str, Any] = resid_dropout lowerCamelCase : List[Any] = embed_dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Dict = classifier_dropout lowerCamelCase : Any = layer_norm_epsilon lowerCamelCase : Dict = initializer_range lowerCamelCase : Dict = use_cache lowerCamelCase : Optional[Any] = bos_token_id lowerCamelCase : int = eos_token_id lowerCamelCase : List[Any] = attention_types lowerCamelCase : Optional[Any] = self.expand_attention_types_params(__magic_name__ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.attention_layers)` == `config.num_layers` """ F'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' """`config.attention_layers` is prepared using `config.attention_types`. """ """Please verify the value of `config.attention_types` argument.""" ) super().__init__(bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ ): lowerCamelCase : Optional[int] = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : Any = input.size() lowerCamelCase : List[Any] = len(lowerCamelCase ) lowerCamelCase : Optional[Any] = shape[dimension] lowerCamelCase : Optional[int] = torch.arange(0, lowerCamelCase, lowerCamelCase ) lowerCamelCase : Dict = torch.div(sizedim - size, lowerCamelCase, rounding_mode="""floor""" ) + 1 lowerCamelCase : int = torch.arange(lowerCamelCase ) + low_indices[:min_length][:, None] lowerCamelCase : str = [slice(lowerCamelCase )] * rank lowerCamelCase : List[str] = indices lowerCamelCase : Dict = input[s] lowerCamelCase : Any = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : List[Any] = torch.arange(1, lowerCamelCase ) lowerCamelCase : Optional[int] = torch.remainder(lowerCamelCase, lowerCamelCase ) lowerCamelCase : List[Any] = remainders == 0 lowerCamelCase : List[Any] = candidates[divisor_indices] lowerCamelCase : Optional[Any] = torch.max(lowerCamelCase ) return largest_divisor, torch.div(lowerCamelCase, lowerCamelCase, rounding_mode="""floor""" ) class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): lowerCamelCase : str = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(__magic_name__ , direction="""inputs""" ) lowerCamelCase : int = {0: """batch""", 1: """past_sequence + sequence"""} else: lowerCamelCase : Tuple = {0: """batch""", 1: """sequence"""} return common_inputs @property def UpperCamelCase__ ( self ): return self._config.num_heads def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ): lowerCamelCase : Optional[int] = super(__magic_name__ , self ).generate_dummy_inputs( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) # We need to order the input in the way they appears in the forward() lowerCamelCase : int = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch lowerCamelCase , lowerCamelCase : Optional[Any] = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values lowerCamelCase : Optional[int] = seqlen + 2 lowerCamelCase : List[Any] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowerCamelCase : str = [ (torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) for _ in range(self.num_layers ) ] lowerCamelCase : Tuple = common_inputs["""attention_mask"""] if self.use_past: lowerCamelCase : str = ordered_inputs["""attention_mask"""].dtype lowerCamelCase : Any = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(__magic_name__ , __magic_name__ , dtype=__magic_name__ )] , dim=1 ) return ordered_inputs @property def UpperCamelCase__ ( self ): return 1_3
287
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 _lowerCamelCase ={ """return_dict""": False, """output_hidden_states""": True, """output_attentions""": True, """torchscript""": True, """torch_dtype""": """float16""", """use_bfloat16""": True, """tf_legacy_loss""": True, """pruned_heads""": {"""a""": 1}, """tie_word_embeddings""": False, """is_decoder""": True, """cross_attention_hidden_size""": 1_2_8, """add_cross_attention""": True, """tie_encoder_decoder""": True, """max_length""": 5_0, """min_length""": 3, """do_sample""": True, """early_stopping""": True, """num_beams""": 3, """num_beam_groups""": 3, """diversity_penalty""": 0.5, """temperature""": 2.0, """top_k""": 1_0, """top_p""": 0.7, """typical_p""": 0.2, """repetition_penalty""": 0.8, """length_penalty""": 0.8, """no_repeat_ngram_size""": 5, """encoder_no_repeat_ngram_size""": 5, """bad_words_ids""": [1, 2, 3], """num_return_sequences""": 3, """chunk_size_feed_forward""": 5, """output_scores""": True, """return_dict_in_generate""": True, """forced_bos_token_id""": 2, """forced_eos_token_id""": 3, """remove_invalid_values""": True, """architectures""": ["""BertModel"""], """finetuning_task""": """translation""", """id2label""": {0: """label"""}, """label2id""": {"""label""": """0"""}, """tokenizer_class""": """BertTokenizerFast""", """prefix""": """prefix""", """bos_token_id""": 6, """pad_token_id""": 7, """eos_token_id""": 8, """sep_token_id""": 9, """decoder_start_token_id""": 1_0, """exponential_decay_length_penalty""": (5, 1.01), """suppress_tokens""": [0, 1], """begin_suppress_tokens""": 2, """task_specific_params""": {"""translation""": """some_params"""}, """problem_type""": """regression""", } @is_staging_test class A__ ( unittest.TestCase): @classmethod def UpperCamelCase__ ( cls ): lowerCamelCase : int = TOKEN HfFolder.save_token(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls ): try: delete_repo(token=cls._token , repo_id="""test-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-config-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-config""" ) except HTTPError: pass def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""test-config""" , use_auth_token=self._token ) lowerCamelCase : Any = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__magic_name__ , repo_id="""test-config""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : Optional[Any] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""valid_org/test-config-org""" , use_auth_token=self._token ) lowerCamelCase : Optional[int] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( __magic_name__ , repo_id="""valid_org/test-config-org""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : List[str] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): CustomConfig.register_for_auto_class() lowerCamelCase : Optional[Any] = CustomConfig(attribute=4_2 ) config.push_to_hub("""test-dynamic-config""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"""AutoConfig""": """custom_configuration.CustomConfig"""} ) lowerCamelCase : List[str] = AutoConfig.from_pretrained(F'''{USER}/test-dynamic-config''' , trust_remote_code=__magic_name__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , """CustomConfig""" ) self.assertEqual(new_config.attribute , 4_2 ) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : str = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated lowerCamelCase : Optional[int] = c.n_embd + 1 # int lowerCamelCase : Optional[int] = c.resid_pdrop + 1.0 # float lowerCamelCase : Tuple = not c.scale_attn_weights # bool lowerCamelCase : Any = c.summary_type + """foo""" # str c.update_from_string( F'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(__magic_name__ , c.n_embd , """mismatch for key: n_embd""" ) self.assertEqual(__magic_name__ , c.resid_pdrop , """mismatch for key: resid_pdrop""" ) self.assertEqual(__magic_name__ , c.scale_attn_weights , """mismatch for key: scale_attn_weights""" ) self.assertEqual(__magic_name__ , c.summary_type , """mismatch for key: summary_type""" ) def UpperCamelCase__ ( self ): lowerCamelCase : str = PretrainedConfig() lowerCamelCase : int = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( __magic_name__ , ["""is_encoder_decoder""", """_name_or_path""", """_commit_hash""", """transformers_version"""] ) lowerCamelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(__magic_name__ , __magic_name__ )] if len(__magic_name__ ) > 0: raise ValueError( """The following keys are set with the default values in""" """ `test_configuration_common.config_common_kwargs` pick another value for them:""" F''' {", ".join(__magic_name__ )}.''' ) def UpperCamelCase__ ( self ): with self.assertRaises(__magic_name__ ): # config is in subfolder, the following should not work without specifying the subfolder lowerCamelCase : Dict = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" ) lowerCamelCase : str = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" , subfolder="""bert""" ) self.assertIsNotNone(__magic_name__ ) def UpperCamelCase__ ( self ): # A mock response for an HTTP head request to emulate server down lowerCamelCase : Dict = mock.Mock() lowerCamelCase : Optional[int] = 5_0_0 lowerCamelCase : List[Any] = {} lowerCamelCase : Tuple = HTTPError lowerCamelCase : Union[str, Any] = {} # Download this model to make sure it's in the cache. lowerCamelCase : List[str] = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=__magic_name__ ) as mock_head: lowerCamelCase : Any = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase__ ( self ): # This test is for deprecated behavior and can be removed in v5 lowerCamelCase : List[str] = BertConfig.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = AutoConfig.from_pretrained("""bert-base-cased""" ) lowerCamelCase : Optional[Any] = ["""config.4.0.0.json"""] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(__magic_name__ ) lowerCamelCase : str = 2 json.dump(configuration.to_dict() , open(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , """w""" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 lowerCamelCase : Any = ["""config.42.0.0.json"""] lowerCamelCase : Optional[Any] = 7_6_8 configuration.save_pretrained(__magic_name__ ) shutil.move(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , os.path.join(__magic_name__ , """config.42.0.0.json""" ) ) lowerCamelCase : int = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 7_6_8 ) def UpperCamelCase__ ( self ): # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. lowerCamelCase : str = """hf-internal-testing/test-two-configs""" import transformers as new_transformers lowerCamelCase : Tuple = """v4.0.0""" lowerCamelCase , lowerCamelCase : Optional[int] = new_transformers.models.auto.AutoConfig.from_pretrained( __magic_name__ , return_unused_kwargs=__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(__magic_name__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers lowerCamelCase : Tuple = """v3.0.0""" lowerCamelCase : Any = old_transformers.models.auto.AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(old_configuration.hidden_size , 7_6_8 )
287
1
import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class A__ : def __init__( self , __magic_name__ , __magic_name__=sys.maxsize ): lowerCamelCase : Any = """bilinear""" lowerCamelCase : Optional[int] = max_size lowerCamelCase : Dict = short_edge_length def __call__( self , __magic_name__ ): lowerCamelCase : Optional[int] = [] for img in imgs: lowerCamelCase , lowerCamelCase : Optional[Any] = img.shape[:2] # later: provide list and randomly choose index for resize lowerCamelCase : Optional[Any] = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img lowerCamelCase : Optional[Any] = size * 1.0 / min(__magic_name__ , __magic_name__ ) if h < w: lowerCamelCase , lowerCamelCase : int = size, scale * w else: lowerCamelCase , lowerCamelCase : str = scale * h, size if max(__magic_name__ , __magic_name__ ) > self.max_size: lowerCamelCase : str = self.max_size * 1.0 / max(__magic_name__ , __magic_name__ ) lowerCamelCase : Tuple = newh * scale lowerCamelCase : str = neww * scale lowerCamelCase : Dict = int(neww + 0.5 ) lowerCamelCase : List[str] = int(newh + 0.5 ) if img.dtype == np.uinta: lowerCamelCase : Tuple = Image.fromarray(__magic_name__ ) lowerCamelCase : str = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) lowerCamelCase : List[str] = np.asarray(__magic_name__ ) else: lowerCamelCase : Optional[Any] = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw lowerCamelCase : List[str] = nn.functional.interpolate( __magic_name__ , (newh, neww) , mode=self.interp_method , align_corners=__magic_name__ ).squeeze(0 ) img_augs.append(__magic_name__ ) return img_augs class A__ : def __init__( self , __magic_name__ ): lowerCamelCase : str = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) lowerCamelCase : Any = cfg.INPUT.FORMAT lowerCamelCase : str = cfg.SIZE_DIVISIBILITY lowerCamelCase : List[str] = cfg.PAD_VALUE lowerCamelCase : Any = cfg.INPUT.MAX_SIZE_TEST lowerCamelCase : List[Any] = cfg.MODEL.DEVICE lowerCamelCase : Union[str, Any] = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) lowerCamelCase : Optional[int] = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) lowerCamelCase : Optional[Any] = lambda __magic_name__ : (x - self.pixel_mean) / self.pixel_std def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : List[Any] = tuple(max(__magic_name__ ) for s in zip(*[img.shape for img in images] ) ) lowerCamelCase : Union[str, Any] = [im.shape[-2:] for im in images] lowerCamelCase : Optional[Any] = [ nn.functional.pad( __magic_name__ , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(__magic_name__ , __magic_name__ ) ] return torch.stack(__magic_name__ ), torch.tensor(__magic_name__ ) def __call__( self , __magic_name__ , __magic_name__=False ): with torch.no_grad(): if not isinstance(__magic_name__ , __magic_name__ ): lowerCamelCase : Optional[int] = [images] if single_image: assert len(__magic_name__ ) == 1 for i in range(len(__magic_name__ ) ): if isinstance(images[i] , torch.Tensor ): images.insert(__magic_name__ , images.pop(__magic_name__ ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( __magic_name__ , torch.as_tensor(img_tensorize(images.pop(__magic_name__ ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge lowerCamelCase : List[Any] = torch.tensor([im.shape[:2] for im in images] ) lowerCamelCase : Any = self.aug(__magic_name__ ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic lowerCamelCase : List[str] = [self.normalizer(__magic_name__ ) for x in images] # now pad them to do the following operations lowerCamelCase , lowerCamelCase : Union[str, Any] = self.pad(__magic_name__ ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad lowerCamelCase : Optional[int] = torch.true_divide(__magic_name__ , __magic_name__ ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def _a ( lowerCamelCase, lowerCamelCase ): boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def _a ( lowerCamelCase, lowerCamelCase ): assert torch.isfinite(lowerCamelCase ).all(), "Box tensor contains infinite or NaN!" lowerCamelCase , lowerCamelCase : List[str] = box_size tensor[:, 0].clamp_(min=0, max=lowerCamelCase ) tensor[:, 1].clamp_(min=0, max=lowerCamelCase ) tensor[:, 2].clamp_(min=0, max=lowerCamelCase ) tensor[:, 3].clamp_(min=0, max=lowerCamelCase )
287
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) lowerCamelCase : Any = tf.convert_to_tensor( [[5, 1_2_1, 1_1, 6_6_0, 1_6, 7_3_0, 2_5_5_4_3, 1_1_0, 8_3, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowerCamelCase : str = model(__magic_name__ )["""last_hidden_state"""] lowerCamelCase : Union[str, Any] = tf.TensorShape((1, 1_0, 7_6_8) ) self.assertEqual(output.shape , __magic_name__ ) # compare the actual values for a slice. lowerCamelCase : Dict = tf.convert_to_tensor( [[[-0.0_254, 0.0_235, 0.1_027], [0.0_606, -0.1_811, -0.0_418], [-0.1_561, -0.1_127, 0.2_687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
287
1
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 _lowerCamelCase =sys.version_info >= (3, 1_0) def _a ( lowerCamelCase=None, lowerCamelCase=None ): return field(default_factory=lambda: default, metadata=lowerCamelCase ) @dataclass class A__ : _UpperCAmelCase : int _UpperCAmelCase : float _UpperCAmelCase : str _UpperCAmelCase : bool @dataclass class A__ : _UpperCAmelCase : int = 42 _UpperCAmelCase : str = field(default="""toto""" , metadata={"""help""": """help message"""}) @dataclass class A__ : _UpperCAmelCase : bool = False _UpperCAmelCase : bool = True _UpperCAmelCase : Optional[bool] = None class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = """titi""" _UpperCAmelCase : List[str] = """toto""" class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = """titi""" _UpperCAmelCase : str = """toto""" _UpperCAmelCase : Any = 42 @dataclass class A__ : _UpperCAmelCase : BasicEnum = "toto" def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = BasicEnum(self.foo ) @dataclass class A__ : _UpperCAmelCase : MixedTypeEnum = "toto" def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = MixedTypeEnum(self.foo ) @dataclass class A__ : _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : Optional[float] = field(default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """help message"""}) _UpperCAmelCase : Optional[str] = None _UpperCAmelCase : Optional[List[str]] = list_field(default=[]) _UpperCAmelCase : Optional[List[int]] = list_field(default=[]) @dataclass class A__ : _UpperCAmelCase : List[int] = list_field(default=[]) _UpperCAmelCase : List[int] = list_field(default=[1, 2, 3]) _UpperCAmelCase : List[str] = list_field(default=["""Hallo""", """Bonjour""", """Hello"""]) _UpperCAmelCase : List[float] = list_field(default=[0.1, 0.2, 0.3]) @dataclass class A__ : _UpperCAmelCase : List[int] = field() _UpperCAmelCase : str = field() _UpperCAmelCase : BasicEnum = field() def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = BasicEnum(self.required_enum ) @dataclass class A__ : _UpperCAmelCase : int _UpperCAmelCase : "BasicEnum" = field() _UpperCAmelCase : "Optional[bool]" = None _UpperCAmelCase : "str" = field(default="""toto""" , metadata={"""help""": """help message"""}) _UpperCAmelCase : "List[str]" = list_field(default=["""Hallo""", """Bonjour""", """Hello"""]) if is_python_no_less_than_3_10: @dataclass class A__ : _UpperCAmelCase : bool = False _UpperCAmelCase : bool = True _UpperCAmelCase : bool | None = None @dataclass class A__ : _UpperCAmelCase : int | None = None _UpperCAmelCase : float | None = field(default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """help message"""}) _UpperCAmelCase : str | None = None _UpperCAmelCase : list[str] | None = list_field(default=[]) _UpperCAmelCase : list[int] | None = list_field(default=[]) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): self.assertEqual(len(a._actions ) , len(b._actions ) ) for x, y in zip(a._actions , b._actions ): lowerCamelCase : Any = {k: v for k, v in vars(__magic_name__ ).items() if k != """container"""} lowerCamelCase : Optional[int] = {k: v for k, v in vars(__magic_name__ ).items() if k != """container"""} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get("""choices""" , __magic_name__ ) and yy.get("""choices""" , __magic_name__ ): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx["""type"""](__magic_name__ ) , yy["""type"""](__magic_name__ ) ) del xx["type"], yy["type"] self.assertEqual(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = HfArgumentParser(__magic_name__ ) lowerCamelCase : Union[str, Any] = argparse.ArgumentParser() expected.add_argument("""--foo""" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument("""--bar""" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument("""--baz""" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument("""--flag""" , type=__magic_name__ , default=__magic_name__ , const=__magic_name__ , nargs="""?""" ) self.argparsersEqual(__magic_name__ , __magic_name__ ) lowerCamelCase : List[Any] = ["""--foo""", """1""", """--baz""", """quux""", """--bar""", """0.5"""] ((lowerCamelCase) , ) : Tuple = parser.parse_args_into_dataclasses(__magic_name__ , look_for_args_file=__magic_name__ ) self.assertFalse(example.flag ) def UpperCamelCase__ ( self ): lowerCamelCase : str = HfArgumentParser(__magic_name__ ) lowerCamelCase : Tuple = argparse.ArgumentParser() expected.add_argument("""--foo""" , default=4_2 , type=__magic_name__ ) expected.add_argument("""--baz""" , default="""toto""" , type=__magic_name__ , help="""help message""" ) self.argparsersEqual(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = argparse.ArgumentParser() expected.add_argument("""--foo""" , type=__magic_name__ , default=__magic_name__ , const=__magic_name__ , nargs="""?""" ) expected.add_argument("""--baz""" , type=__magic_name__ , default=__magic_name__ , const=__magic_name__ , nargs="""?""" ) # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument("""--no_baz""" , action="""store_false""" , default=__magic_name__ , dest="""baz""" ) expected.add_argument("""--opt""" , type=__magic_name__ , default=__magic_name__ ) lowerCamelCase : List[str] = [WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(__magic_name__ ) for dataclass_type in dataclass_types: lowerCamelCase : Dict = HfArgumentParser(__magic_name__ ) self.argparsersEqual(__magic_name__ , __magic_name__ ) lowerCamelCase : Union[str, Any] = parser.parse_args([] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) lowerCamelCase : str = parser.parse_args(["""--foo""", """--no_baz"""] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) lowerCamelCase : Tuple = parser.parse_args(["""--foo""", """--baz"""] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) lowerCamelCase : List[Any] = parser.parse_args(["""--foo""", """True""", """--baz""", """True""", """--opt""", """True"""] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) lowerCamelCase : Any = parser.parse_args(["""--foo""", """False""", """--baz""", """False""", """--opt""", """False"""] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = HfArgumentParser(__magic_name__ ) lowerCamelCase : List[str] = argparse.ArgumentParser() expected.add_argument( """--foo""" , default="""toto""" , choices=["""titi""", """toto""", 4_2] , type=make_choice_type_function(["""titi""", """toto""", 4_2] ) , ) self.argparsersEqual(__magic_name__ , __magic_name__ ) lowerCamelCase : str = parser.parse_args([] ) self.assertEqual(args.foo , """toto""" ) lowerCamelCase : str = parser.parse_args_into_dataclasses([] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.toto ) lowerCamelCase : Dict = parser.parse_args(["""--foo""", """titi"""] ) self.assertEqual(args.foo , """titi""" ) lowerCamelCase : Optional[Any] = parser.parse_args_into_dataclasses(["""--foo""", """titi"""] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.titi ) lowerCamelCase : Any = parser.parse_args(["""--foo""", """42"""] ) self.assertEqual(args.foo , 4_2 ) lowerCamelCase : Tuple = parser.parse_args_into_dataclasses(["""--foo""", """42"""] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.fourtytwo ) def UpperCamelCase__ ( self ): @dataclass class A__ : _UpperCAmelCase : Literal["titi", "toto", 42] = "toto" lowerCamelCase : Optional[int] = HfArgumentParser(__magic_name__ ) lowerCamelCase : List[str] = argparse.ArgumentParser() expected.add_argument( """--foo""" , default="""toto""" , choices=("""titi""", """toto""", 4_2) , type=make_choice_type_function(["""titi""", """toto""", 4_2] ) , ) self.argparsersEqual(__magic_name__ , __magic_name__ ) lowerCamelCase : int = parser.parse_args([] ) self.assertEqual(args.foo , """toto""" ) lowerCamelCase : List[Any] = parser.parse_args(["""--foo""", """titi"""] ) self.assertEqual(args.foo , """titi""" ) lowerCamelCase : List[Any] = parser.parse_args(["""--foo""", """42"""] ) self.assertEqual(args.foo , 4_2 ) def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = HfArgumentParser(__magic_name__ ) lowerCamelCase : Dict = argparse.ArgumentParser() expected.add_argument("""--foo_int""" , nargs="""+""" , default=[] , type=__magic_name__ ) expected.add_argument("""--bar_int""" , nargs="""+""" , default=[1, 2, 3] , type=__magic_name__ ) expected.add_argument("""--foo_str""" , nargs="""+""" , default=["""Hallo""", """Bonjour""", """Hello"""] , type=__magic_name__ ) expected.add_argument("""--foo_float""" , nargs="""+""" , default=[0.1, 0.2, 0.3] , type=__magic_name__ ) self.argparsersEqual(__magic_name__ , __magic_name__ ) lowerCamelCase : Optional[Any] = parser.parse_args([] ) self.assertEqual( __magic_name__ , Namespace(foo_int=[] , bar_int=[1, 2, 3] , foo_str=["""Hallo""", """Bonjour""", """Hello"""] , foo_float=[0.1, 0.2, 0.3] ) , ) lowerCamelCase : Optional[int] = parser.parse_args("""--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7""".split() ) self.assertEqual(__magic_name__ , Namespace(foo_int=[1] , bar_int=[2, 3] , foo_str=["""a""", """b""", """c"""] , foo_float=[0.1, 0.7] ) ) def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = argparse.ArgumentParser() expected.add_argument("""--foo""" , default=__magic_name__ , type=__magic_name__ ) expected.add_argument("""--bar""" , default=__magic_name__ , type=__magic_name__ , help="""help message""" ) expected.add_argument("""--baz""" , default=__magic_name__ , type=__magic_name__ ) expected.add_argument("""--ces""" , nargs="""+""" , default=[] , type=__magic_name__ ) expected.add_argument("""--des""" , nargs="""+""" , default=[] , type=__magic_name__ ) lowerCamelCase : Optional[Any] = [OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(__magic_name__ ) for dataclass_type in dataclass_types: lowerCamelCase : Optional[Any] = HfArgumentParser(__magic_name__ ) self.argparsersEqual(__magic_name__ , __magic_name__ ) lowerCamelCase : Any = parser.parse_args([] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , bar=__magic_name__ , baz=__magic_name__ , ces=[] , des=[] ) ) lowerCamelCase : str = parser.parse_args("""--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3""".split() ) self.assertEqual(__magic_name__ , Namespace(foo=1_2 , bar=3.14 , baz="""42""" , ces=["""a""", """b""", """c"""] , des=[1, 2, 3] ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = HfArgumentParser(__magic_name__ ) lowerCamelCase : List[Any] = argparse.ArgumentParser() expected.add_argument("""--required_list""" , nargs="""+""" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument("""--required_str""" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument( """--required_enum""" , type=make_choice_type_function(["""titi""", """toto"""] ) , choices=["""titi""", """toto"""] , required=__magic_name__ , ) self.argparsersEqual(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = HfArgumentParser(__magic_name__ ) lowerCamelCase : List[str] = argparse.ArgumentParser() expected.add_argument("""--foo""" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument( """--required_enum""" , type=make_choice_type_function(["""titi""", """toto"""] ) , choices=["""titi""", """toto"""] , required=__magic_name__ , ) expected.add_argument("""--opt""" , type=__magic_name__ , default=__magic_name__ ) expected.add_argument("""--baz""" , default="""toto""" , type=__magic_name__ , help="""help message""" ) expected.add_argument("""--foo_str""" , nargs="""+""" , default=["""Hallo""", """Bonjour""", """Hello"""] , type=__magic_name__ ) self.argparsersEqual(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = HfArgumentParser(__magic_name__ ) lowerCamelCase : List[Any] = { """foo""": 1_2, """bar""": 3.14, """baz""": """42""", """flag""": True, } lowerCamelCase : Tuple = parser.parse_dict(__magic_name__ )[0] lowerCamelCase : Optional[Any] = BasicExample(**__magic_name__ ) self.assertEqual(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = HfArgumentParser(__magic_name__ ) lowerCamelCase : List[str] = { """foo""": 1_2, """bar""": 3.14, """baz""": """42""", """flag""": True, """extra""": 4_2, } self.assertRaises(__magic_name__ , parser.parse_dict , __magic_name__ , allow_extra_keys=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = HfArgumentParser(__magic_name__ ) lowerCamelCase : Tuple = { """foo""": 1_2, """bar""": 3.14, """baz""": """42""", """flag""": True, } with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : List[str] = os.path.join(__magic_name__ , """temp_json""" ) os.mkdir(__magic_name__ ) with open(temp_local_path + """.json""" , """w+""" ) as f: json.dump(__magic_name__ , __magic_name__ ) lowerCamelCase : str = parser.parse_yaml_file(Path(temp_local_path + """.json""" ) )[0] lowerCamelCase : Any = BasicExample(**__magic_name__ ) self.assertEqual(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = HfArgumentParser(__magic_name__ ) lowerCamelCase : List[str] = { """foo""": 1_2, """bar""": 3.14, """baz""": """42""", """flag""": True, } with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : int = os.path.join(__magic_name__ , """temp_yaml""" ) os.mkdir(__magic_name__ ) with open(temp_local_path + """.yaml""" , """w+""" ) as f: yaml.dump(__magic_name__ , __magic_name__ ) lowerCamelCase : List[Any] = parser.parse_yaml_file(Path(temp_local_path + """.yaml""" ) )[0] lowerCamelCase : List[Any] = BasicExample(**__magic_name__ ) self.assertEqual(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = HfArgumentParser(__magic_name__ ) self.assertIsNotNone(__magic_name__ )
287
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _lowerCamelCase =get_logger(__name__) class A__ : def __init__( self , __magic_name__ = None ): lowerCamelCase : Dict = ( os.path.join(__magic_name__ , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) lowerCamelCase : List[str] = Extractor def UpperCamelCase__ ( self , __magic_name__ ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" lowerCamelCase : int = os.path.abspath(__magic_name__ ) return os.path.join(self.extract_dir , hash_url_to_filename(__magic_name__ ) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): return force_extract or ( not os.path.isfile(__magic_name__ ) and not (os.path.isdir(__magic_name__ ) and os.listdir(__magic_name__ )) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = False ): lowerCamelCase : Union[str, Any] = self.extractor.infer_extractor_format(__magic_name__ ) if not extractor_format: return input_path lowerCamelCase : int = self._get_output_path(__magic_name__ ) if self._do_extract(__magic_name__ , __magic_name__ ): self.extractor.extract(__magic_name__ , __magic_name__ , __magic_name__ ) return output_path class A__ ( __SCREAMING_SNAKE_CASE): @classmethod @abstractmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): ... @staticmethod @abstractmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): ... class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[bytes] = [] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with open(__magic_name__ , """rb""" ) as f: return f.read(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if not magic_number: lowerCamelCase : Optional[Any] = max(len(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) try: lowerCamelCase : Tuple = cls.read_magic_number(__magic_name__ , __magic_name__ ) except OSError: return False return any(magic_number.startswith(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) class A__ ( __SCREAMING_SNAKE_CASE): @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): return tarfile.is_tarfile(__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): def resolved(__magic_name__ ) -> str: return os.path.realpath(os.path.abspath(__magic_name__ ) ) def badpath(__magic_name__ , __magic_name__ ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__magic_name__ , __magic_name__ ) ).startswith(__magic_name__ ) def badlink(__magic_name__ , __magic_name__ ) -> bool: # Links are interpreted relative to the directory containing the link lowerCamelCase : List[str] = resolved(os.path.join(__magic_name__ , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=__magic_name__ ) lowerCamelCase : Optional[Any] = resolved(__magic_name__ ) for finfo in members: if badpath(finfo.name , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Dict = tarfile.open(__magic_name__ ) tar_file.extractall(__magic_name__ , members=TarExtractor.safemembers(__magic_name__ , __magic_name__ ) ) tar_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = [B"""\x1F\x8B"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with gzip.open(__magic_name__ , """rb""" ) as gzip_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[int] = [ B"""PK\x03\x04""", B"""PK\x05\x06""", # empty archive B"""PK\x07\x08""", # spanned archive ] @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if super().is_extractable(__magic_name__ , magic_number=__magic_name__ ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__magic_name__ , """rb""" ) as fp: lowerCamelCase : List[str] = _EndRecData(__magic_name__ ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: lowerCamelCase : List[Any] = fp.read(__magic_name__ ) # CD is where we expect it to be if len(__magic_name__ ) == sizeCentralDir: lowerCamelCase : str = struct.unpack(__magic_name__ , __magic_name__ ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with zipfile.ZipFile(__magic_name__ , """r""" ) as zip_file: zip_file.extractall(__magic_name__ ) zip_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[str] = [B"""\xFD\x37\x7A\x58\x5A\x00"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with lzma.open(__magic_name__ ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""Rar!\x1a\x07\x00""", B"""Rar!\x1a\x07\x01\x00"""] # RAR_ID # RAR5_ID @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Union[str, Any] = rarfile.RarFile(__magic_name__ ) rf.extractall(__magic_name__ ) rf.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = [B"""\x28\xb5\x2F\xFD"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd lowerCamelCase : int = zstd.ZstdDecompressor() with open(__magic_name__ , """rb""" ) as ifh, open(__magic_name__ , """wb""" ) as ofh: dctx.copy_stream(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""\x42\x5A\x68"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with bza.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x37\x7A\xBC\xAF\x27\x1C"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with pyazr.SevenZipFile(__magic_name__ , """r""" ) as archive: archive.extractall(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x04\x22\x4D\x18"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) _UpperCAmelCase : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def UpperCamelCase__ ( cls ): return max( len(__magic_name__ ) for extractor in cls.extractors.values() if issubclass(__magic_name__ , __magic_name__ ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): try: return MagicNumberBaseExtractor.read_magic_number(__magic_name__ , magic_number_length=__magic_name__ ) except OSError: return b"" @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = False ): warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = cls.infer_extractor_format(__magic_name__ ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): # <Added version="2.4.0"/> lowerCamelCase : Dict = cls._get_magic_number_max_length() lowerCamelCase : Optional[Any] = cls._read_magic_number(__magic_name__ , __magic_name__ ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__magic_name__ , magic_number=__magic_name__ ): return extractor_format @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = "deprecated" , ): os.makedirs(os.path.dirname(__magic_name__ ) , exist_ok=__magic_name__ ) # Prevent parallel extractions lowerCamelCase : Tuple = str(Path(__magic_name__ ).with_suffix(""".lock""" ) ) with FileLock(__magic_name__ ): shutil.rmtree(__magic_name__ , ignore_errors=__magic_name__ ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__magic_name__ , __magic_name__ ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = extractor if extractor != """deprecated""" else extractor_format else: lowerCamelCase : Optional[int] = cls.extractors[extractor_format] return extractor.extract(__magic_name__ , __magic_name__ ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=__magic_name__ , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__magic_name__ ): return extractor.extract(__magic_name__ , __magic_name__ )
287
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) lowerCamelCase : Any = tf.convert_to_tensor( [[5, 1_2_1, 1_1, 6_6_0, 1_6, 7_3_0, 2_5_5_4_3, 1_1_0, 8_3, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowerCamelCase : str = model(__magic_name__ )["""last_hidden_state"""] lowerCamelCase : Union[str, Any] = tf.TensorShape((1, 1_0, 7_6_8) ) self.assertEqual(output.shape , __magic_name__ ) # compare the actual values for a slice. lowerCamelCase : Dict = tf.convert_to_tensor( [[[-0.0_254, 0.0_235, 0.1_027], [0.0_606, -0.1_811, -0.0_418], [-0.1_561, -0.1_127, 0.2_687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
287
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration _lowerCamelCase =5_0_0_0_0_0 _lowerCamelCase , _lowerCamelCase =os.path.split(__file__) _lowerCamelCase =os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.map(**lowerCamelCase ) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.filter(**lowerCamelCase ) def _a ( ): lowerCamelCase : Optional[Any] = {"""num examples""": SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : Any = datasets.Features({"""text""": datasets.Value("""string""" ), """numbers""": datasets.Value("""float32""" )} ) lowerCamelCase : Tuple = generate_example_dataset( os.path.join(lowerCamelCase, """dataset.arrow""" ), lowerCamelCase, num_examples=lowerCamelCase ) lowerCamelCase : Tuple = transformers.AutoTokenizer.from_pretrained("""bert-base-cased""", use_fast=lowerCamelCase ) def tokenize(lowerCamelCase ): return tokenizer(examples["""text"""] ) lowerCamelCase : List[str] = map(lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""numpy""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""pandas""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""torch""", columns="""numbers""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""tensorflow""", columns="""numbers""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : Union[str, Any] = filter(lowerCamelCase ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(lowerCamelCase, """wb""" ) as f: f.write(json.dumps(lowerCamelCase ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
287
1
from typing import TYPE_CHECKING from ....utils import _LazyModule _lowerCamelCase ={"""tokenization_tapex""": ["""TapexTokenizer"""]} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys _lowerCamelCase =_LazyModule(__name__, globals()["""__file__"""], _import_structure)
287
def _a ( lowerCamelCase ): if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True lowerCamelCase : Any = 4 lowerCamelCase : List[str] = (1 << p) - 1 for _ in range(p - 2 ): lowerCamelCase : List[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(1_1))
287
1
import warnings from .generation import TFGenerationMixin class A__ ( __SCREAMING_SNAKE_CASE): # warning at import time warnings.warn( """Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will """ """be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.""" , __SCREAMING_SNAKE_CASE , )
287
import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , *__magic_name__ , **__magic_name__ ): warnings.warn( """The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use OwlViTImageProcessor instead.""" , __magic_name__ , ) super().__init__(*__magic_name__ , **__magic_name__ )
287
1
import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase =[ ("""bert.bert""", """visual_bert"""), ("""bert.cls""", """cls"""), ("""bert.classifier""", """cls"""), ("""token_type_embeddings_visual""", """visual_token_type_embeddings"""), ("""position_embeddings_visual""", """visual_position_embeddings"""), ("""projection""", """visual_projection"""), ] _lowerCamelCase =[ """nlvr2_coco_pre_trained.th""", """nlvr2_fine_tuned.th""", """nlvr2_pre_trained.th""", """vcr_coco_pre_train.th""", """vcr_fine_tune.th""", """vcr_pre_train.th""", """vqa_coco_pre_trained.th""", """vqa_fine_tuned.th""", """vqa_pre_trained.th""", ] def _a ( lowerCamelCase ): lowerCamelCase : Any = torch.load(lowerCamelCase, map_location="""cpu""" ) return sd def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase=rename_keys_prefix ): lowerCamelCase : str = OrderedDict() lowerCamelCase : Tuple = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue lowerCamelCase : int = key for name_pair in rename_keys_prefix: lowerCamelCase : Dict = new_key.replace(name_pair[0], name_pair[1] ) lowerCamelCase : Dict = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately lowerCamelCase : Any = new_d["""cls.predictions.bias"""] return new_d @torch.no_grad() def _a ( lowerCamelCase, lowerCamelCase ): assert ( checkpoint_path.split("""/""" )[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: lowerCamelCase : Any = """pretraining""" if "vcr" in checkpoint_path: lowerCamelCase : Dict = {"""visual_embedding_dim""": 512} elif "vqa_advanced" in checkpoint_path: lowerCamelCase : Tuple = {"""visual_embedding_dim""": 2048} elif "vqa" in checkpoint_path: lowerCamelCase : List[str] = {"""visual_embedding_dim""": 2048} elif "nlvr" in checkpoint_path: lowerCamelCase : int = {"""visual_embedding_dim""": 1024} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''' ) else: if "vcr" in checkpoint_path: lowerCamelCase : List[str] = {"""visual_embedding_dim""": 512} lowerCamelCase : Union[str, Any] = """multichoice""" elif "vqa_advanced" in checkpoint_path: lowerCamelCase : List[Any] = {"""visual_embedding_dim""": 2048} lowerCamelCase : str = """vqa_advanced""" elif "vqa" in checkpoint_path: lowerCamelCase : str = {"""visual_embedding_dim""": 2048, """num_labels""": 3129} lowerCamelCase : Optional[int] = """vqa""" elif "nlvr" in checkpoint_path: lowerCamelCase : Tuple = { """visual_embedding_dim""": 1024, """num_labels""": 2, } lowerCamelCase : List[str] = """nlvr""" lowerCamelCase : Dict = VisualBertConfig(**lowerCamelCase ) # Load State Dict lowerCamelCase : List[Any] = load_state_dict(lowerCamelCase ) lowerCamelCase : List[str] = get_new_dict(lowerCamelCase, lowerCamelCase ) if model_type == "pretraining": lowerCamelCase : Optional[int] = VisualBertForPreTraining(lowerCamelCase ) elif model_type == "vqa": lowerCamelCase : Optional[int] = VisualBertForQuestionAnswering(lowerCamelCase ) elif model_type == "nlvr": lowerCamelCase : int = VisualBertForVisualReasoning(lowerCamelCase ) elif model_type == "multichoice": lowerCamelCase : Optional[int] = VisualBertForMultipleChoice(lowerCamelCase ) model.load_state_dict(lowerCamelCase ) # Save Checkpoints Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) model.save_pretrained(lowerCamelCase ) if __name__ == "__main__": _lowerCamelCase =argparse.ArgumentParser() # Required parameters parser.add_argument("""orig_checkpoint_path""", type=str, help="""A path to .th on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", type=str, help="""Path to the output PyTorch model.""") _lowerCamelCase =parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
287
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """nvidia/segformer-b0-finetuned-ade-512-512""": ( """https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json""" ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """segformer""" def __init__( self , __magic_name__=3 , __magic_name__=4 , __magic_name__=[2, 2, 2, 2] , __magic_name__=[8, 4, 2, 1] , __magic_name__=[3_2, 6_4, 1_6_0, 2_5_6] , __magic_name__=[7, 3, 3, 3] , __magic_name__=[4, 2, 2, 2] , __magic_name__=[1, 2, 5, 8] , __magic_name__=[4, 4, 4, 4] , __magic_name__="gelu" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=0.02 , __magic_name__=0.1 , __magic_name__=1e-6 , __magic_name__=2_5_6 , __magic_name__=2_5_5 , **__magic_name__ , ): super().__init__(**__magic_name__ ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( """Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be""" """ removed, as the behaviour will default to that of reshape_last_stage = True.""" , __magic_name__ , ) lowerCamelCase : Optional[Any] = num_channels lowerCamelCase : str = num_encoder_blocks lowerCamelCase : Any = depths lowerCamelCase : List[Any] = sr_ratios lowerCamelCase : int = hidden_sizes lowerCamelCase : Union[str, Any] = patch_sizes lowerCamelCase : Optional[Any] = strides lowerCamelCase : Dict = mlp_ratios lowerCamelCase : str = num_attention_heads lowerCamelCase : Any = hidden_act lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase : Dict = classifier_dropout_prob lowerCamelCase : Tuple = initializer_range lowerCamelCase : Dict = drop_path_rate lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : Any = decoder_hidden_size lowerCamelCase : str = kwargs.get("""reshape_last_stage""" , __magic_name__ ) lowerCamelCase : Dict = semantic_loss_ignore_index class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-4 @property def UpperCamelCase__ ( self ): return 1_2
287
1
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class A__ ( unittest.TestCase): @property def UpperCamelCase__ ( self ): torch.manual_seed(0 ) lowerCamelCase : int = UNetaDModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) return model def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = self.dummy_uncond_unet lowerCamelCase : int = KarrasVeScheduler() lowerCamelCase : str = KarrasVePipeline(unet=__magic_name__ , scheduler=__magic_name__ ) pipe.to(__magic_name__ ) pipe.set_progress_bar_config(disable=__magic_name__ ) lowerCamelCase : Tuple = torch.manual_seed(0 ) lowerCamelCase : List[str] = pipe(num_inference_steps=2 , generator=__magic_name__ , output_type="""numpy""" ).images lowerCamelCase : List[str] = torch.manual_seed(0 ) lowerCamelCase : str = pipe(num_inference_steps=2 , generator=__magic_name__ , output_type="""numpy""" , return_dict=__magic_name__ )[0] lowerCamelCase : Any = image[0, -3:, -3:, -1] lowerCamelCase : str = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) lowerCamelCase : int = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : Dict = """google/ncsnpp-celebahq-256""" lowerCamelCase : Optional[int] = UNetaDModel.from_pretrained(__magic_name__ ) lowerCamelCase : Union[str, Any] = KarrasVeScheduler() lowerCamelCase : Tuple = KarrasVePipeline(unet=__magic_name__ , scheduler=__magic_name__ ) pipe.to(__magic_name__ ) pipe.set_progress_bar_config(disable=__magic_name__ ) lowerCamelCase : int = torch.manual_seed(0 ) lowerCamelCase : Any = pipe(num_inference_steps=2_0 , generator=__magic_name__ , output_type="""numpy""" ).images lowerCamelCase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 2_5_6, 2_5_6, 3) lowerCamelCase : Union[str, Any] = np.array([0.578, 0.5_811, 0.5_924, 0.5_809, 0.587, 0.5_886, 0.5_861, 0.5_802, 0.586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
287
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = """gpt_neo""" _UpperCAmelCase : Union[str, Any] = ["""past_key_values"""] _UpperCAmelCase : List[Any] = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , __magic_name__=5_0_2_5_7 , __magic_name__=2_0_4_8 , __magic_name__=2_0_4_8 , __magic_name__=2_4 , __magic_name__=[[["global", "local"], 1_2]] , __magic_name__=1_6 , __magic_name__=None , __magic_name__=2_5_6 , __magic_name__="gelu_new" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=1e-5 , __magic_name__=0.02 , __magic_name__=True , __magic_name__=5_0_2_5_6 , __magic_name__=5_0_2_5_6 , **__magic_name__ , ): lowerCamelCase : List[Any] = vocab_size lowerCamelCase : str = max_position_embeddings lowerCamelCase : str = hidden_size lowerCamelCase : Optional[int] = num_layers lowerCamelCase : str = num_heads lowerCamelCase : Optional[Any] = intermediate_size lowerCamelCase : List[Any] = window_size lowerCamelCase : int = activation_function lowerCamelCase : Union[str, Any] = resid_dropout lowerCamelCase : List[Any] = embed_dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Dict = classifier_dropout lowerCamelCase : Any = layer_norm_epsilon lowerCamelCase : Dict = initializer_range lowerCamelCase : Dict = use_cache lowerCamelCase : Optional[Any] = bos_token_id lowerCamelCase : int = eos_token_id lowerCamelCase : List[Any] = attention_types lowerCamelCase : Optional[Any] = self.expand_attention_types_params(__magic_name__ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.attention_layers)` == `config.num_layers` """ F'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' """`config.attention_layers` is prepared using `config.attention_types`. """ """Please verify the value of `config.attention_types` argument.""" ) super().__init__(bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ ): lowerCamelCase : Optional[int] = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : Any = input.size() lowerCamelCase : List[Any] = len(lowerCamelCase ) lowerCamelCase : Optional[Any] = shape[dimension] lowerCamelCase : Optional[int] = torch.arange(0, lowerCamelCase, lowerCamelCase ) lowerCamelCase : Dict = torch.div(sizedim - size, lowerCamelCase, rounding_mode="""floor""" ) + 1 lowerCamelCase : int = torch.arange(lowerCamelCase ) + low_indices[:min_length][:, None] lowerCamelCase : str = [slice(lowerCamelCase )] * rank lowerCamelCase : List[str] = indices lowerCamelCase : Dict = input[s] lowerCamelCase : Any = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : List[Any] = torch.arange(1, lowerCamelCase ) lowerCamelCase : Optional[int] = torch.remainder(lowerCamelCase, lowerCamelCase ) lowerCamelCase : List[Any] = remainders == 0 lowerCamelCase : List[Any] = candidates[divisor_indices] lowerCamelCase : Optional[Any] = torch.max(lowerCamelCase ) return largest_divisor, torch.div(lowerCamelCase, lowerCamelCase, rounding_mode="""floor""" ) class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): lowerCamelCase : str = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(__magic_name__ , direction="""inputs""" ) lowerCamelCase : int = {0: """batch""", 1: """past_sequence + sequence"""} else: lowerCamelCase : Tuple = {0: """batch""", 1: """sequence"""} return common_inputs @property def UpperCamelCase__ ( self ): return self._config.num_heads def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ): lowerCamelCase : Optional[int] = super(__magic_name__ , self ).generate_dummy_inputs( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) # We need to order the input in the way they appears in the forward() lowerCamelCase : int = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch lowerCamelCase , lowerCamelCase : Optional[Any] = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values lowerCamelCase : Optional[int] = seqlen + 2 lowerCamelCase : List[Any] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowerCamelCase : str = [ (torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) for _ in range(self.num_layers ) ] lowerCamelCase : Tuple = common_inputs["""attention_mask"""] if self.use_past: lowerCamelCase : str = ordered_inputs["""attention_mask"""].dtype lowerCamelCase : Any = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(__magic_name__ , __magic_name__ , dtype=__magic_name__ )] , dim=1 ) return ordered_inputs @property def UpperCamelCase__ ( self ): return 1_3
287
1
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): for param, grad_param in zip(model_a.parameters(), model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=True ): model.train() lowerCamelCase : Dict = model(lowerCamelCase ) lowerCamelCase : Any = F.mse_loss(lowerCamelCase, target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase=False ): set_seed(42 ) lowerCamelCase : Tuple = RegressionModel() lowerCamelCase : Any = deepcopy(lowerCamelCase ) lowerCamelCase : Any = RegressionDataset(length=80 ) lowerCamelCase : Dict = DataLoader(lowerCamelCase, batch_size=16 ) model.to(accelerator.device ) if sched: lowerCamelCase : int = AdamW(params=model.parameters(), lr=1e-3 ) lowerCamelCase : Optional[Any] = AdamW(params=ddp_model.parameters(), lr=1e-3 ) lowerCamelCase : str = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) lowerCamelCase : Tuple = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) # Make a copy of `model` if sched: lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = accelerator.prepare(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: lowerCamelCase , lowerCamelCase : List[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def _a ( lowerCamelCase ): # Test when on a single CPU or GPU that the context manager does nothing lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : int = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad, ddp_param.grad ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : List[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase ): # Test on distributed setup that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : int = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Any = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : str = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Optional[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : Any = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : Any = get_training_setup(lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : str = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(lowerCamelCase ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Any = ddp_input[torch.randperm(len(lowerCamelCase ) )] GradientState._reset_state() def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : List[Any] = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase, lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(lowerCamelCase )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]["lr"]}\nDDP opt: {ddp_opt.param_groups[0]["lr"]}\n''' lowerCamelCase : Union[str, Any] = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(lowerCamelCase )) if accelerator.num_processes > 1: check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def _a ( ): lowerCamelCase : int = Accelerator() lowerCamelCase : Optional[Any] = RegressionDataset(length=80 ) lowerCamelCase : List[str] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase : int = RegressionDataset(length=96 ) lowerCamelCase : Optional[int] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if iteration < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if batch_num < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def _a ( ): lowerCamelCase : List[Any] = Accelerator() lowerCamelCase : int = accelerator.state if state.local_process_index == 0: print("""**Test `accumulate` gradient accumulation with dataloader break**""" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("""**Test NOOP `no_sync` context manager**""" ) test_noop_sync(lowerCamelCase ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("""**Test Distributed `no_sync` context manager**""" ) test_distributed_sync(lowerCamelCase ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation(lowerCamelCase, lowerCamelCase ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("""<""", """2.0""" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, """`split_batches=False`, `dispatch_batches=False`**""", ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation_with_opt_and_scheduler(lowerCamelCase, lowerCamelCase ) def _a ( lowerCamelCase ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
287
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
287
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _lowerCamelCase ={} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =["""GPTSw3Tokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys _lowerCamelCase =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
287
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): for param, grad_param in zip(model_a.parameters(), model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=True ): model.train() lowerCamelCase : Dict = model(lowerCamelCase ) lowerCamelCase : Any = F.mse_loss(lowerCamelCase, target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase=False ): set_seed(42 ) lowerCamelCase : Tuple = RegressionModel() lowerCamelCase : Any = deepcopy(lowerCamelCase ) lowerCamelCase : Any = RegressionDataset(length=80 ) lowerCamelCase : Dict = DataLoader(lowerCamelCase, batch_size=16 ) model.to(accelerator.device ) if sched: lowerCamelCase : int = AdamW(params=model.parameters(), lr=1e-3 ) lowerCamelCase : Optional[Any] = AdamW(params=ddp_model.parameters(), lr=1e-3 ) lowerCamelCase : str = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) lowerCamelCase : Tuple = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) # Make a copy of `model` if sched: lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = accelerator.prepare(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: lowerCamelCase , lowerCamelCase : List[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def _a ( lowerCamelCase ): # Test when on a single CPU or GPU that the context manager does nothing lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : int = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad, ddp_param.grad ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : List[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase ): # Test on distributed setup that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : int = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Any = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : str = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Optional[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : Any = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : Any = get_training_setup(lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : str = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(lowerCamelCase ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Any = ddp_input[torch.randperm(len(lowerCamelCase ) )] GradientState._reset_state() def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : List[Any] = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase, lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(lowerCamelCase )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]["lr"]}\nDDP opt: {ddp_opt.param_groups[0]["lr"]}\n''' lowerCamelCase : Union[str, Any] = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(lowerCamelCase )) if accelerator.num_processes > 1: check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def _a ( ): lowerCamelCase : int = Accelerator() lowerCamelCase : Optional[Any] = RegressionDataset(length=80 ) lowerCamelCase : List[str] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase : int = RegressionDataset(length=96 ) lowerCamelCase : Optional[int] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if iteration < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if batch_num < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def _a ( ): lowerCamelCase : List[Any] = Accelerator() lowerCamelCase : int = accelerator.state if state.local_process_index == 0: print("""**Test `accumulate` gradient accumulation with dataloader break**""" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("""**Test NOOP `no_sync` context manager**""" ) test_noop_sync(lowerCamelCase ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("""**Test Distributed `no_sync` context manager**""" ) test_distributed_sync(lowerCamelCase ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation(lowerCamelCase, lowerCamelCase ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("""<""", """2.0""" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, """`split_batches=False`, `dispatch_batches=False`**""", ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation_with_opt_and_scheduler(lowerCamelCase, lowerCamelCase ) def _a ( lowerCamelCase ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
287
1
from __future__ import annotations from collections.abc import Callable _lowerCamelCase =list[list[float | int]] def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : int = len(lowerCamelCase ) lowerCamelCase : Matrix = [[0 for _ in range(size + 1 )] for _ in range(lowerCamelCase )] lowerCamelCase : int lowerCamelCase : int lowerCamelCase : int lowerCamelCase : int lowerCamelCase : int lowerCamelCase : float for row in range(lowerCamelCase ): for col in range(lowerCamelCase ): lowerCamelCase : Union[str, Any] = matrix[row][col] lowerCamelCase : str = vector[row][0] lowerCamelCase : Union[str, Any] = 0 lowerCamelCase : Any = 0 while row < size and col < size: # pivoting lowerCamelCase : Optional[Any] = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowerCamelCase, lowerCamelCase ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: lowerCamelCase , lowerCamelCase : Tuple = augmented[pivot_row], augmented[row] for rowa in range(row + 1, lowerCamelCase ): lowerCamelCase : Union[str, Any] = augmented[rowa][col] / augmented[row][col] lowerCamelCase : str = 0 for cola in range(col + 1, size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1, lowerCamelCase ): for row in range(lowerCamelCase ): lowerCamelCase : List[str] = augmented[row][col] / augmented[col][col] for cola in range(lowerCamelCase, size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row], 10 )] for row in range(lowerCamelCase ) ] def _a ( lowerCamelCase ): lowerCamelCase : int = len(lowerCamelCase ) lowerCamelCase : Matrix = [[0 for _ in range(lowerCamelCase )] for _ in range(lowerCamelCase )] lowerCamelCase : Matrix = [[0] for _ in range(lowerCamelCase )] lowerCamelCase : Matrix lowerCamelCase : int lowerCamelCase : int lowerCamelCase : int for x_val, y_val in enumerate(lowerCamelCase ): for col in range(lowerCamelCase ): lowerCamelCase : Optional[Any] = (x_val + 1) ** (size - col - 1) lowerCamelCase : List[str] = y_val lowerCamelCase : int = solve(lowerCamelCase, lowerCamelCase ) def interpolated_func(lowerCamelCase ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowerCamelCase ) ) return interpolated_func def _a ( lowerCamelCase ): return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def _a ( lowerCamelCase = question_function, lowerCamelCase = 10 ): lowerCamelCase : list[int] = [func(lowerCamelCase ) for x_val in range(1, order + 1 )] lowerCamelCase : list[Callable[[int], int]] = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1, order + 1 ) ] lowerCamelCase : int = 0 lowerCamelCase : Callable[[int], int] lowerCamelCase : int for poly in polynomials: lowerCamelCase : Optional[Any] = 1 while func(lowerCamelCase ) == poly(lowerCamelCase ): x_val += 1 ret += poly(lowerCamelCase ) return ret if __name__ == "__main__": print(f'''{solution() = }''')
287
from scipy.stats import pearsonr import datasets _lowerCamelCase =""" Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ _lowerCamelCase =""" Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ _lowerCamelCase =""" @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class A__ ( datasets.Metric): def UpperCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"""] , ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__=False ): if return_pvalue: lowerCamelCase : Optional[Any] = pearsonr(__magic_name__ , __magic_name__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(__magic_name__ , __magic_name__ )[0] )}
287
1
import unittest from diffusers.models.unet_ad_blocks import * # noqa F403 from diffusers.utils import torch_device from .test_unet_blocks_common import UNetBlockTesterMixin class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Any = DownBlockaD # noqa F405 _UpperCAmelCase : List[str] = """down""" def UpperCamelCase__ ( self ): lowerCamelCase : Any = [-0.0_232, -0.9_869, 0.8_054, -0.0_637, -0.1_688, -1.4_264, 0.4_470, -1.3_394, 0.0_904] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Union[str, Any] = ResnetDownsampleBlockaD # noqa F405 _UpperCAmelCase : Union[str, Any] = """down""" def UpperCamelCase__ ( self ): lowerCamelCase : Dict = [0.0_710, 0.2_410, -0.7_320, -1.0_757, -1.1_343, 0.3_540, -0.0_133, -0.2_576, 0.0_948] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : List[str] = AttnDownBlockaD # noqa F405 _UpperCAmelCase : Union[str, Any] = """down""" def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = [0.0_636, 0.8_964, -0.6_234, -1.0_131, 0.0_844, 0.4_935, 0.3_437, 0.0_911, -0.2_957] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Optional[int] = CrossAttnDownBlockaD # noqa F405 _UpperCAmelCase : List[str] = """down""" def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Dict = super().prepare_init_args_and_inputs_for_common() lowerCamelCase : int = 3_2 return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = [0.2_238, -0.7_396, -0.2_255, -0.3_829, 0.1_925, 1.1_665, 0.0_603, -0.7_295, 0.1_983] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : List[Any] = SimpleCrossAttnDownBlockaD # noqa F405 _UpperCAmelCase : Any = """down""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_encoder_hidden_states=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : List[str] = super().prepare_init_args_and_inputs_for_common() lowerCamelCase : int = 3_2 return init_dict, inputs_dict @unittest.skipIf(torch_device == """mps""" , """MPS result is not consistent""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = [0.7_921, -0.0_992, -0.1_962, -0.7_695, -0.4_242, 0.7_804, 0.4_737, 0.2_765, 0.3_338] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : List[Any] = SkipDownBlockaD # noqa F405 _UpperCAmelCase : int = """down""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_skip_sample=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = [-0.0_845, -0.2_087, -0.2_465, 0.0_971, 0.1_900, -0.0_484, 0.2_664, 0.4_179, 0.5_069] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : List[str] = AttnSkipDownBlockaD # noqa F405 _UpperCAmelCase : int = """down""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_skip_sample=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = [0.5_539, 0.1_609, 0.4_924, 0.0_537, -0.1_995, 0.4_050, 0.0_979, -0.2_721, -0.0_642] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Optional[Any] = DownEncoderBlockaD # noqa F405 _UpperCAmelCase : Tuple = """down""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_temb=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = { """in_channels""": 3_2, """out_channels""": 3_2, } lowerCamelCase : str = self.dummy_input return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = [1.1_102, 0.5_302, 0.4_872, -0.0_023, -0.8_042, 0.0_483, -0.3_489, -0.5_632, 0.7_626] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Dict = AttnDownEncoderBlockaD # noqa F405 _UpperCAmelCase : Any = """down""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_temb=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = { """in_channels""": 3_2, """out_channels""": 3_2, } lowerCamelCase : Optional[int] = self.dummy_input return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = [0.8_966, -0.1_486, 0.8_568, 0.8_141, -0.9_046, -0.1_342, -0.0_972, -0.7_417, 0.1_538] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : str = UNetMidBlockaD # noqa F405 _UpperCAmelCase : Optional[Any] = """mid""" def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = { """in_channels""": 3_2, """temb_channels""": 1_2_8, } lowerCamelCase : int = self.dummy_input return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : Dict = [-0.1_062, 1.7_248, 0.3_494, 1.4_569, -0.0_910, -1.2_421, -0.9_984, 0.6_736, 1.0_028] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : List[Any] = UNetMidBlockaDCrossAttn # noqa F405 _UpperCAmelCase : List[Any] = """mid""" def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Tuple = super().prepare_init_args_and_inputs_for_common() lowerCamelCase : Tuple = 3_2 return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = [0.0_187, 2.4_220, 0.4_484, 1.1_203, -0.6_121, -1.5_122, -0.8_270, 0.7_851, 1.8_335] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Optional[Any] = UNetMidBlockaDSimpleCrossAttn # noqa F405 _UpperCAmelCase : List[Any] = """mid""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_encoder_hidden_states=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Union[str, Any] = super().prepare_init_args_and_inputs_for_common() lowerCamelCase : str = 3_2 return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = [0.7_143, 1.9_974, 0.5_448, 1.3_977, 0.1_282, -1.1_237, -1.4_238, 0.5_530, 0.8_880] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : int = UpBlockaD # noqa F405 _UpperCAmelCase : Any = """up""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = [-0.2_041, -0.4_165, -0.3_022, 0.0_041, -0.6_628, -0.7_053, 0.1_928, -0.0_325, 0.0_523] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Dict = ResnetUpsampleBlockaD # noqa F405 _UpperCAmelCase : Union[str, Any] = """up""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = [0.2_287, 0.3_549, -0.1_346, 0.4_797, -0.1_715, -0.9_649, 0.7_305, -0.5_864, -0.6_244] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Optional[int] = CrossAttnUpBlockaD # noqa F405 _UpperCAmelCase : int = """up""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Optional[int] = super().prepare_init_args_and_inputs_for_common() lowerCamelCase : Optional[Any] = 3_2 return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = [-0.1_403, -0.3_515, -0.0_420, -0.1_425, 0.3_167, 0.5_094, -0.2_181, 0.5_931, 0.5_582] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Union[str, Any] = SimpleCrossAttnUpBlockaD # noqa F405 _UpperCAmelCase : int = """up""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=__magic_name__ , include_encoder_hidden_states=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Dict = super().prepare_init_args_and_inputs_for_common() lowerCamelCase : str = 3_2 return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : Any = [0.2_645, 0.1_480, 0.0_909, 0.8_044, -0.9_758, -0.9_083, 0.0_994, -1.1_453, -0.7_402] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : List[Any] = AttnUpBlockaD # noqa F405 _UpperCAmelCase : List[Any] = """up""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=__magic_name__ ) @unittest.skipIf(torch_device == """mps""" , """MPS result is not consistent""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = [0.0_979, 0.1_326, 0.0_021, 0.0_659, 0.2_249, 0.0_059, 0.1_132, 0.5_952, 0.1_033] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Optional[int] = SkipUpBlockaD # noqa F405 _UpperCAmelCase : Optional[int] = """up""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = [-0.0_893, -0.1_234, -0.1_506, -0.0_332, 0.0_123, -0.0_211, 0.0_566, 0.0_143, 0.0_362] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Any = AttnSkipUpBlockaD # noqa F405 _UpperCAmelCase : Any = """up""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_res_hidden_states_tuple=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : int = [0.0_361, 0.0_617, 0.2_787, -0.0_350, 0.0_342, 0.3_421, -0.0_843, 0.0_913, 0.3_015] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : int = UpDecoderBlockaD # noqa F405 _UpperCAmelCase : str = """up""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_temb=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = {"""in_channels""": 3_2, """out_channels""": 3_2} lowerCamelCase : Any = self.dummy_input return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = [0.4_404, 0.1_998, -0.9_886, -0.3_320, -0.3_128, -0.7_034, -0.6_955, -0.2_338, -0.3_137] super().test_output(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : List[str] = AttnUpDecoderBlockaD # noqa F405 _UpperCAmelCase : List[str] = """up""" @property def UpperCamelCase__ ( self ): return super().get_dummy_input(include_temb=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = {"""in_channels""": 3_2, """out_channels""": 3_2} lowerCamelCase : int = self.dummy_input return init_dict, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = [0.6_738, 0.4_491, 0.1_055, 1.0_710, 0.7_316, 0.3_339, 0.3_352, 0.1_023, 0.3_568] super().test_output(__magic_name__ )
287
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """microsoft/conditional-detr-resnet-50""": ( """https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """conditional_detr""" _UpperCAmelCase : Optional[int] = ["""past_key_values"""] _UpperCAmelCase : Optional[Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , __magic_name__=True , __magic_name__=None , __magic_name__=3 , __magic_name__=3_0_0 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=True , __magic_name__="relu" , __magic_name__=2_5_6 , __magic_name__=0.1 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.02 , __magic_name__=1.0 , __magic_name__=False , __magic_name__="sine" , __magic_name__="resnet50" , __magic_name__=True , __magic_name__=False , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=1 , __magic_name__=1 , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=0.25 , **__magic_name__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) lowerCamelCase : Optional[int] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(__magic_name__ , __magic_name__ ): lowerCamelCase : List[Any] = backbone_config.get("""model_type""" ) lowerCamelCase : Dict = CONFIG_MAPPING[backbone_model_type] lowerCamelCase : str = config_class.from_dict(__magic_name__ ) lowerCamelCase : Dict = use_timm_backbone lowerCamelCase : str = backbone_config lowerCamelCase : Tuple = num_channels lowerCamelCase : Dict = num_queries lowerCamelCase : Any = d_model lowerCamelCase : Optional[Any] = encoder_ffn_dim lowerCamelCase : List[str] = encoder_layers lowerCamelCase : Union[str, Any] = encoder_attention_heads lowerCamelCase : Any = decoder_ffn_dim lowerCamelCase : Dict = decoder_layers lowerCamelCase : Union[str, Any] = decoder_attention_heads lowerCamelCase : Dict = dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Union[str, Any] = activation_dropout lowerCamelCase : Optional[int] = activation_function lowerCamelCase : int = init_std lowerCamelCase : str = init_xavier_std lowerCamelCase : Tuple = encoder_layerdrop lowerCamelCase : str = decoder_layerdrop lowerCamelCase : Tuple = encoder_layers lowerCamelCase : Optional[int] = auxiliary_loss lowerCamelCase : Optional[Any] = position_embedding_type lowerCamelCase : Optional[int] = backbone lowerCamelCase : Union[str, Any] = use_pretrained_backbone lowerCamelCase : str = dilation # Hungarian matcher lowerCamelCase : Optional[Any] = class_cost lowerCamelCase : Dict = bbox_cost lowerCamelCase : Tuple = giou_cost # Loss coefficients lowerCamelCase : Union[str, Any] = mask_loss_coefficient lowerCamelCase : Dict = dice_loss_coefficient lowerCamelCase : Optional[int] = cls_loss_coefficient lowerCamelCase : Optional[int] = bbox_loss_coefficient lowerCamelCase : Optional[int] = giou_loss_coefficient lowerCamelCase : Optional[int] = focal_alpha super().__init__(is_encoder_decoder=__magic_name__ , **__magic_name__ ) @property def UpperCamelCase__ ( self ): return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): return self.d_model def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: lowerCamelCase : Optional[int] = self.backbone_config.to_dict() lowerCamelCase : Optional[Any] = self.__class__.model_type return output class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-5 @property def UpperCamelCase__ ( self ): return 1_2
287
1
import os def _a ( ): with open(os.path.dirname(lowerCamelCase ) + """/p022_names.txt""" ) as file: lowerCamelCase : str = str(file.readlines()[0] ) lowerCamelCase : str = names.replace("""\"""", """""" ).split(""",""" ) names.sort() lowerCamelCase : int = 0 lowerCamelCase : Any = 0 for i, name in enumerate(lowerCamelCase ): for letter in name: name_score += ord(lowerCamelCase ) - 64 total_score += (i + 1) * name_score lowerCamelCase : Union[str, Any] = 0 return total_score if __name__ == "__main__": print(solution())
287
import json import sys def _a ( lowerCamelCase, lowerCamelCase ): with open(lowerCamelCase, encoding="""utf-8""" ) as f: lowerCamelCase : List[Any] = json.load(lowerCamelCase ) lowerCamelCase : Optional[Any] = ["""<details>""", """<summary>Show updated benchmarks!</summary>""", """ """] for benchmark_name in sorted(lowerCamelCase ): lowerCamelCase : List[Any] = results[benchmark_name] lowerCamelCase : Union[str, Any] = benchmark_name.split("""/""" )[-1] output_md.append(F'''### Benchmark: {benchmark_file_name}''' ) lowerCamelCase : Any = """| metric |""" lowerCamelCase : str = """|--------|""" lowerCamelCase : List[Any] = """| new / old (diff) |""" for metric_name in sorted(lowerCamelCase ): lowerCamelCase : List[Any] = benchmark_res[metric_name] lowerCamelCase : Tuple = metric_vals["""new"""] lowerCamelCase : int = metric_vals.get("""old""", lowerCamelCase ) lowerCamelCase : Dict = metric_vals.get("""diff""", lowerCamelCase ) lowerCamelCase : Dict = F''' {new_val:f}''' if isinstance(lowerCamelCase, (int, float) ) else """None""" if old_val is not None: val_str += F''' / {old_val:f}''' if isinstance(lowerCamelCase, (int, float) ) else "None" if dif_val is not None: val_str += F''' ({dif_val:f})''' if isinstance(lowerCamelCase, (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append("""</details>""" ) with open(lowerCamelCase, """w""", encoding="""utf-8""" ) as f: f.writelines("""\n""".join(lowerCamelCase ) ) if __name__ == "__main__": _lowerCamelCase =sys.argv[1] _lowerCamelCase =sys.argv[2] format_json_to_md(input_json_file, output_md_file)
287
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _lowerCamelCase =logging.getLogger(__name__) def _a ( lowerCamelCase, lowerCamelCase ): return (preds == labels).mean() @dataclass class A__ : _UpperCAmelCase : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""}) _UpperCAmelCase : Optional[str] = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""}) _UpperCAmelCase : Optional[str] = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""}) _UpperCAmelCase : Optional[str] = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class A__ : _UpperCAmelCase : str = field(metadata={"""help""": """The name of the task to train on: """ + """, """.join(processors.keys())}) _UpperCAmelCase : str = field(metadata={"""help""": """Should contain the data files for the task."""}) _UpperCAmelCase : int = field( default=128 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) _UpperCAmelCase : bool = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Overwrite the cached training and evaluation sets"""}) def _a ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCamelCase : str = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[Any] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' """ --overwrite_output_dir to overcome.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN, ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""", training_args.local_rank, training_args.device, training_args.n_gpu, bool(training_args.local_rank != -1 ), training_args.fpaa, ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""", lowerCamelCase ) # Set seed set_seed(training_args.seed ) try: lowerCamelCase : List[Any] = processors[data_args.task_name]() lowerCamelCase : List[Any] = processor.get_labels() lowerCamelCase : Any = len(lowerCamelCase ) except KeyError: raise ValueError("""Task not found: %s""" % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCamelCase : int = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path, num_labels=lowerCamelCase, finetuning_task=data_args.task_name, cache_dir=model_args.cache_dir, ) lowerCamelCase : int = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path, cache_dir=model_args.cache_dir, ) lowerCamelCase : Dict = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path, from_tf=bool(""".ckpt""" in model_args.model_name_or_path ), config=lowerCamelCase, cache_dir=model_args.cache_dir, ) # Get datasets lowerCamelCase : Tuple = ( MultipleChoiceDataset( data_dir=data_args.data_dir, tokenizer=lowerCamelCase, task=data_args.task_name, max_seq_length=data_args.max_seq_length, overwrite_cache=data_args.overwrite_cache, mode=Split.train, ) if training_args.do_train else None ) lowerCamelCase : int = ( MultipleChoiceDataset( data_dir=data_args.data_dir, tokenizer=lowerCamelCase, task=data_args.task_name, max_seq_length=data_args.max_seq_length, overwrite_cache=data_args.overwrite_cache, mode=Split.dev, ) if training_args.do_eval else None ) def compute_metrics(lowerCamelCase ) -> Dict: lowerCamelCase : Optional[Any] = np.argmax(p.predictions, axis=1 ) return {"acc": simple_accuracy(lowerCamelCase, p.label_ids )} # Data collator lowerCamelCase : Tuple = DataCollatorWithPadding(lowerCamelCase, pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer lowerCamelCase : str = Trainer( model=lowerCamelCase, args=lowerCamelCase, train_dataset=lowerCamelCase, eval_dataset=lowerCamelCase, compute_metrics=lowerCamelCase, data_collator=lowerCamelCase, ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowerCamelCase : List[Any] = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) lowerCamelCase : Tuple = trainer.evaluate() lowerCamelCase : Optional[Any] = os.path.join(training_args.output_dir, """eval_results.txt""" ) if trainer.is_world_master(): with open(lowerCamelCase, """w""" ) as writer: logger.info("""***** Eval results *****""" ) for key, value in result.items(): logger.info(""" %s = %s""", lowerCamelCase, lowerCamelCase ) writer.write("""%s = %s\n""" % (key, value) ) results.update(lowerCamelCase ) return results def _a ( lowerCamelCase ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
287
def _a ( lowerCamelCase ): return " ".join( """""".join(word[::-1] ) if len(lowerCamelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("""Hey wollef sroirraw"""))
287
1
import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): # Initialise PyTorch model lowerCamelCase : int = FunnelConfig.from_json_file(lowerCamelCase ) print(F'''Building PyTorch model from configuration: {config}''' ) lowerCamelCase : str = FunnelBaseModel(lowerCamelCase ) if base_model else FunnelModel(lowerCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_funnel(lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict(), lowerCamelCase ) if __name__ == "__main__": _lowerCamelCase =argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--base_model""", action="""store_true""", help="""Whether you want just the base model (no decoder) or not.""" ) _lowerCamelCase =parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
287
# Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from packaging import version from .. import __version__ from .constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD from .doc import ( add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, copy_func, replace_return_docstrings, ) from .generic import ( ContextManagers, ExplicitEnum, ModelOutput, PaddingStrategy, TensorType, add_model_info_to_auto_map, cached_property, can_return_loss, expand_dims, find_labels, flatten_dict, infer_framework, is_jax_tensor, is_numpy_array, is_tensor, is_tf_symbolic_tensor, is_tf_tensor, is_torch_device, is_torch_dtype, is_torch_tensor, reshape, squeeze, strtobool, tensor_size, to_numpy, to_py_obj, transpose, working_or_temp_dir, ) from .hub import ( CLOUDFRONT_DISTRIB_PREFIX, DISABLE_TELEMETRY, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, EntryNotFoundError, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, cached_file, default_cache_path, define_sagemaker_information, download_url, extract_commit_hash, get_cached_models, get_file_from_repo, get_full_repo_name, has_file, http_user_agent, is_offline_mode, is_remote_url, move_cache, send_example_telemetry, try_to_load_from_cache, ) from .import_utils import ( ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, TORCH_FX_REQUIRED_VERSION, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, ccl_version, direct_transformers_import, get_torch_version, is_accelerate_available, is_apex_available, is_bitsandbytes_available, is_bsa_available, is_coloredlogs_available, is_cython_available, is_datasets_available, is_decord_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_jieba_available, is_jumanpp_available, is_kenlm_available, is_keras_nlp_available, is_librosa_available, is_natten_available, is_ninja_available, is_onnx_available, is_openai_available, is_optimum_available, is_pandas_available, is_peft_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytest_available, is_pytorch_quantization_available, is_rjieba_available, is_sacremoses_available, is_safetensors_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_sudachi_available, is_tensorflow_probability_available, is_tensorflow_text_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_bfaa_cpu_available, is_torch_bfaa_gpu_available, is_torch_compile_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_neuroncore_available, is_torch_tensorrt_fx_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_torchdistx_available, is_torchdynamo_available, is_torchvision_available, is_training_run_on_sagemaker, is_vision_available, requires_backends, torch_only_method, ) _lowerCamelCase ="""pytorch_model.bin""" _lowerCamelCase ="""pytorch_model.bin.index.json""" _lowerCamelCase ="""adapter_config.json""" _lowerCamelCase ="""adapter_model.bin""" _lowerCamelCase ="""adapter_model.safetensors""" _lowerCamelCase ="""tf_model.h5""" _lowerCamelCase ="""tf_model.h5.index.json""" _lowerCamelCase ="""model.ckpt""" _lowerCamelCase ="""flax_model.msgpack""" _lowerCamelCase ="""flax_model.msgpack.index.json""" _lowerCamelCase ="""model.safetensors""" _lowerCamelCase ="""model.safetensors.index.json""" _lowerCamelCase ="""config.json""" _lowerCamelCase ="""preprocessor_config.json""" _lowerCamelCase =FEATURE_EXTRACTOR_NAME _lowerCamelCase ="""generation_config.json""" _lowerCamelCase ="""modelcard.json""" _lowerCamelCase ="""▁""" _lowerCamelCase =SENTENCEPIECE_UNDERLINE # Kept for backward compatibility _lowerCamelCase =[ [[0, 1, 0, 1], [1, 0, 0, 1]] ] * 2 # Needs to have 0s and 1s only since XLM uses it for langs too. _lowerCamelCase =[[7, 6, 0, 0, 1], [1, 2, 3, 0, 0], [0, 0, 0, 4, 5]] _lowerCamelCase =[[1, 1, 1, 1, 1], [1, 1, 1, 0, 0], [0, 0, 0, 1, 1]] def _a ( lowerCamelCase ): if version.parse(lowerCamelCase ) < version.parse(lowerCamelCase ): if "dev" in min_version: lowerCamelCase : Optional[int] = ( """This example requires a source install from HuggingFace Transformers (see """ """`https://huggingface.co/docs/transformers/installation#install-from-source`),""" ) else: lowerCamelCase : int = F'''This example requires a minimum version of {min_version},''' error_message += F''' but the version found is {__version__}.\n''' raise ImportError( error_message + """Check out https://github.com/huggingface/transformers/tree/main/examples#important-note for the examples corresponding to other """ """versions of HuggingFace Transformers.""" )
287
1
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """nvidia/segformer-b0-finetuned-ade-512-512""": ( """https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json""" ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """segformer""" def __init__( self , __magic_name__=3 , __magic_name__=4 , __magic_name__=[2, 2, 2, 2] , __magic_name__=[8, 4, 2, 1] , __magic_name__=[3_2, 6_4, 1_6_0, 2_5_6] , __magic_name__=[7, 3, 3, 3] , __magic_name__=[4, 2, 2, 2] , __magic_name__=[1, 2, 5, 8] , __magic_name__=[4, 4, 4, 4] , __magic_name__="gelu" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=0.02 , __magic_name__=0.1 , __magic_name__=1e-6 , __magic_name__=2_5_6 , __magic_name__=2_5_5 , **__magic_name__ , ): super().__init__(**__magic_name__ ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( """Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be""" """ removed, as the behaviour will default to that of reshape_last_stage = True.""" , __magic_name__ , ) lowerCamelCase : Optional[Any] = num_channels lowerCamelCase : str = num_encoder_blocks lowerCamelCase : Any = depths lowerCamelCase : List[Any] = sr_ratios lowerCamelCase : int = hidden_sizes lowerCamelCase : Union[str, Any] = patch_sizes lowerCamelCase : Optional[Any] = strides lowerCamelCase : Dict = mlp_ratios lowerCamelCase : str = num_attention_heads lowerCamelCase : Any = hidden_act lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase : Dict = classifier_dropout_prob lowerCamelCase : Tuple = initializer_range lowerCamelCase : Dict = drop_path_rate lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : Any = decoder_hidden_size lowerCamelCase : str = kwargs.get("""reshape_last_stage""" , __magic_name__ ) lowerCamelCase : Dict = semantic_loss_ignore_index class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-4 @property def UpperCamelCase__ ( self ): return 1_2
287
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = """camembert""" def __init__( self , __magic_name__=3_0_5_2_2 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , __magic_name__=None , **__magic_name__ , ): super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : int = num_hidden_layers lowerCamelCase : int = num_attention_heads lowerCamelCase : Optional[int] = hidden_act lowerCamelCase : List[Any] = intermediate_size lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Optional[int] = attention_probs_dropout_prob lowerCamelCase : Optional[int] = max_position_embeddings lowerCamelCase : str = type_vocab_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : int = layer_norm_eps lowerCamelCase : Any = position_embedding_type lowerCamelCase : Optional[int] = use_cache lowerCamelCase : Union[str, Any] = classifier_dropout class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): if self.task == "multiple-choice": lowerCamelCase : List[str] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase : List[str] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
287
1
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration _lowerCamelCase ={ """tiny.en""": """https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt""", """tiny""": """https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt""", """base.en""": """https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt""", """base""": """https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt""", """small.en""": """https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt""", """small""": """https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt""", """medium.en""": """https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt""", """medium""": """https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt""", """large""": """https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt""", """large-v2""": """https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt""", } def _a ( lowerCamelCase ): lowerCamelCase : str = ["""layers""", """blocks"""] for k in ignore_keys: state_dict.pop(lowerCamelCase, lowerCamelCase ) _lowerCamelCase ={ """blocks""": """layers""", """mlp.0""": """fc1""", """mlp.2""": """fc2""", """mlp_ln""": """final_layer_norm""", """.attn.query""": """.self_attn.q_proj""", """.attn.key""": """.self_attn.k_proj""", """.attn.value""": """.self_attn.v_proj""", """.attn_ln""": """.self_attn_layer_norm""", """.attn.out""": """.self_attn.out_proj""", """.cross_attn.query""": """.encoder_attn.q_proj""", """.cross_attn.key""": """.encoder_attn.k_proj""", """.cross_attn.value""": """.encoder_attn.v_proj""", """.cross_attn_ln""": """.encoder_attn_layer_norm""", """.cross_attn.out""": """.encoder_attn.out_proj""", """decoder.ln.""": """decoder.layer_norm.""", """encoder.ln.""": """encoder.layer_norm.""", """token_embedding""": """embed_tokens""", """encoder.positional_embedding""": """encoder.embed_positions.weight""", """decoder.positional_embedding""": """decoder.embed_positions.weight""", """ln_post""": """layer_norm""", } def _a ( lowerCamelCase ): lowerCamelCase : Union[str, Any] = list(s_dict.keys() ) for key in keys: lowerCamelCase : List[Any] = key for k, v in WHISPER_MAPPING.items(): if k in key: lowerCamelCase : List[Any] = new_key.replace(lowerCamelCase, lowerCamelCase ) print(F'''{key} -> {new_key}''' ) lowerCamelCase : Tuple = s_dict.pop(lowerCamelCase ) return s_dict def _a ( lowerCamelCase ): lowerCamelCase , lowerCamelCase : Optional[int] = emb.weight.shape lowerCamelCase : Optional[int] = nn.Linear(lowerCamelCase, lowerCamelCase, bias=lowerCamelCase ) lowerCamelCase : int = emb.weight.data return lin_layer def _a ( lowerCamelCase, lowerCamelCase ): os.makedirs(lowerCamelCase, exist_ok=lowerCamelCase ) lowerCamelCase : List[Any] = os.path.basename(lowerCamelCase ) lowerCamelCase : int = url.split("""/""" )[-2] lowerCamelCase : int = os.path.join(lowerCamelCase, lowerCamelCase ) if os.path.exists(lowerCamelCase ) and not os.path.isfile(lowerCamelCase ): raise RuntimeError(F'''{download_target} exists and is not a regular file''' ) if os.path.isfile(lowerCamelCase ): lowerCamelCase : Optional[int] = open(lowerCamelCase, """rb""" ).read() if hashlib.shaaaa(lowerCamelCase ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(F'''{download_target} exists, but the SHA256 checksum does not match; re-downloading the file''' ) with urllib.request.urlopen(lowerCamelCase ) as source, open(lowerCamelCase, """wb""" ) as output: with tqdm( total=int(source.info().get("""Content-Length""" ) ), ncols=80, unit="""iB""", unit_scale=lowerCamelCase, unit_divisor=1024 ) as loop: while True: lowerCamelCase : str = source.read(8192 ) if not buffer: break output.write(lowerCamelCase ) loop.update(len(lowerCamelCase ) ) lowerCamelCase : List[Any] = open(lowerCamelCase, """rb""" ).read() if hashlib.shaaaa(lowerCamelCase ).hexdigest() != expected_shaaaa: raise RuntimeError( """Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.""" ) return model_bytes def _a ( lowerCamelCase, lowerCamelCase ): if ".pt" not in checkpoint_path: lowerCamelCase : str = _download(_MODELS[checkpoint_path] ) else: lowerCamelCase : Dict = torch.load(lowerCamelCase, map_location="""cpu""" ) lowerCamelCase : int = original_checkpoint["""dims"""] lowerCamelCase : Union[str, Any] = original_checkpoint["""model_state_dict"""] lowerCamelCase : Union[str, Any] = state_dict["""decoder.token_embedding.weight"""] remove_ignore_keys_(lowerCamelCase ) rename_keys(lowerCamelCase ) lowerCamelCase : str = True lowerCamelCase : str = state_dict["""decoder.layers.0.fc1.weight"""].shape[0] lowerCamelCase : Optional[Any] = WhisperConfig( vocab_size=dimensions["""n_vocab"""], encoder_ffn_dim=lowerCamelCase, decoder_ffn_dim=lowerCamelCase, num_mel_bins=dimensions["""n_mels"""], d_model=dimensions["""n_audio_state"""], max_target_positions=dimensions["""n_text_ctx"""], encoder_layers=dimensions["""n_audio_layer"""], encoder_attention_heads=dimensions["""n_audio_head"""], decoder_layers=dimensions["""n_text_layer"""], decoder_attention_heads=dimensions["""n_text_state"""], max_source_positions=dimensions["""n_audio_ctx"""], ) lowerCamelCase : str = WhisperForConditionalGeneration(lowerCamelCase ) lowerCamelCase , lowerCamelCase : int = model.model.load_state_dict(lowerCamelCase, strict=lowerCamelCase ) if len(lowerCamelCase ) > 0 and not set(lowerCamelCase ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( """Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,""" F''' but all the following weights are missing {missing}''' ) if tie_embeds: lowerCamelCase : List[Any] = make_linear_from_emb(model.model.decoder.embed_tokens ) else: lowerCamelCase : Optional[int] = proj_out_weights model.save_pretrained(lowerCamelCase ) if __name__ == "__main__": _lowerCamelCase =argparse.ArgumentParser() # # Required parameters parser.add_argument("""--checkpoint_path""", type=str, help="""Patht to the downloaded checkpoints""") parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") _lowerCamelCase =parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
287
_lowerCamelCase ={ "joule": 1.0, "kilojoule": 1_0_0_0, "megajoule": 1_0_0_0_0_0_0, "gigajoule": 1_0_0_0_0_0_0_0_0_0, "wattsecond": 1.0, "watthour": 3_6_0_0, "kilowatthour": 3_6_0_0_0_0_0, "newtonmeter": 1.0, "calorie_nutr": 4_1_8_6.8, "kilocalorie_nutr": 4_1_8_6_8_0_0.0_0, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1_0_5_5.0_5_5_8_5, "footpound": 1.355818, } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowerCamelCase : Dict = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {", ".join(lowerCamelCase )}''' ) raise ValueError(lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
287
1
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _lowerCamelCase =logging.getLogger() @unittest.skip("""Temporarily disable the doc tests.""") @require_torch @require_tf @slow class A__ ( unittest.TestCase): def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = True , ): lowerCamelCase : Any = [file for file in os.listdir(__magic_name__ ) if os.path.isfile(os.path.join(__magic_name__ , __magic_name__ ) )] if identifier is not None: lowerCamelCase : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(__magic_name__ , __magic_name__ ): for n_ in n_identifier: lowerCamelCase : Any = [file for file in files if n_ not in file] else: lowerCamelCase : Union[str, Any] = [file for file in files if n_identifier not in file] lowerCamelCase : Any = ignore_files or [] ignore_files.append("""__init__.py""" ) lowerCamelCase : Dict = [file for file in files if file not in ignore_files] for file in files: # Open all files print("""Testing""" , __magic_name__ ) if only_modules: lowerCamelCase : List[Any] = file.split(""".""" )[0] try: lowerCamelCase : List[Any] = getattr(__magic_name__ , __magic_name__ ) lowerCamelCase : Union[str, Any] = doctest.DocTestSuite(__magic_name__ ) lowerCamelCase : Optional[Any] = unittest.TextTestRunner().run(__magic_name__ ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F'''{module_identifier} is not a module.''' ) else: lowerCamelCase : List[str] = doctest.testfile(str("""..""" / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = Path("""src/transformers""" ) lowerCamelCase : Dict = """modeling""" lowerCamelCase : Optional[Any] = [ """modeling_ctrl.py""", """modeling_tf_ctrl.py""", ] self.analyze_directory(__magic_name__ , identifier=__magic_name__ , ignore_files=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = Path("""src/transformers""" ) lowerCamelCase : Union[str, Any] = """tokenization""" self.analyze_directory(__magic_name__ , identifier=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = Path("""src/transformers""" ) lowerCamelCase : Dict = """configuration""" self.analyze_directory(__magic_name__ , identifier=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = Path("""src/transformers""" ) lowerCamelCase : int = ["""configuration""", """modeling""", """tokenization"""] self.analyze_directory(__magic_name__ , n_identifier=__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = Path("""docs/source""" ) lowerCamelCase : Tuple = ["""favicon.ico"""] self.analyze_directory(__magic_name__ , ignore_files=__magic_name__ , only_modules=__magic_name__ )
287
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=__SCREAMING_SNAKE_CASE) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = field(default="""automatic-speech-recognition""" , metadata={"""include_in_asdict_even_if_is_default""": True}) _UpperCAmelCase : ClassVar[Features] = Features({"""audio""": Audio()}) _UpperCAmelCase : ClassVar[Features] = Features({"""transcription""": Value("""string""")}) _UpperCAmelCase : str = "audio" _UpperCAmelCase : str = "transcription" def UpperCamelCase__ ( self , __magic_name__ ): if self.audio_column not in features: raise ValueError(F'''Column {self.audio_column} is not present in features.''' ) if not isinstance(features[self.audio_column] , __magic_name__ ): raise ValueError(F'''Column {self.audio_column} is not an Audio type.''' ) lowerCamelCase : Optional[Any] = copy.deepcopy(self ) lowerCamelCase : List[Any] = self.input_schema.copy() lowerCamelCase : Tuple = features[self.audio_column] lowerCamelCase : int = input_schema return task_template @property def UpperCamelCase__ ( self ): return {self.audio_column: "audio", self.transcription_column: "transcription"}
287
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def _a ( lowerCamelCase=None ): if subparsers is not None: lowerCamelCase : int = subparsers.add_parser("""test""" ) else: lowerCamelCase : int = argparse.ArgumentParser("""Accelerate test command""" ) parser.add_argument( """--config_file""", default=lowerCamelCase, help=( """The path to use to store the config file. Will default to a file named default_config.yaml in the cache """ """location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have """ """such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed """ """with 'huggingface'.""" ), ) if subparsers is not None: parser.set_defaults(func=lowerCamelCase ) return parser def _a ( lowerCamelCase ): lowerCamelCase : Optional[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["""test_utils""", """scripts""", """test_script.py"""] ) if args.config_file is None: lowerCamelCase : Optional[int] = script_name else: lowerCamelCase : int = F'''--config_file={args.config_file} {script_name}''' lowerCamelCase : Union[str, Any] = ["""accelerate-launch"""] + test_args.split() lowerCamelCase : Optional[int] = execute_subprocess_async(lowerCamelCase, env=os.environ.copy() ) if result.returncode == 0: print("""Test is a success! You are ready for your distributed training!""" ) def _a ( ): lowerCamelCase : str = test_command_parser() lowerCamelCase : Dict = parser.parse_args() test_command(lowerCamelCase ) if __name__ == "__main__": main()
287
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower_vision_model""" def __init__( self , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=3 , __magic_name__=1_6 , __magic_name__=2_8_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__=True , __magic_name__=False , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : str = num_hidden_layers lowerCamelCase : Optional[int] = num_channels lowerCamelCase : List[str] = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Any = initializer_factor lowerCamelCase : Tuple = layer_norm_eps lowerCamelCase : Tuple = stop_gradient lowerCamelCase : Optional[int] = share_layernorm lowerCamelCase : str = remove_last_layer @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : str = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """bridgetower_text_model""" def __init__( self , __magic_name__=5_0_2_6_5 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=1 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_4 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Any = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : Tuple = hidden_act lowerCamelCase : Optional[int] = initializer_factor lowerCamelCase : Any = intermediate_size lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : str = max_position_embeddings lowerCamelCase : Union[str, Any] = type_vocab_size lowerCamelCase : Optional[int] = layer_norm_eps lowerCamelCase : Optional[int] = position_embedding_type lowerCamelCase : List[str] = use_cache lowerCamelCase : List[str] = pad_token_id lowerCamelCase : List[str] = bos_token_id lowerCamelCase : Optional[int] = eos_token_id @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : Optional[int] = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower""" def __init__( self , __magic_name__=True , __magic_name__="gelu" , __magic_name__=7_6_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__="add" , __magic_name__=1_2 , __magic_name__=6 , __magic_name__=False , __magic_name__=False , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): # TODO: remove this once the Hub files are updated. lowerCamelCase : int = kwargs.pop("""text_config_dict""" , __magic_name__ ) lowerCamelCase : str = kwargs.pop("""vision_config_dict""" , __magic_name__ ) super().__init__(**__magic_name__ ) lowerCamelCase : str = share_cross_modal_transformer_layers lowerCamelCase : Union[str, Any] = hidden_act lowerCamelCase : str = hidden_size lowerCamelCase : Tuple = initializer_factor lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = share_link_tower_layers lowerCamelCase : List[Any] = link_tower_type lowerCamelCase : Tuple = num_attention_heads lowerCamelCase : int = num_hidden_layers lowerCamelCase : Union[str, Any] = tie_word_embeddings lowerCamelCase : Tuple = init_layernorm_from_vision_encoder if text_config is None: lowerCamelCase : Any = {} logger.info("""`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.""" ) if vision_config is None: lowerCamelCase : int = {} logger.info("""`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.""" ) lowerCamelCase : Any = BridgeTowerTextConfig(**__magic_name__ ) lowerCamelCase : Optional[Any] = BridgeTowerVisionConfig(**__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , **__magic_name__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = copy.deepcopy(self.__dict__ ) lowerCamelCase : int = self.text_config.to_dict() lowerCamelCase : Dict = self.vision_config.to_dict() lowerCamelCase : List[str] = self.__class__.model_type return output
287
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase =logging.get_logger(__name__) def _a ( lowerCamelCase ): lowerCamelCase : int = DPTConfig(embedding_type="""hybrid""" ) if "large" in checkpoint_url: lowerCamelCase : Any = 1024 lowerCamelCase : int = 4096 lowerCamelCase : List[str] = 24 lowerCamelCase : List[str] = 16 lowerCamelCase : Optional[Any] = [5, 11, 17, 23] lowerCamelCase : Union[str, Any] = [256, 512, 1024, 1024] lowerCamelCase : Any = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: lowerCamelCase : Optional[int] = 768 lowerCamelCase : List[str] = [1, 1, 1, 0.5] lowerCamelCase : List[Any] = [256, 512, 768, 768] lowerCamelCase : Optional[Any] = 150 lowerCamelCase : List[Any] = 16 lowerCamelCase : Tuple = (1, 384, 384) lowerCamelCase : Any = False lowerCamelCase : Tuple = """project""" if "ade" in checkpoint_url: lowerCamelCase : Optional[int] = True lowerCamelCase : Optional[Any] = 768 lowerCamelCase : Optional[Any] = [1, 1, 1, 0.5] lowerCamelCase : str = 150 lowerCamelCase : Optional[int] = 16 lowerCamelCase : Optional[int] = """huggingface/label-files""" lowerCamelCase : Tuple = """ade20k-id2label.json""" lowerCamelCase : Dict = json.load(open(cached_download(hf_hub_url(lowerCamelCase, lowerCamelCase, repo_type="""dataset""" ) ), """r""" ) ) lowerCamelCase : List[Any] = {int(lowerCamelCase ): v for k, v in idalabel.items()} lowerCamelCase : int = idalabel lowerCamelCase : Optional[Any] = {v: k for k, v in idalabel.items()} lowerCamelCase : Any = [1, 150, 480, 480] return config, expected_shape def _a ( lowerCamelCase ): lowerCamelCase : Dict = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(lowerCamelCase, lowerCamelCase ) def _a ( lowerCamelCase ): if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): lowerCamelCase : Union[str, Any] = name.replace("""pretrained.model""", """dpt.encoder""" ) if "pretrained.model" in name: lowerCamelCase : Union[str, Any] = name.replace("""pretrained.model""", """dpt.embeddings""" ) if "patch_embed" in name: lowerCamelCase : int = name.replace("""patch_embed""", """""" ) if "pos_embed" in name: lowerCamelCase : Tuple = name.replace("""pos_embed""", """position_embeddings""" ) if "attn.proj" in name: lowerCamelCase : List[str] = name.replace("""attn.proj""", """attention.output.dense""" ) if "proj" in name and "project" not in name: lowerCamelCase : Dict = name.replace("""proj""", """projection""" ) if "blocks" in name: lowerCamelCase : Any = name.replace("""blocks""", """layer""" ) if "mlp.fc1" in name: lowerCamelCase : Dict = name.replace("""mlp.fc1""", """intermediate.dense""" ) if "mlp.fc2" in name: lowerCamelCase : Optional[int] = name.replace("""mlp.fc2""", """output.dense""" ) if "norm1" in name and "backbone" not in name: lowerCamelCase : Optional[Any] = name.replace("""norm1""", """layernorm_before""" ) if "norm2" in name and "backbone" not in name: lowerCamelCase : Optional[Any] = name.replace("""norm2""", """layernorm_after""" ) if "scratch.output_conv" in name: lowerCamelCase : Any = name.replace("""scratch.output_conv""", """head""" ) if "scratch" in name: lowerCamelCase : Dict = name.replace("""scratch""", """neck""" ) if "layer1_rn" in name: lowerCamelCase : Dict = name.replace("""layer1_rn""", """convs.0""" ) if "layer2_rn" in name: lowerCamelCase : Any = name.replace("""layer2_rn""", """convs.1""" ) if "layer3_rn" in name: lowerCamelCase : Optional[Any] = name.replace("""layer3_rn""", """convs.2""" ) if "layer4_rn" in name: lowerCamelCase : Optional[int] = name.replace("""layer4_rn""", """convs.3""" ) if "refinenet" in name: lowerCamelCase : int = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 lowerCamelCase : int = name.replace(F'''refinenet{layer_idx}''', F'''fusion_stage.layers.{abs(layer_idx-4 )}''' ) if "out_conv" in name: lowerCamelCase : Optional[Any] = name.replace("""out_conv""", """projection""" ) if "resConfUnit1" in name: lowerCamelCase : Union[str, Any] = name.replace("""resConfUnit1""", """residual_layer1""" ) if "resConfUnit2" in name: lowerCamelCase : Dict = name.replace("""resConfUnit2""", """residual_layer2""" ) if "conv1" in name: lowerCamelCase : int = name.replace("""conv1""", """convolution1""" ) if "conv2" in name: lowerCamelCase : Dict = name.replace("""conv2""", """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: lowerCamelCase : Optional[Any] = name.replace("""pretrained.act_postprocess1.0.project.0""", """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: lowerCamelCase : Union[str, Any] = name.replace("""pretrained.act_postprocess2.0.project.0""", """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: lowerCamelCase : Union[str, Any] = name.replace("""pretrained.act_postprocess3.0.project.0""", """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: lowerCamelCase : Optional[Any] = name.replace("""pretrained.act_postprocess4.0.project.0""", """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: lowerCamelCase : Tuple = name.replace("""pretrained.act_postprocess1.3""", """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: lowerCamelCase : str = name.replace("""pretrained.act_postprocess1.4""", """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: lowerCamelCase : Dict = name.replace("""pretrained.act_postprocess2.3""", """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: lowerCamelCase : Union[str, Any] = name.replace("""pretrained.act_postprocess2.4""", """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: lowerCamelCase : List[Any] = name.replace("""pretrained.act_postprocess3.3""", """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: lowerCamelCase : Union[str, Any] = name.replace("""pretrained.act_postprocess4.3""", """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: lowerCamelCase : int = name.replace("""pretrained.act_postprocess4.4""", """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: lowerCamelCase : List[str] = name.replace("""pretrained""", """dpt""" ) if "bn" in name: lowerCamelCase : List[str] = name.replace("""bn""", """batch_norm""" ) if "head" in name: lowerCamelCase : Optional[int] = name.replace("""head""", """head.head""" ) if "encoder.norm" in name: lowerCamelCase : List[Any] = name.replace("""encoder.norm""", """layernorm""" ) if "auxlayer" in name: lowerCamelCase : Tuple = name.replace("""auxlayer""", """auxiliary_head.head""" ) if "backbone" in name: lowerCamelCase : Union[str, Any] = name.replace("""backbone""", """backbone.bit.encoder""" ) if ".." in name: lowerCamelCase : Optional[int] = name.replace("""..""", """.""" ) if "stem.conv" in name: lowerCamelCase : Union[str, Any] = name.replace("""stem.conv""", """bit.embedder.convolution""" ) if "blocks" in name: lowerCamelCase : Dict = name.replace("""blocks""", """layers""" ) if "convolution" in name and "backbone" in name: lowerCamelCase : Optional[Any] = name.replace("""convolution""", """conv""" ) if "layer" in name and "backbone" in name: lowerCamelCase : Union[str, Any] = name.replace("""layer""", """layers""" ) if "backbone.bit.encoder.bit" in name: lowerCamelCase : str = name.replace("""backbone.bit.encoder.bit""", """backbone.bit""" ) if "embedder.conv" in name: lowerCamelCase : int = name.replace("""embedder.conv""", """embedder.convolution""" ) if "backbone.bit.encoder.stem.norm" in name: lowerCamelCase : List[str] = name.replace("""backbone.bit.encoder.stem.norm""", """backbone.bit.embedder.norm""" ) return name def _a ( lowerCamelCase, lowerCamelCase ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase : List[Any] = state_dict.pop(F'''dpt.encoder.layer.{i}.attn.qkv.weight''' ) lowerCamelCase : str = state_dict.pop(F'''dpt.encoder.layer.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase : Dict = in_proj_weight[: config.hidden_size, :] lowerCamelCase : int = in_proj_bias[: config.hidden_size] lowerCamelCase : int = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase : Union[str, Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase : int = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase : Union[str, Any] = in_proj_bias[-config.hidden_size :] def _a ( ): lowerCamelCase : Optional[Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase : Optional[Any] = Image.open(requests.get(lowerCamelCase, stream=lowerCamelCase ).raw ) return im @torch.no_grad() def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowerCamelCase , lowerCamelCase : Dict = get_dpt_config(lowerCamelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") lowerCamelCase : Any = torch.load(lowerCamelCase, map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(lowerCamelCase ) # rename keys for key in state_dict.copy().keys(): lowerCamelCase : List[Any] = state_dict.pop(lowerCamelCase ) lowerCamelCase : str = val # read in qkv matrices read_in_q_k_v(lowerCamelCase, lowerCamelCase ) # load HuggingFace model lowerCamelCase : Optional[int] = DPTForSemanticSegmentation(lowerCamelCase ) if """ade""" in checkpoint_url else DPTForDepthEstimation(lowerCamelCase ) model.load_state_dict(lowerCamelCase ) model.eval() # Check outputs on an image lowerCamelCase : Union[str, Any] = 480 if """ade""" in checkpoint_url else 384 lowerCamelCase : List[str] = DPTImageProcessor(size=lowerCamelCase ) lowerCamelCase : Union[str, Any] = prepare_img() lowerCamelCase : Dict = image_processor(lowerCamelCase, return_tensors="""pt""" ) # forward pass lowerCamelCase : int = model(**lowerCamelCase ).logits if """ade""" in checkpoint_url else model(**lowerCamelCase ).predicted_depth if show_prediction: lowerCamelCase : Union[str, Any] = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ), size=(image.size[1], image.size[0]), mode="""bicubic""", align_corners=lowerCamelCase, ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCamelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCamelCase ) if push_to_hub: model.push_to_hub("""ybelkada/dpt-hybrid-midas""" ) image_processor.push_to_hub("""ybelkada/dpt-hybrid-midas""" ) if __name__ == "__main__": _lowerCamelCase =argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt""", type=str, help="""URL of the original DPT checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=False, help="""Path to the output PyTorch model directory.""", ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) parser.add_argument( """--model_name""", default="""dpt-large""", type=str, help="""Name of the model, in case you're pushing to the hub.""", ) parser.add_argument( """--show_prediction""", action="""store_true""", ) _lowerCamelCase =parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
287
def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = set(range(3, lowerCamelCase, 2 ) ) primes.add(2 ) for p in range(3, lowerCamelCase, 2 ): if p not in primes: continue primes.difference_update(set(range(p * p, lowerCamelCase, lowerCamelCase ) ) ) lowerCamelCase : Any = [float(lowerCamelCase ) for n in range(limit + 1 )] for p in primes: for n in range(lowerCamelCase, limit + 1, lowerCamelCase ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'''{solution() = }''')
287
1
_lowerCamelCase =[ """DownloadConfig""", """DownloadManager""", """DownloadMode""", """StreamingDownloadManager""", ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
287
from __future__ import annotations import collections import pprint from pathlib import Path def _a ( lowerCamelCase ): return "".join(sorted(lowerCamelCase ) ) def _a ( lowerCamelCase ): return word_by_signature[signature(lowerCamelCase )] _lowerCamelCase =Path(__file__).parent.joinpath("""words.txt""").read_text(encoding="""utf-8""") _lowerCamelCase =sorted({word.strip().lower() for word in data.splitlines()}) _lowerCamelCase =collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _lowerCamelCase ={word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open("""anagrams.txt""", """w""") as file: file.write("""all_anagrams = \n """) file.write(pprint.pformat(all_anagrams))
287
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from .config import config_command_parser from .config_args import default_config_file, load_config_from_file # noqa: F401 from .default import default_command_parser from .update import update_command_parser def _a ( lowerCamelCase=None ): lowerCamelCase : List[Any] = argparse.ArgumentParser(add_help=lowerCamelCase, allow_abbrev=lowerCamelCase ) # The main config parser lowerCamelCase : str = config_command_parser(lowerCamelCase ) # The subparser to add commands to lowerCamelCase : str = config_parser.add_subparsers(title="""subcommands""", dest="""subcommand""" ) # Then add other parsers with the parent parser default_command_parser(lowerCamelCase, parents=[parent_parser] ) update_command_parser(lowerCamelCase, parents=[parent_parser] ) return config_parser def _a ( ): lowerCamelCase : int = get_config_parser() lowerCamelCase : Any = config_parser.parse_args() if not hasattr(lowerCamelCase, """func""" ): config_parser.print_help() exit(1 ) # Run args.func(lowerCamelCase ) if __name__ == "__main__": main()
287
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = ["""pixel_values"""] def __init__( self , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , __magic_name__ = PILImageResampling.BILINEAR , __magic_name__ = True , __magic_name__ = 1 / 2_5_5 , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = size if size is not None else {"""shortest_edge""": 3_8_4} lowerCamelCase : Tuple = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Dict = do_resize lowerCamelCase : List[Any] = size # Default value set here for backwards compatibility where the value in config is None lowerCamelCase : Any = crop_pct if crop_pct is not None else 2_2_4 / 2_5_6 lowerCamelCase : Union[str, Any] = resample lowerCamelCase : str = do_rescale lowerCamelCase : Union[str, Any] = rescale_factor lowerCamelCase : Tuple = do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCamelCase : int = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = PILImageResampling.BICUBIC , __magic_name__ = None , **__magic_name__ , ): lowerCamelCase : Union[str, Any] = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) lowerCamelCase : str = size["""shortest_edge"""] if shortest_edge < 3_8_4: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct lowerCamelCase : List[str] = int(shortest_edge / crop_pct ) lowerCamelCase : Optional[Any] = get_resize_output_image_size(__magic_name__ , size=__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Optional[int] = resize(image=__magic_name__ , size=__magic_name__ , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=__magic_name__ , size=(shortest_edge, shortest_edge) , data_format=__magic_name__ , **__magic_name__ ) else: # warping (no cropping) when evaluated at 384 or larger return resize( __magic_name__ , size=(shortest_edge, shortest_edge) , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return rescale(__magic_name__ , scale=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return normalize(__magic_name__ , mean=__magic_name__ , std=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = ChannelDimension.FIRST , **__magic_name__ , ): lowerCamelCase : str = do_resize if do_resize is not None else self.do_resize lowerCamelCase : Optional[Any] = crop_pct if crop_pct is not None else self.crop_pct lowerCamelCase : Optional[int] = resample if resample is not None else self.resample lowerCamelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else self.image_mean lowerCamelCase : Tuple = image_std if image_std is not None else self.image_std lowerCamelCase : Dict = size if size is not None else self.size lowerCamelCase : str = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : List[str] = make_list_of_images(__magic_name__ ) if not valid_images(__magic_name__ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_resize and size["shortest_edge"] < 3_8_4 and crop_pct is None: raise ValueError("""crop_pct must be specified if size < 384.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. lowerCamelCase : Optional[Any] = [to_numpy_array(__magic_name__ ) for image in images] if do_resize: lowerCamelCase : List[Any] = [self.resize(image=__magic_name__ , size=__magic_name__ , crop_pct=__magic_name__ , resample=__magic_name__ ) for image in images] if do_rescale: lowerCamelCase : Union[str, Any] = [self.rescale(image=__magic_name__ , scale=__magic_name__ ) for image in images] if do_normalize: lowerCamelCase : List[Any] = [self.normalize(image=__magic_name__ , mean=__magic_name__ , std=__magic_name__ ) for image in images] lowerCamelCase : Optional[int] = [to_channel_dimension_format(__magic_name__ , __magic_name__ ) for image in images] lowerCamelCase : List[str] = {"""pixel_values""": images} return BatchFeature(data=__magic_name__ , tensor_type=__magic_name__ )
287
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """microsoft/trocr-base-handwritten""": ( """https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json""" ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[int] = """trocr""" _UpperCAmelCase : str = ["""past_key_values"""] _UpperCAmelCase : Union[str, Any] = { """num_attention_heads""": """decoder_attention_heads""", """hidden_size""": """d_model""", """num_hidden_layers""": """decoder_layers""", } def __init__( self , __magic_name__=5_0_2_6_5 , __magic_name__=1_0_2_4 , __magic_name__=1_2 , __magic_name__=1_6 , __magic_name__=4_0_9_6 , __magic_name__="gelu" , __magic_name__=5_1_2 , __magic_name__=0.1 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=0.0 , __magic_name__=True , __magic_name__=False , __magic_name__=True , __magic_name__=True , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , **__magic_name__ , ): lowerCamelCase : Union[str, Any] = vocab_size lowerCamelCase : int = d_model lowerCamelCase : Any = decoder_layers lowerCamelCase : Optional[Any] = decoder_attention_heads lowerCamelCase : Optional[Any] = decoder_ffn_dim lowerCamelCase : Tuple = activation_function lowerCamelCase : Any = max_position_embeddings lowerCamelCase : Tuple = dropout lowerCamelCase : str = attention_dropout lowerCamelCase : Tuple = activation_dropout lowerCamelCase : str = init_std lowerCamelCase : Dict = decoder_layerdrop lowerCamelCase : str = use_cache lowerCamelCase : Tuple = scale_embedding lowerCamelCase : int = use_learned_position_embeddings lowerCamelCase : Union[str, Any] = layernorm_embedding super().__init__( pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , decoder_start_token_id=__magic_name__ , **__magic_name__ , )
287
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 _lowerCamelCase ={ """return_dict""": False, """output_hidden_states""": True, """output_attentions""": True, """torchscript""": True, """torch_dtype""": """float16""", """use_bfloat16""": True, """tf_legacy_loss""": True, """pruned_heads""": {"""a""": 1}, """tie_word_embeddings""": False, """is_decoder""": True, """cross_attention_hidden_size""": 1_2_8, """add_cross_attention""": True, """tie_encoder_decoder""": True, """max_length""": 5_0, """min_length""": 3, """do_sample""": True, """early_stopping""": True, """num_beams""": 3, """num_beam_groups""": 3, """diversity_penalty""": 0.5, """temperature""": 2.0, """top_k""": 1_0, """top_p""": 0.7, """typical_p""": 0.2, """repetition_penalty""": 0.8, """length_penalty""": 0.8, """no_repeat_ngram_size""": 5, """encoder_no_repeat_ngram_size""": 5, """bad_words_ids""": [1, 2, 3], """num_return_sequences""": 3, """chunk_size_feed_forward""": 5, """output_scores""": True, """return_dict_in_generate""": True, """forced_bos_token_id""": 2, """forced_eos_token_id""": 3, """remove_invalid_values""": True, """architectures""": ["""BertModel"""], """finetuning_task""": """translation""", """id2label""": {0: """label"""}, """label2id""": {"""label""": """0"""}, """tokenizer_class""": """BertTokenizerFast""", """prefix""": """prefix""", """bos_token_id""": 6, """pad_token_id""": 7, """eos_token_id""": 8, """sep_token_id""": 9, """decoder_start_token_id""": 1_0, """exponential_decay_length_penalty""": (5, 1.01), """suppress_tokens""": [0, 1], """begin_suppress_tokens""": 2, """task_specific_params""": {"""translation""": """some_params"""}, """problem_type""": """regression""", } @is_staging_test class A__ ( unittest.TestCase): @classmethod def UpperCamelCase__ ( cls ): lowerCamelCase : int = TOKEN HfFolder.save_token(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls ): try: delete_repo(token=cls._token , repo_id="""test-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-config-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-config""" ) except HTTPError: pass def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""test-config""" , use_auth_token=self._token ) lowerCamelCase : Any = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__magic_name__ , repo_id="""test-config""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : Optional[Any] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""valid_org/test-config-org""" , use_auth_token=self._token ) lowerCamelCase : Optional[int] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( __magic_name__ , repo_id="""valid_org/test-config-org""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : List[str] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): CustomConfig.register_for_auto_class() lowerCamelCase : Optional[Any] = CustomConfig(attribute=4_2 ) config.push_to_hub("""test-dynamic-config""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"""AutoConfig""": """custom_configuration.CustomConfig"""} ) lowerCamelCase : List[str] = AutoConfig.from_pretrained(F'''{USER}/test-dynamic-config''' , trust_remote_code=__magic_name__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , """CustomConfig""" ) self.assertEqual(new_config.attribute , 4_2 ) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : str = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated lowerCamelCase : Optional[int] = c.n_embd + 1 # int lowerCamelCase : Optional[int] = c.resid_pdrop + 1.0 # float lowerCamelCase : Tuple = not c.scale_attn_weights # bool lowerCamelCase : Any = c.summary_type + """foo""" # str c.update_from_string( F'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(__magic_name__ , c.n_embd , """mismatch for key: n_embd""" ) self.assertEqual(__magic_name__ , c.resid_pdrop , """mismatch for key: resid_pdrop""" ) self.assertEqual(__magic_name__ , c.scale_attn_weights , """mismatch for key: scale_attn_weights""" ) self.assertEqual(__magic_name__ , c.summary_type , """mismatch for key: summary_type""" ) def UpperCamelCase__ ( self ): lowerCamelCase : str = PretrainedConfig() lowerCamelCase : int = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( __magic_name__ , ["""is_encoder_decoder""", """_name_or_path""", """_commit_hash""", """transformers_version"""] ) lowerCamelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(__magic_name__ , __magic_name__ )] if len(__magic_name__ ) > 0: raise ValueError( """The following keys are set with the default values in""" """ `test_configuration_common.config_common_kwargs` pick another value for them:""" F''' {", ".join(__magic_name__ )}.''' ) def UpperCamelCase__ ( self ): with self.assertRaises(__magic_name__ ): # config is in subfolder, the following should not work without specifying the subfolder lowerCamelCase : Dict = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" ) lowerCamelCase : str = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" , subfolder="""bert""" ) self.assertIsNotNone(__magic_name__ ) def UpperCamelCase__ ( self ): # A mock response for an HTTP head request to emulate server down lowerCamelCase : Dict = mock.Mock() lowerCamelCase : Optional[int] = 5_0_0 lowerCamelCase : List[Any] = {} lowerCamelCase : Tuple = HTTPError lowerCamelCase : Union[str, Any] = {} # Download this model to make sure it's in the cache. lowerCamelCase : List[str] = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=__magic_name__ ) as mock_head: lowerCamelCase : Any = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase__ ( self ): # This test is for deprecated behavior and can be removed in v5 lowerCamelCase : List[str] = BertConfig.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = AutoConfig.from_pretrained("""bert-base-cased""" ) lowerCamelCase : Optional[Any] = ["""config.4.0.0.json"""] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(__magic_name__ ) lowerCamelCase : str = 2 json.dump(configuration.to_dict() , open(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , """w""" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 lowerCamelCase : Any = ["""config.42.0.0.json"""] lowerCamelCase : Optional[Any] = 7_6_8 configuration.save_pretrained(__magic_name__ ) shutil.move(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , os.path.join(__magic_name__ , """config.42.0.0.json""" ) ) lowerCamelCase : int = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 7_6_8 ) def UpperCamelCase__ ( self ): # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. lowerCamelCase : str = """hf-internal-testing/test-two-configs""" import transformers as new_transformers lowerCamelCase : Tuple = """v4.0.0""" lowerCamelCase , lowerCamelCase : Optional[int] = new_transformers.models.auto.AutoConfig.from_pretrained( __magic_name__ , return_unused_kwargs=__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(__magic_name__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers lowerCamelCase : Tuple = """v3.0.0""" lowerCamelCase : Any = old_transformers.models.auto.AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(old_configuration.hidden_size , 7_6_8 )
287
1
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class A__ : def __init__( self , __magic_name__ , __magic_name__=1_3 , __magic_name__=3_0 , __magic_name__=2 , __magic_name__=3 , __magic_name__=True , __magic_name__=True , __magic_name__=3_2 , __magic_name__=2 , __magic_name__=4 , __magic_name__=3_7 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=1_0 , __magic_name__=0.02 , __magic_name__=3 , __magic_name__=None , __magic_name__=2 , ): lowerCamelCase : Any = parent lowerCamelCase : Union[str, Any] = batch_size lowerCamelCase : str = image_size lowerCamelCase : str = patch_size lowerCamelCase : Optional[int] = num_channels lowerCamelCase : Union[str, Any] = is_training lowerCamelCase : Optional[Any] = use_labels lowerCamelCase : Union[str, Any] = hidden_size lowerCamelCase : Union[str, Any] = num_hidden_layers lowerCamelCase : Optional[int] = num_attention_heads lowerCamelCase : List[str] = intermediate_size lowerCamelCase : Union[str, Any] = hidden_act lowerCamelCase : Optional[Any] = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : str = type_sequence_label_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : Dict = scope lowerCamelCase : Tuple = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) lowerCamelCase : Tuple = (image_size // patch_size) ** 2 lowerCamelCase : List[Any] = num_patches + 2 def UpperCamelCase__ ( self ): lowerCamelCase : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase : Optional[int] = None if self.use_labels: lowerCamelCase : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase : Tuple = self.get_config() return config, pixel_values, labels def UpperCamelCase__ ( self ): return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__magic_name__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Dict = TFDeiTModel(config=__magic_name__ ) lowerCamelCase : Any = model(__magic_name__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : List[Any] = TFDeiTForMaskedImageModeling(config=__magic_name__ ) lowerCamelCase : List[str] = model(__magic_name__ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCamelCase : int = 1 lowerCamelCase : Optional[int] = TFDeiTForMaskedImageModeling(__magic_name__ ) lowerCamelCase : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase : List[str] = model(__magic_name__ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Dict = self.type_sequence_label_size lowerCamelCase : Optional[Any] = TFDeiTForImageClassification(__magic_name__ ) lowerCamelCase : List[str] = model(__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase : str = 1 lowerCamelCase : List[Any] = TFDeiTForImageClassification(__magic_name__ ) lowerCamelCase : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase : Tuple = model(__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = self.prepare_config_and_inputs() lowerCamelCase , lowerCamelCase , lowerCamelCase : Dict = config_and_inputs lowerCamelCase : Any = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Union[str, Any] = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) _UpperCAmelCase : Union[str, Any] = ( { """feature-extraction""": TFDeiTModel, """image-classification""": (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) _UpperCAmelCase : str = False _UpperCAmelCase : Dict = False _UpperCAmelCase : str = False _UpperCAmelCase : Optional[Any] = False def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = TFDeiTModelTester(self ) lowerCamelCase : List[Any] = ConfigTester(self , config_class=__magic_name__ , has_text_modality=__magic_name__ , hidden_size=3_7 ) def UpperCamelCase__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def UpperCamelCase__ ( self ): pass def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase : List[str] = model_class(__magic_name__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__magic_name__ , tf.keras.layers.Dense ) ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase : Union[str, Any] = model_class(__magic_name__ ) lowerCamelCase : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase : List[Any] = [*signature.parameters.keys()] lowerCamelCase : Union[str, Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__=False ): lowerCamelCase : Union[str, Any] = super()._prepare_for_class(__magic_name__ , __magic_name__ , return_labels=__magic_name__ ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def UpperCamelCase__ ( self ): for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase : str = TFDeiTModel.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) def _a ( ): lowerCamelCase : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class A__ ( unittest.TestCase): @cached_property def UpperCamelCase__ ( self ): return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def UpperCamelCase__ ( self ): lowerCamelCase : str = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) lowerCamelCase : Dict = self.default_image_processor lowerCamelCase : List[str] = prepare_img() lowerCamelCase : str = image_processor(images=__magic_name__ , return_tensors="""tf""" ) # forward pass lowerCamelCase : Optional[Any] = model(**__magic_name__ ) # verify the logits lowerCamelCase : str = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __magic_name__ ) lowerCamelCase : Optional[Any] = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __magic_name__ , atol=1e-4 ) )
287
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) lowerCamelCase : Any = tf.convert_to_tensor( [[5, 1_2_1, 1_1, 6_6_0, 1_6, 7_3_0, 2_5_5_4_3, 1_1_0, 8_3, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowerCamelCase : str = model(__magic_name__ )["""last_hidden_state"""] lowerCamelCase : Union[str, Any] = tf.TensorShape((1, 1_0, 7_6_8) ) self.assertEqual(output.shape , __magic_name__ ) # compare the actual values for a slice. lowerCamelCase : Dict = tf.convert_to_tensor( [[[-0.0_254, 0.0_235, 0.1_027], [0.0_606, -0.1_811, -0.0_418], [-0.1_561, -0.1_127, 0.2_687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
287
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCamelCase ={ """configuration_longformer""": [ """LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LongformerConfig""", """LongformerOnnxConfig""", ], """tokenization_longformer""": ["""LongformerTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =["""LongformerTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =[ """LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """LongformerForMaskedLM""", """LongformerForMultipleChoice""", """LongformerForQuestionAnswering""", """LongformerForSequenceClassification""", """LongformerForTokenClassification""", """LongformerModel""", """LongformerPreTrainedModel""", """LongformerSelfAttention""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =[ """TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFLongformerForMaskedLM""", """TFLongformerForMultipleChoice""", """TFLongformerForQuestionAnswering""", """TFLongformerForSequenceClassification""", """TFLongformerForTokenClassification""", """TFLongformerModel""", """TFLongformerPreTrainedModel""", """TFLongformerSelfAttention""", ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys _lowerCamelCase =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
287
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _lowerCamelCase =get_logger(__name__) class A__ : def __init__( self , __magic_name__ = None ): lowerCamelCase : Dict = ( os.path.join(__magic_name__ , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) lowerCamelCase : List[str] = Extractor def UpperCamelCase__ ( self , __magic_name__ ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" lowerCamelCase : int = os.path.abspath(__magic_name__ ) return os.path.join(self.extract_dir , hash_url_to_filename(__magic_name__ ) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): return force_extract or ( not os.path.isfile(__magic_name__ ) and not (os.path.isdir(__magic_name__ ) and os.listdir(__magic_name__ )) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = False ): lowerCamelCase : Union[str, Any] = self.extractor.infer_extractor_format(__magic_name__ ) if not extractor_format: return input_path lowerCamelCase : int = self._get_output_path(__magic_name__ ) if self._do_extract(__magic_name__ , __magic_name__ ): self.extractor.extract(__magic_name__ , __magic_name__ , __magic_name__ ) return output_path class A__ ( __SCREAMING_SNAKE_CASE): @classmethod @abstractmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): ... @staticmethod @abstractmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): ... class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[bytes] = [] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with open(__magic_name__ , """rb""" ) as f: return f.read(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if not magic_number: lowerCamelCase : Optional[Any] = max(len(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) try: lowerCamelCase : Tuple = cls.read_magic_number(__magic_name__ , __magic_name__ ) except OSError: return False return any(magic_number.startswith(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) class A__ ( __SCREAMING_SNAKE_CASE): @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): return tarfile.is_tarfile(__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): def resolved(__magic_name__ ) -> str: return os.path.realpath(os.path.abspath(__magic_name__ ) ) def badpath(__magic_name__ , __magic_name__ ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__magic_name__ , __magic_name__ ) ).startswith(__magic_name__ ) def badlink(__magic_name__ , __magic_name__ ) -> bool: # Links are interpreted relative to the directory containing the link lowerCamelCase : List[str] = resolved(os.path.join(__magic_name__ , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=__magic_name__ ) lowerCamelCase : Optional[Any] = resolved(__magic_name__ ) for finfo in members: if badpath(finfo.name , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Dict = tarfile.open(__magic_name__ ) tar_file.extractall(__magic_name__ , members=TarExtractor.safemembers(__magic_name__ , __magic_name__ ) ) tar_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = [B"""\x1F\x8B"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with gzip.open(__magic_name__ , """rb""" ) as gzip_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[int] = [ B"""PK\x03\x04""", B"""PK\x05\x06""", # empty archive B"""PK\x07\x08""", # spanned archive ] @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if super().is_extractable(__magic_name__ , magic_number=__magic_name__ ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__magic_name__ , """rb""" ) as fp: lowerCamelCase : List[str] = _EndRecData(__magic_name__ ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: lowerCamelCase : List[Any] = fp.read(__magic_name__ ) # CD is where we expect it to be if len(__magic_name__ ) == sizeCentralDir: lowerCamelCase : str = struct.unpack(__magic_name__ , __magic_name__ ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with zipfile.ZipFile(__magic_name__ , """r""" ) as zip_file: zip_file.extractall(__magic_name__ ) zip_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[str] = [B"""\xFD\x37\x7A\x58\x5A\x00"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with lzma.open(__magic_name__ ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""Rar!\x1a\x07\x00""", B"""Rar!\x1a\x07\x01\x00"""] # RAR_ID # RAR5_ID @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Union[str, Any] = rarfile.RarFile(__magic_name__ ) rf.extractall(__magic_name__ ) rf.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = [B"""\x28\xb5\x2F\xFD"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd lowerCamelCase : int = zstd.ZstdDecompressor() with open(__magic_name__ , """rb""" ) as ifh, open(__magic_name__ , """wb""" ) as ofh: dctx.copy_stream(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""\x42\x5A\x68"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with bza.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x37\x7A\xBC\xAF\x27\x1C"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with pyazr.SevenZipFile(__magic_name__ , """r""" ) as archive: archive.extractall(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x04\x22\x4D\x18"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) _UpperCAmelCase : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def UpperCamelCase__ ( cls ): return max( len(__magic_name__ ) for extractor in cls.extractors.values() if issubclass(__magic_name__ , __magic_name__ ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): try: return MagicNumberBaseExtractor.read_magic_number(__magic_name__ , magic_number_length=__magic_name__ ) except OSError: return b"" @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = False ): warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = cls.infer_extractor_format(__magic_name__ ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): # <Added version="2.4.0"/> lowerCamelCase : Dict = cls._get_magic_number_max_length() lowerCamelCase : Optional[Any] = cls._read_magic_number(__magic_name__ , __magic_name__ ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__magic_name__ , magic_number=__magic_name__ ): return extractor_format @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = "deprecated" , ): os.makedirs(os.path.dirname(__magic_name__ ) , exist_ok=__magic_name__ ) # Prevent parallel extractions lowerCamelCase : Tuple = str(Path(__magic_name__ ).with_suffix(""".lock""" ) ) with FileLock(__magic_name__ ): shutil.rmtree(__magic_name__ , ignore_errors=__magic_name__ ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__magic_name__ , __magic_name__ ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = extractor if extractor != """deprecated""" else extractor_format else: lowerCamelCase : Optional[int] = cls.extractors[extractor_format] return extractor.extract(__magic_name__ , __magic_name__ ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=__magic_name__ , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__magic_name__ ): return extractor.extract(__magic_name__ , __magic_name__ )
287
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = """camembert""" def __init__( self , __magic_name__=3_0_5_2_2 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , __magic_name__=None , **__magic_name__ , ): super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : int = num_hidden_layers lowerCamelCase : int = num_attention_heads lowerCamelCase : Optional[int] = hidden_act lowerCamelCase : List[Any] = intermediate_size lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Optional[int] = attention_probs_dropout_prob lowerCamelCase : Optional[int] = max_position_embeddings lowerCamelCase : str = type_vocab_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : int = layer_norm_eps lowerCamelCase : Any = position_embedding_type lowerCamelCase : Optional[int] = use_cache lowerCamelCase : Union[str, Any] = classifier_dropout class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): if self.task == "multiple-choice": lowerCamelCase : List[str] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase : List[str] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
287
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration _lowerCamelCase =5_0_0_0_0_0 _lowerCamelCase , _lowerCamelCase =os.path.split(__file__) _lowerCamelCase =os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.map(**lowerCamelCase ) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.filter(**lowerCamelCase ) def _a ( ): lowerCamelCase : Optional[Any] = {"""num examples""": SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : Any = datasets.Features({"""text""": datasets.Value("""string""" ), """numbers""": datasets.Value("""float32""" )} ) lowerCamelCase : Tuple = generate_example_dataset( os.path.join(lowerCamelCase, """dataset.arrow""" ), lowerCamelCase, num_examples=lowerCamelCase ) lowerCamelCase : Tuple = transformers.AutoTokenizer.from_pretrained("""bert-base-cased""", use_fast=lowerCamelCase ) def tokenize(lowerCamelCase ): return tokenizer(examples["""text"""] ) lowerCamelCase : List[str] = map(lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""numpy""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""pandas""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""torch""", columns="""numbers""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""tensorflow""", columns="""numbers""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : Union[str, Any] = filter(lowerCamelCase ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(lowerCamelCase, """wb""" ) as f: f.write(json.dumps(lowerCamelCase ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
287
1
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class A__ ( unittest.TestCase): def __init__( self , __magic_name__ , __magic_name__=7 , __magic_name__=3 , __magic_name__=1_8 , __magic_name__=3_0 , __magic_name__=4_0_0 , __magic_name__=True , __magic_name__=None , __magic_name__=True , ): lowerCamelCase : List[str] = size if size is not None else {"""height""": 1_8, """width""": 1_8} lowerCamelCase : Any = parent lowerCamelCase : Union[str, Any] = batch_size lowerCamelCase : int = num_channels lowerCamelCase : Union[str, Any] = image_size lowerCamelCase : int = min_resolution lowerCamelCase : Union[str, Any] = max_resolution lowerCamelCase : Dict = do_resize lowerCamelCase : List[Any] = size lowerCamelCase : Any = do_normalize def UpperCamelCase__ ( self ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8_866_443_634_033_203, 0.6_618_829_369_544_983, 0.3_891_746_401_786_804], [-0.6_042_559_146_881_104, -0.02_295_008_860_528_469, 0.5_423_797_369_003_296], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Dict = ImageGPTImageProcessor if is_vision_available() else None def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = ImageGPTImageProcessingTester(self ) @property def UpperCamelCase__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ): lowerCamelCase : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__magic_name__ , """clusters""" ) ) self.assertTrue(hasattr(__magic_name__ , """do_resize""" ) ) self.assertTrue(hasattr(__magic_name__ , """size""" ) ) self.assertTrue(hasattr(__magic_name__ , """do_normalize""" ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 1_8, """width""": 1_8} ) lowerCamelCase : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 ) self.assertEqual(image_processor.size , {"""height""": 4_2, """width""": 4_2} ) def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) lowerCamelCase : List[Any] = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(__magic_name__ , obj[key] ) ) else: self.assertEqual(obj[key] , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : int = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase : Any = os.path.join(__magic_name__ , """image_processor.json""" ) image_processor_first.to_json_file(__magic_name__ ) lowerCamelCase : Tuple = self.image_processing_class.from_json_file(__magic_name__ ).to_dict() lowerCamelCase : Optional[int] = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(__magic_name__ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(__magic_name__ ) lowerCamelCase : int = self.image_processing_class.from_pretrained(__magic_name__ ).to_dict() lowerCamelCase : int = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(__magic_name__ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , __magic_name__ ) @unittest.skip("""ImageGPT requires clusters at initialization""" ) def UpperCamelCase__ ( self ): pass def _a ( ): lowerCamelCase : Union[str, Any] = load_dataset("""hf-internal-testing/fixtures_image_utils""", split="""test""" ) lowerCamelCase : Tuple = Image.open(dataset[4]["""file"""] ) lowerCamelCase : Optional[int] = Image.open(dataset[5]["""file"""] ) lowerCamelCase : str = [imagea, imagea] return images @require_vision @require_torch class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = ImageGPTImageProcessor.from_pretrained("""openai/imagegpt-small""" ) lowerCamelCase : Optional[int] = prepare_images() # test non-batched lowerCamelCase : Any = image_processing(images[0] , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1_0_2_4) ) lowerCamelCase : Any = [3_0_6, 1_9_1, 1_9_1] self.assertEqual(encoding.input_ids[0, :3].tolist() , __magic_name__ ) # test batched lowerCamelCase : List[str] = image_processing(__magic_name__ , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1_0_2_4) ) lowerCamelCase : int = [3_0_3, 1_3, 1_3] self.assertEqual(encoding.input_ids[1, -3:].tolist() , __magic_name__ )
287
def _a ( lowerCamelCase ): if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True lowerCamelCase : Any = 4 lowerCamelCase : List[str] = (1 << p) - 1 for _ in range(p - 2 ): lowerCamelCase : List[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(1_1))
287
1
from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class A__ : def __init__( self , __magic_name__ , __magic_name__=1_3 , __magic_name__=7 , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=9_9 , __magic_name__=3_2 , __magic_name__=2 , __magic_name__=4 , __magic_name__=3_7 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=1_6 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=3 , __magic_name__=4 , __magic_name__=None , __magic_name__=1_0_0_0 , ): lowerCamelCase : Union[str, Any] = parent lowerCamelCase : List[Any] = batch_size lowerCamelCase : List[Any] = seq_length lowerCamelCase : Union[str, Any] = is_training lowerCamelCase : Union[str, Any] = use_input_mask lowerCamelCase : str = use_token_type_ids lowerCamelCase : Optional[int] = use_labels lowerCamelCase : int = vocab_size lowerCamelCase : List[str] = hidden_size lowerCamelCase : Tuple = num_hidden_layers lowerCamelCase : str = num_attention_heads lowerCamelCase : str = intermediate_size lowerCamelCase : str = hidden_act lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : List[Any] = attention_probs_dropout_prob lowerCamelCase : List[Any] = max_position_embeddings lowerCamelCase : int = type_vocab_size lowerCamelCase : Dict = type_sequence_label_size lowerCamelCase : Optional[int] = initializer_range lowerCamelCase : str = num_labels lowerCamelCase : Dict = num_choices lowerCamelCase : List[str] = scope lowerCamelCase : Optional[int] = range_bbox def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment lowerCamelCase : int = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowerCamelCase : Dict = bbox[i, j, 3] lowerCamelCase : Any = bbox[i, j, 1] lowerCamelCase : str = t if bbox[i, j, 2] < bbox[i, j, 0]: lowerCamelCase : Union[str, Any] = bbox[i, j, 2] lowerCamelCase : Tuple = bbox[i, j, 0] lowerCamelCase : Optional[Any] = t lowerCamelCase : str = tf.convert_to_tensor(__magic_name__ ) lowerCamelCase : int = None if self.use_input_mask: lowerCamelCase : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase : Optional[int] = None if self.use_token_type_ids: lowerCamelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase : Optional[int] = None lowerCamelCase : Optional[int] = None lowerCamelCase : int = None if self.use_labels: lowerCamelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase : Tuple = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase : str = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : int = TFLayoutLMModel(config=__magic_name__ ) lowerCamelCase : Optional[int] = model(__magic_name__ , __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ ) lowerCamelCase : Tuple = model(__magic_name__ , __magic_name__ , token_type_ids=__magic_name__ ) lowerCamelCase : Any = model(__magic_name__ , __magic_name__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Dict = TFLayoutLMForMaskedLM(config=__magic_name__ ) lowerCamelCase : Tuple = model(__magic_name__ , __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Tuple = self.num_labels lowerCamelCase : List[str] = TFLayoutLMForSequenceClassification(config=__magic_name__ ) lowerCamelCase : Dict = model(__magic_name__ , __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Any = self.num_labels lowerCamelCase : Dict = TFLayoutLMForTokenClassification(config=__magic_name__ ) lowerCamelCase : Optional[Any] = model(__magic_name__ , __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : List[str] = TFLayoutLMForQuestionAnswering(config=__magic_name__ ) lowerCamelCase : Union[str, Any] = model(__magic_name__ , __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = self.prepare_config_and_inputs() ( ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ) : Any = config_and_inputs lowerCamelCase : Any = { """input_ids""": input_ids, """bbox""": bbox, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_tf class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : List[str] = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) _UpperCAmelCase : str = ( { """feature-extraction""": TFLayoutLMModel, """fill-mask""": TFLayoutLMForMaskedLM, """text-classification""": TFLayoutLMForSequenceClassification, """token-classification""": TFLayoutLMForTokenClassification, """zero-shot""": TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) _UpperCAmelCase : Dict = False _UpperCAmelCase : Tuple = True _UpperCAmelCase : Union[str, Any] = 10 def UpperCamelCase__ ( self ): lowerCamelCase : Any = TFLayoutLMModelTester(self ) lowerCamelCase : str = ConfigTester(self , config_class=__magic_name__ , hidden_size=3_7 ) def UpperCamelCase__ ( self ): self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__magic_name__ ) @slow def UpperCamelCase__ ( self ): for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase : str = TFLayoutLMModel.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) @unittest.skip("""Onnx compliancy broke with TF 2.10""" ) def UpperCamelCase__ ( self ): pass def _a ( ): # Here we prepare a batch of 2 sequences to test a LayoutLM forward pass on: # fmt: off lowerCamelCase : int = tf.convert_to_tensor([[101,1019,1014,1016,1037,1_2849,4747,1004,1_4246,2278,5439,4524,5002,2930,2193,2930,4341,3208,1005,1055,2171,2848,1_1300,3531,102],[101,4070,4034,7020,1024,3058,1015,1013,2861,1013,6070,1_9274,2772,6205,2_7814,1_6147,1_6147,4343,2047,1_0283,1_0969,1_4389,1012,2338,102]] ) # noqa: E231 lowerCamelCase : Union[str, Any] = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 lowerCamelCase : Optional[int] = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1000,1000,1000,1000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1000,1000,1000,1000]]] ) # noqa: E231 lowerCamelCase : Optional[int] = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) lowerCamelCase : str = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : Dict = TFLayoutLMModel.from_pretrained("""microsoft/layoutlm-base-uncased""" ) lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Union[str, Any] = prepare_layoutlm_batch_inputs() # forward pass lowerCamelCase : List[str] = model(input_ids=__magic_name__ , bbox=__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ ) # test the sequence output on [0, :3, :3] lowerCamelCase : List[str] = tf.convert_to_tensor( [[0.1_785, -0.1_947, -0.0_425], [-0.3_254, -0.2_807, 0.2_553], [-0.5_391, -0.3_322, 0.3_364]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , __magic_name__ , atol=1e-3 ) ) # test the pooled output on [1, :3] lowerCamelCase : List[Any] = tf.convert_to_tensor([-0.6_580, -0.0_214, 0.8_552] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , __magic_name__ , atol=1e-3 ) ) @slow def UpperCamelCase__ ( self ): # initialize model with randomly initialized sequence classification head lowerCamelCase : Union[str, Any] = TFLayoutLMForSequenceClassification.from_pretrained("""microsoft/layoutlm-base-uncased""" , num_labels=2 ) lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[str] = prepare_layoutlm_batch_inputs() # forward pass lowerCamelCase : Union[str, Any] = model( input_ids=__magic_name__ , bbox=__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar lowerCamelCase : List[str] = outputs.loss lowerCamelCase : Dict = (2,) self.assertEqual(loss.shape , __magic_name__ ) # test the shape of the logits lowerCamelCase : int = outputs.logits lowerCamelCase : str = (2, 2) self.assertEqual(logits.shape , __magic_name__ ) @slow def UpperCamelCase__ ( self ): # initialize model with randomly initialized token classification head lowerCamelCase : int = TFLayoutLMForTokenClassification.from_pretrained("""microsoft/layoutlm-base-uncased""" , num_labels=1_3 ) lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Tuple = prepare_layoutlm_batch_inputs() # forward pass lowerCamelCase : int = model( input_ids=__magic_name__ , bbox=__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ ) # test the shape of the logits lowerCamelCase : Optional[int] = outputs.logits lowerCamelCase : Optional[Any] = tf.convert_to_tensor((2, 2_5, 1_3) ) self.assertEqual(logits.shape , __magic_name__ ) @slow def UpperCamelCase__ ( self ): # initialize model with randomly initialized token classification head lowerCamelCase : Optional[Any] = TFLayoutLMForQuestionAnswering.from_pretrained("""microsoft/layoutlm-base-uncased""" ) lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = prepare_layoutlm_batch_inputs() # forward pass lowerCamelCase : Dict = model(input_ids=__magic_name__ , bbox=__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ ) # test the shape of the logits lowerCamelCase : Dict = tf.convert_to_tensor((2, 2_5) ) self.assertEqual(outputs.start_logits.shape , __magic_name__ ) self.assertEqual(outputs.end_logits.shape , __magic_name__ )
287
import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , *__magic_name__ , **__magic_name__ ): warnings.warn( """The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use OwlViTImageProcessor instead.""" , __magic_name__ , ) super().__init__(*__magic_name__ , **__magic_name__ )
287
1
from __future__ import annotations def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if len(lowerCamelCase ) == 0: raise ValueError("""find_max() arg is an empty sequence""" ) if ( left >= len(lowerCamelCase ) or left < -len(lowerCamelCase ) or right >= len(lowerCamelCase ) or right < -len(lowerCamelCase ) ): raise IndexError("""list index out of range""" ) if left == right: return nums[left] lowerCamelCase : List[Any] = (left + right) >> 1 # the middle lowerCamelCase : int = find_max(lowerCamelCase, lowerCamelCase, lowerCamelCase ) # find max in range[left, mid] lowerCamelCase : int = find_max(lowerCamelCase, mid + 1, lowerCamelCase ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
287
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """nvidia/segformer-b0-finetuned-ade-512-512""": ( """https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json""" ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """segformer""" def __init__( self , __magic_name__=3 , __magic_name__=4 , __magic_name__=[2, 2, 2, 2] , __magic_name__=[8, 4, 2, 1] , __magic_name__=[3_2, 6_4, 1_6_0, 2_5_6] , __magic_name__=[7, 3, 3, 3] , __magic_name__=[4, 2, 2, 2] , __magic_name__=[1, 2, 5, 8] , __magic_name__=[4, 4, 4, 4] , __magic_name__="gelu" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=0.02 , __magic_name__=0.1 , __magic_name__=1e-6 , __magic_name__=2_5_6 , __magic_name__=2_5_5 , **__magic_name__ , ): super().__init__(**__magic_name__ ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( """Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be""" """ removed, as the behaviour will default to that of reshape_last_stage = True.""" , __magic_name__ , ) lowerCamelCase : Optional[Any] = num_channels lowerCamelCase : str = num_encoder_blocks lowerCamelCase : Any = depths lowerCamelCase : List[Any] = sr_ratios lowerCamelCase : int = hidden_sizes lowerCamelCase : Union[str, Any] = patch_sizes lowerCamelCase : Optional[Any] = strides lowerCamelCase : Dict = mlp_ratios lowerCamelCase : str = num_attention_heads lowerCamelCase : Any = hidden_act lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase : Dict = classifier_dropout_prob lowerCamelCase : Tuple = initializer_range lowerCamelCase : Dict = drop_path_rate lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : Any = decoder_hidden_size lowerCamelCase : str = kwargs.get("""reshape_last_stage""" , __magic_name__ ) lowerCamelCase : Dict = semantic_loss_ignore_index class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-4 @property def UpperCamelCase__ ( self ): return 1_2
287
1
from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class A__ : def __init__( self , __magic_name__ = None ): if components is None: lowerCamelCase : Optional[Any] = [] lowerCamelCase : List[str] = list(__magic_name__ ) def __len__( self ): return len(self.__components ) def __str__( self ): return "(" + ",".join(map(__magic_name__ , self.__components ) ) + ")" def __add__( self , __magic_name__ ): lowerCamelCase : Any = len(self ) if size == len(__magic_name__ ): lowerCamelCase : Any = [self.__components[i] + other.component(__magic_name__ ) for i in range(__magic_name__ )] return Vector(__magic_name__ ) else: raise Exception("""must have the same size""" ) def __sub__( self , __magic_name__ ): lowerCamelCase : List[str] = len(self ) if size == len(__magic_name__ ): lowerCamelCase : str = [self.__components[i] - other.component(__magic_name__ ) for i in range(__magic_name__ )] return Vector(__magic_name__ ) else: # error case raise Exception("""must have the same size""" ) @overload def __mul__( self , __magic_name__ ): ... @overload def __mul__( self , __magic_name__ ): ... def __mul__( self , __magic_name__ ): if isinstance(__magic_name__ , (float, int) ): lowerCamelCase : Any = [c * other for c in self.__components] return Vector(__magic_name__ ) elif isinstance(__magic_name__ , __magic_name__ ) and len(self ) == len(__magic_name__ ): lowerCamelCase : List[Any] = len(self ) lowerCamelCase : Optional[int] = [self.__components[i] * other.component(__magic_name__ ) for i in range(__magic_name__ )] return sum(__magic_name__ ) else: # error case raise Exception("""invalid operand!""" ) def UpperCamelCase__ ( self ): return Vector(self.__components ) def UpperCamelCase__ ( self , __magic_name__ ): if isinstance(__magic_name__ , __magic_name__ ) and -len(self.__components ) <= i < len(self.__components ): return self.__components[i] else: raise Exception("""index out of range""" ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): assert -len(self.__components ) <= pos < len(self.__components ) lowerCamelCase : Union[str, Any] = value def UpperCamelCase__ ( self ): if len(self.__components ) == 0: raise Exception("""Vector is empty""" ) lowerCamelCase : Union[str, Any] = [c**2 for c in self.__components] return math.sqrt(sum(__magic_name__ ) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = False ): lowerCamelCase : Tuple = self * other lowerCamelCase : List[Any] = self.euclidean_length() * other.euclidean_length() if deg: return math.degrees(math.acos(num / den ) ) else: return math.acos(num / den ) def _a ( lowerCamelCase ): assert isinstance(lowerCamelCase, lowerCamelCase ) return Vector([0] * dimension ) def _a ( lowerCamelCase, lowerCamelCase ): assert isinstance(lowerCamelCase, lowerCamelCase ) and (isinstance(lowerCamelCase, lowerCamelCase )) lowerCamelCase : Tuple = [0] * dimension lowerCamelCase : Tuple = 1 return Vector(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): assert ( isinstance(lowerCamelCase, lowerCamelCase ) and isinstance(lowerCamelCase, lowerCamelCase ) and (isinstance(lowerCamelCase, (int, float) )) ) return x * scalar + y def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): random.seed(lowerCamelCase ) lowerCamelCase : str = [random.randint(lowerCamelCase, lowerCamelCase ) for _ in range(lowerCamelCase )] return Vector(lowerCamelCase ) class A__ : def __init__( self , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Union[str, Any] = matrix lowerCamelCase : Optional[Any] = w lowerCamelCase : List[Any] = h def __str__( self ): lowerCamelCase : Dict = """""" for i in range(self.__height ): ans += "|" for j in range(self.__width ): if j < self.__width - 1: ans += str(self.__matrix[i][j] ) + "," else: ans += str(self.__matrix[i][j] ) + "|\n" return ans def __add__( self , __magic_name__ ): if self.__width == other.width() and self.__height == other.height(): lowerCamelCase : Dict = [] for i in range(self.__height ): lowerCamelCase : List[Any] = [ self.__matrix[i][j] + other.component(__magic_name__ , __magic_name__ ) for j in range(self.__width ) ] matrix.append(__magic_name__ ) return Matrix(__magic_name__ , self.__width , self.__height ) else: raise Exception("""matrix must have the same dimension!""" ) def __sub__( self , __magic_name__ ): if self.__width == other.width() and self.__height == other.height(): lowerCamelCase : Optional[int] = [] for i in range(self.__height ): lowerCamelCase : Optional[Any] = [ self.__matrix[i][j] - other.component(__magic_name__ , __magic_name__ ) for j in range(self.__width ) ] matrix.append(__magic_name__ ) return Matrix(__magic_name__ , self.__width , self.__height ) else: raise Exception("""matrices must have the same dimension!""" ) @overload def __mul__( self , __magic_name__ ): ... @overload def __mul__( self , __magic_name__ ): ... def __mul__( self , __magic_name__ ): if isinstance(__magic_name__ , __magic_name__ ): # matrix-vector if len(__magic_name__ ) == self.__width: lowerCamelCase : Tuple = zero_vector(self.__height ) for i in range(self.__height ): lowerCamelCase : Union[str, Any] = [ self.__matrix[i][j] * other.component(__magic_name__ ) for j in range(self.__width ) ] ans.change_component(__magic_name__ , sum(__magic_name__ ) ) return ans else: raise Exception( """vector must have the same size as the """ """number of columns of the matrix!""" ) elif isinstance(__magic_name__ , (int, float) ): # matrix-scalar lowerCamelCase : Union[str, Any] = [ [self.__matrix[i][j] * other for j in range(self.__width )] for i in range(self.__height ) ] return Matrix(__magic_name__ , self.__width , self.__height ) return None def UpperCamelCase__ ( self ): return self.__height def UpperCamelCase__ ( self ): return self.__width def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): if 0 <= x < self.__height and 0 <= y < self.__width: return self.__matrix[x][y] else: raise Exception("""change_component: indices out of bounds""" ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ ): if 0 <= x < self.__height and 0 <= y < self.__width: lowerCamelCase : Tuple = value else: raise Exception("""change_component: indices out of bounds""" ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): if self.__height != self.__width: raise Exception("""Matrix is not square""" ) lowerCamelCase : str = self.__matrix[:x] + self.__matrix[x + 1 :] for i in range(len(__magic_name__ ) ): lowerCamelCase : List[str] = minor[i][:y] + minor[i][y + 1 :] return Matrix(__magic_name__ , self.__width - 1 , self.__height - 1 ).determinant() def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): if self.__height != self.__width: raise Exception("""Matrix is not square""" ) if 0 <= x < self.__height and 0 <= y < self.__width: return (-1) ** (x + y) * self.minor(__magic_name__ , __magic_name__ ) else: raise Exception("""Indices out of bounds""" ) def UpperCamelCase__ ( self ): if self.__height != self.__width: raise Exception("""Matrix is not square""" ) if self.__height < 1: raise Exception("""Matrix has no element""" ) elif self.__height == 1: return self.__matrix[0][0] elif self.__height == 2: return ( self.__matrix[0][0] * self.__matrix[1][1] - self.__matrix[0][1] * self.__matrix[1][0] ) else: lowerCamelCase : Tuple = [ self.__matrix[0][y] * self.cofactor(0 , __magic_name__ ) for y in range(self.__width ) ] return sum(__magic_name__ ) def _a ( lowerCamelCase ): lowerCamelCase : list[list[float]] = [[0] * n for _ in range(lowerCamelCase )] return Matrix(lowerCamelCase, lowerCamelCase, lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): random.seed(lowerCamelCase ) lowerCamelCase : list[list[float]] = [ [random.randint(lowerCamelCase, lowerCamelCase ) for _ in range(lowerCamelCase )] for _ in range(lowerCamelCase ) ] return Matrix(lowerCamelCase, lowerCamelCase, lowerCamelCase )
287
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = """gpt_neo""" _UpperCAmelCase : Union[str, Any] = ["""past_key_values"""] _UpperCAmelCase : List[Any] = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , __magic_name__=5_0_2_5_7 , __magic_name__=2_0_4_8 , __magic_name__=2_0_4_8 , __magic_name__=2_4 , __magic_name__=[[["global", "local"], 1_2]] , __magic_name__=1_6 , __magic_name__=None , __magic_name__=2_5_6 , __magic_name__="gelu_new" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=1e-5 , __magic_name__=0.02 , __magic_name__=True , __magic_name__=5_0_2_5_6 , __magic_name__=5_0_2_5_6 , **__magic_name__ , ): lowerCamelCase : List[Any] = vocab_size lowerCamelCase : str = max_position_embeddings lowerCamelCase : str = hidden_size lowerCamelCase : Optional[int] = num_layers lowerCamelCase : str = num_heads lowerCamelCase : Optional[Any] = intermediate_size lowerCamelCase : List[Any] = window_size lowerCamelCase : int = activation_function lowerCamelCase : Union[str, Any] = resid_dropout lowerCamelCase : List[Any] = embed_dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Dict = classifier_dropout lowerCamelCase : Any = layer_norm_epsilon lowerCamelCase : Dict = initializer_range lowerCamelCase : Dict = use_cache lowerCamelCase : Optional[Any] = bos_token_id lowerCamelCase : int = eos_token_id lowerCamelCase : List[Any] = attention_types lowerCamelCase : Optional[Any] = self.expand_attention_types_params(__magic_name__ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.attention_layers)` == `config.num_layers` """ F'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' """`config.attention_layers` is prepared using `config.attention_types`. """ """Please verify the value of `config.attention_types` argument.""" ) super().__init__(bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ ): lowerCamelCase : Optional[int] = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : Any = input.size() lowerCamelCase : List[Any] = len(lowerCamelCase ) lowerCamelCase : Optional[Any] = shape[dimension] lowerCamelCase : Optional[int] = torch.arange(0, lowerCamelCase, lowerCamelCase ) lowerCamelCase : Dict = torch.div(sizedim - size, lowerCamelCase, rounding_mode="""floor""" ) + 1 lowerCamelCase : int = torch.arange(lowerCamelCase ) + low_indices[:min_length][:, None] lowerCamelCase : str = [slice(lowerCamelCase )] * rank lowerCamelCase : List[str] = indices lowerCamelCase : Dict = input[s] lowerCamelCase : Any = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : List[Any] = torch.arange(1, lowerCamelCase ) lowerCamelCase : Optional[int] = torch.remainder(lowerCamelCase, lowerCamelCase ) lowerCamelCase : List[Any] = remainders == 0 lowerCamelCase : List[Any] = candidates[divisor_indices] lowerCamelCase : Optional[Any] = torch.max(lowerCamelCase ) return largest_divisor, torch.div(lowerCamelCase, lowerCamelCase, rounding_mode="""floor""" ) class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): lowerCamelCase : str = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(__magic_name__ , direction="""inputs""" ) lowerCamelCase : int = {0: """batch""", 1: """past_sequence + sequence"""} else: lowerCamelCase : Tuple = {0: """batch""", 1: """sequence"""} return common_inputs @property def UpperCamelCase__ ( self ): return self._config.num_heads def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ): lowerCamelCase : Optional[int] = super(__magic_name__ , self ).generate_dummy_inputs( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) # We need to order the input in the way they appears in the forward() lowerCamelCase : int = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch lowerCamelCase , lowerCamelCase : Optional[Any] = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values lowerCamelCase : Optional[int] = seqlen + 2 lowerCamelCase : List[Any] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowerCamelCase : str = [ (torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) for _ in range(self.num_layers ) ] lowerCamelCase : Tuple = common_inputs["""attention_mask"""] if self.use_past: lowerCamelCase : str = ordered_inputs["""attention_mask"""].dtype lowerCamelCase : Any = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(__magic_name__ , __magic_name__ , dtype=__magic_name__ )] , dim=1 ) return ordered_inputs @property def UpperCamelCase__ ( self ): return 1_3
287
1
from __future__ import annotations def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowerCamelCase : int = list(range(len(lowerCamelCase ) ) ) lowerCamelCase : Optional[int] = [v / w for v, w in zip(lowerCamelCase, lowerCamelCase )] index.sort(key=lambda lowerCamelCase : ratio[i], reverse=lowerCamelCase ) lowerCamelCase : float = 0 lowerCamelCase : list[float] = [0] * len(lowerCamelCase ) for i in index: if weight[i] <= capacity: lowerCamelCase : Union[str, Any] = 1 max_value += value[i] capacity -= weight[i] else: lowerCamelCase : Tuple = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
287
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
287
1
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: _lowerCamelCase =None _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _lowerCamelCase ={ """vocab_file""": { """facebook/mbart-large-en-ro""": ( """https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model""" ), """facebook/mbart-large-cc25""": ( """https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """facebook/mbart-large-en-ro""": """https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json""", """facebook/mbart-large-cc25""": """https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json""", }, } _lowerCamelCase ={ """facebook/mbart-large-en-ro""": 1_0_2_4, """facebook/mbart-large-cc25""": 1_0_2_4, } # fmt: off _lowerCamelCase =["""ar_AR""", """cs_CZ""", """de_DE""", """en_XX""", """es_XX""", """et_EE""", """fi_FI""", """fr_XX""", """gu_IN""", """hi_IN""", """it_IT""", """ja_XX""", """kk_KZ""", """ko_KR""", """lt_LT""", """lv_LV""", """my_MM""", """ne_NP""", """nl_XX""", """ro_RO""", """ru_RU""", """si_LK""", """tr_TR""", """vi_VN""", """zh_CN"""] class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : int = VOCAB_FILES_NAMES _UpperCAmelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : Optional[int] = ["""input_ids""", """attention_mask"""] _UpperCAmelCase : Optional[int] = MBartTokenizer _UpperCAmelCase : List[int] = [] _UpperCAmelCase : List[int] = [] def __init__( self , __magic_name__=None , __magic_name__=None , __magic_name__="<s>" , __magic_name__="</s>" , __magic_name__="</s>" , __magic_name__="<s>" , __magic_name__="<unk>" , __magic_name__="<pad>" , __magic_name__="<mask>" , __magic_name__=None , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase : Tuple = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else mask_token super().__init__( vocab_file=__magic_name__ , tokenizer_file=__magic_name__ , bos_token=__magic_name__ , eos_token=__magic_name__ , sep_token=__magic_name__ , cls_token=__magic_name__ , unk_token=__magic_name__ , pad_token=__magic_name__ , mask_token=__magic_name__ , src_lang=__magic_name__ , tgt_lang=__magic_name__ , additional_special_tokens=__magic_name__ , **__magic_name__ , ) lowerCamelCase : List[Any] = vocab_file lowerCamelCase : Optional[int] = False if not self.vocab_file else True lowerCamelCase : List[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"""additional_special_tokens""": _additional_special_tokens} ) lowerCamelCase : Union[str, Any] = { lang_code: self.convert_tokens_to_ids(__magic_name__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCamelCase : Any = src_lang if src_lang is not None else """en_XX""" lowerCamelCase : Optional[Any] = self.convert_tokens_to_ids(self._src_lang ) lowerCamelCase : Dict = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def UpperCamelCase__ ( self ): return self._src_lang @src_lang.setter def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): lowerCamelCase : Tuple = [self.sep_token_id] lowerCamelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , **__magic_name__ ): if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) lowerCamelCase : Optional[Any] = src_lang lowerCamelCase : Union[str, Any] = self(__magic_name__ , add_special_tokens=__magic_name__ , return_tensors=__magic_name__ , **__magic_name__ ) lowerCamelCase : Optional[Any] = self.convert_tokens_to_ids(__magic_name__ ) lowerCamelCase : int = tgt_lang_id return inputs def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = "en_XX" , __magic_name__ = None , __magic_name__ = "ro_RO" , **__magic_name__ , ): lowerCamelCase : Union[str, Any] = src_lang lowerCamelCase : Optional[Any] = tgt_lang return super().prepare_seqaseq_batch(__magic_name__ , __magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self ): return self.set_src_lang_special_tokens(self.src_lang ) def UpperCamelCase__ ( self ): return self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : str = self.convert_tokens_to_ids(__magic_name__ ) lowerCamelCase : List[str] = [] lowerCamelCase : Any = [self.eos_token_id, self.cur_lang_code] lowerCamelCase : Optional[int] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase : Dict = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase : Any = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : Any = self.convert_tokens_to_ids(__magic_name__ ) lowerCamelCase : int = [] lowerCamelCase : Tuple = [self.eos_token_id, self.cur_lang_code] lowerCamelCase : Optional[int] = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase : Tuple = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase : Union[str, Any] = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(__magic_name__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory.''' ) return lowerCamelCase : List[str] = os.path.join( __magic_name__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__magic_name__ ): copyfile(self.vocab_file , __magic_name__ ) return (out_vocab_file,)
287
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): for param, grad_param in zip(model_a.parameters(), model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=True ): model.train() lowerCamelCase : Dict = model(lowerCamelCase ) lowerCamelCase : Any = F.mse_loss(lowerCamelCase, target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase=False ): set_seed(42 ) lowerCamelCase : Tuple = RegressionModel() lowerCamelCase : Any = deepcopy(lowerCamelCase ) lowerCamelCase : Any = RegressionDataset(length=80 ) lowerCamelCase : Dict = DataLoader(lowerCamelCase, batch_size=16 ) model.to(accelerator.device ) if sched: lowerCamelCase : int = AdamW(params=model.parameters(), lr=1e-3 ) lowerCamelCase : Optional[Any] = AdamW(params=ddp_model.parameters(), lr=1e-3 ) lowerCamelCase : str = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) lowerCamelCase : Tuple = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) # Make a copy of `model` if sched: lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = accelerator.prepare(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: lowerCamelCase , lowerCamelCase : List[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def _a ( lowerCamelCase ): # Test when on a single CPU or GPU that the context manager does nothing lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : int = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad, ddp_param.grad ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : List[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase ): # Test on distributed setup that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : int = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Any = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : str = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Optional[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : Any = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : Any = get_training_setup(lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : str = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(lowerCamelCase ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Any = ddp_input[torch.randperm(len(lowerCamelCase ) )] GradientState._reset_state() def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : List[Any] = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase, lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(lowerCamelCase )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]["lr"]}\nDDP opt: {ddp_opt.param_groups[0]["lr"]}\n''' lowerCamelCase : Union[str, Any] = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(lowerCamelCase )) if accelerator.num_processes > 1: check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def _a ( ): lowerCamelCase : int = Accelerator() lowerCamelCase : Optional[Any] = RegressionDataset(length=80 ) lowerCamelCase : List[str] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase : int = RegressionDataset(length=96 ) lowerCamelCase : Optional[int] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if iteration < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if batch_num < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def _a ( ): lowerCamelCase : List[Any] = Accelerator() lowerCamelCase : int = accelerator.state if state.local_process_index == 0: print("""**Test `accumulate` gradient accumulation with dataloader break**""" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("""**Test NOOP `no_sync` context manager**""" ) test_noop_sync(lowerCamelCase ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("""**Test Distributed `no_sync` context manager**""" ) test_distributed_sync(lowerCamelCase ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation(lowerCamelCase, lowerCamelCase ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("""<""", """2.0""" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, """`split_batches=False`, `dispatch_batches=False`**""", ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation_with_opt_and_scheduler(lowerCamelCase, lowerCamelCase ) def _a ( lowerCamelCase ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
287
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower_vision_model""" def __init__( self , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=3 , __magic_name__=1_6 , __magic_name__=2_8_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__=True , __magic_name__=False , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : str = num_hidden_layers lowerCamelCase : Optional[int] = num_channels lowerCamelCase : List[str] = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Any = initializer_factor lowerCamelCase : Tuple = layer_norm_eps lowerCamelCase : Tuple = stop_gradient lowerCamelCase : Optional[int] = share_layernorm lowerCamelCase : str = remove_last_layer @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : str = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """bridgetower_text_model""" def __init__( self , __magic_name__=5_0_2_6_5 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=1 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_4 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Any = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : Tuple = hidden_act lowerCamelCase : Optional[int] = initializer_factor lowerCamelCase : Any = intermediate_size lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : str = max_position_embeddings lowerCamelCase : Union[str, Any] = type_vocab_size lowerCamelCase : Optional[int] = layer_norm_eps lowerCamelCase : Optional[int] = position_embedding_type lowerCamelCase : List[str] = use_cache lowerCamelCase : List[str] = pad_token_id lowerCamelCase : List[str] = bos_token_id lowerCamelCase : Optional[int] = eos_token_id @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : Optional[int] = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower""" def __init__( self , __magic_name__=True , __magic_name__="gelu" , __magic_name__=7_6_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__="add" , __magic_name__=1_2 , __magic_name__=6 , __magic_name__=False , __magic_name__=False , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): # TODO: remove this once the Hub files are updated. lowerCamelCase : int = kwargs.pop("""text_config_dict""" , __magic_name__ ) lowerCamelCase : str = kwargs.pop("""vision_config_dict""" , __magic_name__ ) super().__init__(**__magic_name__ ) lowerCamelCase : str = share_cross_modal_transformer_layers lowerCamelCase : Union[str, Any] = hidden_act lowerCamelCase : str = hidden_size lowerCamelCase : Tuple = initializer_factor lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = share_link_tower_layers lowerCamelCase : List[Any] = link_tower_type lowerCamelCase : Tuple = num_attention_heads lowerCamelCase : int = num_hidden_layers lowerCamelCase : Union[str, Any] = tie_word_embeddings lowerCamelCase : Tuple = init_layernorm_from_vision_encoder if text_config is None: lowerCamelCase : Any = {} logger.info("""`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.""" ) if vision_config is None: lowerCamelCase : int = {} logger.info("""`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.""" ) lowerCamelCase : Any = BridgeTowerTextConfig(**__magic_name__ ) lowerCamelCase : Optional[Any] = BridgeTowerVisionConfig(**__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , **__magic_name__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = copy.deepcopy(self.__dict__ ) lowerCamelCase : int = self.text_config.to_dict() lowerCamelCase : Dict = self.vision_config.to_dict() lowerCamelCase : List[str] = self.__class__.model_type return output
287
from scipy.stats import pearsonr import datasets _lowerCamelCase =""" Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ _lowerCamelCase =""" Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ _lowerCamelCase =""" @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class A__ ( datasets.Metric): def UpperCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"""] , ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__=False ): if return_pvalue: lowerCamelCase : Optional[Any] = pearsonr(__magic_name__ , __magic_name__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(__magic_name__ , __magic_name__ )[0] )}
287
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : int = """megatron-bert""" def __init__( self , __magic_name__=2_9_0_5_6 , __magic_name__=1_0_2_4 , __magic_name__=2_4 , __magic_name__=1_6 , __magic_name__=4_0_9_6 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=0 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ): super().__init__(pad_token_id=__magic_name__ , **__magic_name__ ) lowerCamelCase : List[Any] = vocab_size lowerCamelCase : List[str] = hidden_size lowerCamelCase : Tuple = num_hidden_layers lowerCamelCase : Optional[int] = num_attention_heads lowerCamelCase : Tuple = hidden_act lowerCamelCase : Union[str, Any] = intermediate_size lowerCamelCase : List[Any] = hidden_dropout_prob lowerCamelCase : Tuple = attention_probs_dropout_prob lowerCamelCase : int = max_position_embeddings lowerCamelCase : Tuple = type_vocab_size lowerCamelCase : str = initializer_range lowerCamelCase : int = layer_norm_eps lowerCamelCase : Optional[Any] = position_embedding_type lowerCamelCase : Optional[int] = use_cache
287
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """microsoft/conditional-detr-resnet-50""": ( """https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """conditional_detr""" _UpperCAmelCase : Optional[int] = ["""past_key_values"""] _UpperCAmelCase : Optional[Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , __magic_name__=True , __magic_name__=None , __magic_name__=3 , __magic_name__=3_0_0 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=True , __magic_name__="relu" , __magic_name__=2_5_6 , __magic_name__=0.1 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.02 , __magic_name__=1.0 , __magic_name__=False , __magic_name__="sine" , __magic_name__="resnet50" , __magic_name__=True , __magic_name__=False , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=1 , __magic_name__=1 , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=0.25 , **__magic_name__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) lowerCamelCase : Optional[int] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(__magic_name__ , __magic_name__ ): lowerCamelCase : List[Any] = backbone_config.get("""model_type""" ) lowerCamelCase : Dict = CONFIG_MAPPING[backbone_model_type] lowerCamelCase : str = config_class.from_dict(__magic_name__ ) lowerCamelCase : Dict = use_timm_backbone lowerCamelCase : str = backbone_config lowerCamelCase : Tuple = num_channels lowerCamelCase : Dict = num_queries lowerCamelCase : Any = d_model lowerCamelCase : Optional[Any] = encoder_ffn_dim lowerCamelCase : List[str] = encoder_layers lowerCamelCase : Union[str, Any] = encoder_attention_heads lowerCamelCase : Any = decoder_ffn_dim lowerCamelCase : Dict = decoder_layers lowerCamelCase : Union[str, Any] = decoder_attention_heads lowerCamelCase : Dict = dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Union[str, Any] = activation_dropout lowerCamelCase : Optional[int] = activation_function lowerCamelCase : int = init_std lowerCamelCase : str = init_xavier_std lowerCamelCase : Tuple = encoder_layerdrop lowerCamelCase : str = decoder_layerdrop lowerCamelCase : Tuple = encoder_layers lowerCamelCase : Optional[int] = auxiliary_loss lowerCamelCase : Optional[Any] = position_embedding_type lowerCamelCase : Optional[int] = backbone lowerCamelCase : Union[str, Any] = use_pretrained_backbone lowerCamelCase : str = dilation # Hungarian matcher lowerCamelCase : Optional[Any] = class_cost lowerCamelCase : Dict = bbox_cost lowerCamelCase : Tuple = giou_cost # Loss coefficients lowerCamelCase : Union[str, Any] = mask_loss_coefficient lowerCamelCase : Dict = dice_loss_coefficient lowerCamelCase : Optional[int] = cls_loss_coefficient lowerCamelCase : Optional[int] = bbox_loss_coefficient lowerCamelCase : Optional[int] = giou_loss_coefficient lowerCamelCase : Optional[int] = focal_alpha super().__init__(is_encoder_decoder=__magic_name__ , **__magic_name__ ) @property def UpperCamelCase__ ( self ): return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): return self.d_model def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: lowerCamelCase : Optional[int] = self.backbone_config.to_dict() lowerCamelCase : Optional[Any] = self.__class__.model_type return output class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-5 @property def UpperCamelCase__ ( self ): return 1_2
287
1
_lowerCamelCase ={ "joule": 1.0, "kilojoule": 1_0_0_0, "megajoule": 1_0_0_0_0_0_0, "gigajoule": 1_0_0_0_0_0_0_0_0_0, "wattsecond": 1.0, "watthour": 3_6_0_0, "kilowatthour": 3_6_0_0_0_0_0, "newtonmeter": 1.0, "calorie_nutr": 4_1_8_6.8, "kilocalorie_nutr": 4_1_8_6_8_0_0.0_0, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1_0_5_5.0_5_5_8_5, "footpound": 1.355818, } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowerCamelCase : Dict = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {", ".join(lowerCamelCase )}''' ) raise ValueError(lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
287
import json import sys def _a ( lowerCamelCase, lowerCamelCase ): with open(lowerCamelCase, encoding="""utf-8""" ) as f: lowerCamelCase : List[Any] = json.load(lowerCamelCase ) lowerCamelCase : Optional[Any] = ["""<details>""", """<summary>Show updated benchmarks!</summary>""", """ """] for benchmark_name in sorted(lowerCamelCase ): lowerCamelCase : List[Any] = results[benchmark_name] lowerCamelCase : Union[str, Any] = benchmark_name.split("""/""" )[-1] output_md.append(F'''### Benchmark: {benchmark_file_name}''' ) lowerCamelCase : Any = """| metric |""" lowerCamelCase : str = """|--------|""" lowerCamelCase : List[Any] = """| new / old (diff) |""" for metric_name in sorted(lowerCamelCase ): lowerCamelCase : List[Any] = benchmark_res[metric_name] lowerCamelCase : Tuple = metric_vals["""new"""] lowerCamelCase : int = metric_vals.get("""old""", lowerCamelCase ) lowerCamelCase : Dict = metric_vals.get("""diff""", lowerCamelCase ) lowerCamelCase : Dict = F''' {new_val:f}''' if isinstance(lowerCamelCase, (int, float) ) else """None""" if old_val is not None: val_str += F''' / {old_val:f}''' if isinstance(lowerCamelCase, (int, float) ) else "None" if dif_val is not None: val_str += F''' ({dif_val:f})''' if isinstance(lowerCamelCase, (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append("""</details>""" ) with open(lowerCamelCase, """w""", encoding="""utf-8""" ) as f: f.writelines("""\n""".join(lowerCamelCase ) ) if __name__ == "__main__": _lowerCamelCase =sys.argv[1] _lowerCamelCase =sys.argv[2] format_json_to_md(input_json_file, output_md_file)
287
1
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _lowerCamelCase =get_logger(__name__) class A__ : def __init__( self , __magic_name__ = None ): lowerCamelCase : Dict = ( os.path.join(__magic_name__ , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) lowerCamelCase : List[str] = Extractor def UpperCamelCase__ ( self , __magic_name__ ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" lowerCamelCase : int = os.path.abspath(__magic_name__ ) return os.path.join(self.extract_dir , hash_url_to_filename(__magic_name__ ) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): return force_extract or ( not os.path.isfile(__magic_name__ ) and not (os.path.isdir(__magic_name__ ) and os.listdir(__magic_name__ )) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = False ): lowerCamelCase : Union[str, Any] = self.extractor.infer_extractor_format(__magic_name__ ) if not extractor_format: return input_path lowerCamelCase : int = self._get_output_path(__magic_name__ ) if self._do_extract(__magic_name__ , __magic_name__ ): self.extractor.extract(__magic_name__ , __magic_name__ , __magic_name__ ) return output_path class A__ ( __SCREAMING_SNAKE_CASE): @classmethod @abstractmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): ... @staticmethod @abstractmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): ... class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[bytes] = [] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with open(__magic_name__ , """rb""" ) as f: return f.read(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if not magic_number: lowerCamelCase : Optional[Any] = max(len(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) try: lowerCamelCase : Tuple = cls.read_magic_number(__magic_name__ , __magic_name__ ) except OSError: return False return any(magic_number.startswith(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) class A__ ( __SCREAMING_SNAKE_CASE): @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): return tarfile.is_tarfile(__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): def resolved(__magic_name__ ) -> str: return os.path.realpath(os.path.abspath(__magic_name__ ) ) def badpath(__magic_name__ , __magic_name__ ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__magic_name__ , __magic_name__ ) ).startswith(__magic_name__ ) def badlink(__magic_name__ , __magic_name__ ) -> bool: # Links are interpreted relative to the directory containing the link lowerCamelCase : List[str] = resolved(os.path.join(__magic_name__ , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=__magic_name__ ) lowerCamelCase : Optional[Any] = resolved(__magic_name__ ) for finfo in members: if badpath(finfo.name , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Dict = tarfile.open(__magic_name__ ) tar_file.extractall(__magic_name__ , members=TarExtractor.safemembers(__magic_name__ , __magic_name__ ) ) tar_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = [B"""\x1F\x8B"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with gzip.open(__magic_name__ , """rb""" ) as gzip_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[int] = [ B"""PK\x03\x04""", B"""PK\x05\x06""", # empty archive B"""PK\x07\x08""", # spanned archive ] @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if super().is_extractable(__magic_name__ , magic_number=__magic_name__ ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__magic_name__ , """rb""" ) as fp: lowerCamelCase : List[str] = _EndRecData(__magic_name__ ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: lowerCamelCase : List[Any] = fp.read(__magic_name__ ) # CD is where we expect it to be if len(__magic_name__ ) == sizeCentralDir: lowerCamelCase : str = struct.unpack(__magic_name__ , __magic_name__ ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with zipfile.ZipFile(__magic_name__ , """r""" ) as zip_file: zip_file.extractall(__magic_name__ ) zip_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[str] = [B"""\xFD\x37\x7A\x58\x5A\x00"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with lzma.open(__magic_name__ ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""Rar!\x1a\x07\x00""", B"""Rar!\x1a\x07\x01\x00"""] # RAR_ID # RAR5_ID @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Union[str, Any] = rarfile.RarFile(__magic_name__ ) rf.extractall(__magic_name__ ) rf.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = [B"""\x28\xb5\x2F\xFD"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd lowerCamelCase : int = zstd.ZstdDecompressor() with open(__magic_name__ , """rb""" ) as ifh, open(__magic_name__ , """wb""" ) as ofh: dctx.copy_stream(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""\x42\x5A\x68"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with bza.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x37\x7A\xBC\xAF\x27\x1C"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with pyazr.SevenZipFile(__magic_name__ , """r""" ) as archive: archive.extractall(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x04\x22\x4D\x18"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) _UpperCAmelCase : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def UpperCamelCase__ ( cls ): return max( len(__magic_name__ ) for extractor in cls.extractors.values() if issubclass(__magic_name__ , __magic_name__ ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): try: return MagicNumberBaseExtractor.read_magic_number(__magic_name__ , magic_number_length=__magic_name__ ) except OSError: return b"" @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = False ): warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = cls.infer_extractor_format(__magic_name__ ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): # <Added version="2.4.0"/> lowerCamelCase : Dict = cls._get_magic_number_max_length() lowerCamelCase : Optional[Any] = cls._read_magic_number(__magic_name__ , __magic_name__ ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__magic_name__ , magic_number=__magic_name__ ): return extractor_format @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = "deprecated" , ): os.makedirs(os.path.dirname(__magic_name__ ) , exist_ok=__magic_name__ ) # Prevent parallel extractions lowerCamelCase : Tuple = str(Path(__magic_name__ ).with_suffix(""".lock""" ) ) with FileLock(__magic_name__ ): shutil.rmtree(__magic_name__ , ignore_errors=__magic_name__ ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__magic_name__ , __magic_name__ ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = extractor if extractor != """deprecated""" else extractor_format else: lowerCamelCase : Optional[int] = cls.extractors[extractor_format] return extractor.extract(__magic_name__ , __magic_name__ ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=__magic_name__ , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__magic_name__ ): return extractor.extract(__magic_name__ , __magic_name__ )
287
def _a ( lowerCamelCase ): return " ".join( """""".join(word[::-1] ) if len(lowerCamelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("""Hey wollef sroirraw"""))
287
1
import os import pytest from attr import dataclass _lowerCamelCase ="""us-east-1""" # defaults region @dataclass class A__ : _UpperCAmelCase : str _UpperCAmelCase : Any = """arn:aws:iam::558105141721:role/sagemaker_execution_role""" _UpperCAmelCase : Optional[int] = { """task_name""": """mnli""", """per_device_train_batch_size""": 16, """per_device_eval_batch_size""": 16, """do_train""": True, """do_eval""": True, """do_predict""": True, """output_dir""": """/opt/ml/model""", """overwrite_output_dir""": True, """max_steps""": 500, """save_steps""": 5500, } _UpperCAmelCase : Dict = {**hyperparameters, """max_steps""": 1000} @property def UpperCamelCase__ ( self ): if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def UpperCamelCase__ ( self ): return F'''{self.framework}-transfromers-test''' @property def UpperCamelCase__ ( self ): return F'''./tests/sagemaker/scripts/{self.framework}''' @property def UpperCamelCase__ ( self ): if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope="""class""" ) def _a ( lowerCamelCase ): lowerCamelCase : List[Any] = SageMakerTestEnvironment(framework=request.cls.framework )
287
# Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from packaging import version from .. import __version__ from .constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD from .doc import ( add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, copy_func, replace_return_docstrings, ) from .generic import ( ContextManagers, ExplicitEnum, ModelOutput, PaddingStrategy, TensorType, add_model_info_to_auto_map, cached_property, can_return_loss, expand_dims, find_labels, flatten_dict, infer_framework, is_jax_tensor, is_numpy_array, is_tensor, is_tf_symbolic_tensor, is_tf_tensor, is_torch_device, is_torch_dtype, is_torch_tensor, reshape, squeeze, strtobool, tensor_size, to_numpy, to_py_obj, transpose, working_or_temp_dir, ) from .hub import ( CLOUDFRONT_DISTRIB_PREFIX, DISABLE_TELEMETRY, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, EntryNotFoundError, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, cached_file, default_cache_path, define_sagemaker_information, download_url, extract_commit_hash, get_cached_models, get_file_from_repo, get_full_repo_name, has_file, http_user_agent, is_offline_mode, is_remote_url, move_cache, send_example_telemetry, try_to_load_from_cache, ) from .import_utils import ( ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, TORCH_FX_REQUIRED_VERSION, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, ccl_version, direct_transformers_import, get_torch_version, is_accelerate_available, is_apex_available, is_bitsandbytes_available, is_bsa_available, is_coloredlogs_available, is_cython_available, is_datasets_available, is_decord_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_jieba_available, is_jumanpp_available, is_kenlm_available, is_keras_nlp_available, is_librosa_available, is_natten_available, is_ninja_available, is_onnx_available, is_openai_available, is_optimum_available, is_pandas_available, is_peft_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytest_available, is_pytorch_quantization_available, is_rjieba_available, is_sacremoses_available, is_safetensors_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_sudachi_available, is_tensorflow_probability_available, is_tensorflow_text_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_bfaa_cpu_available, is_torch_bfaa_gpu_available, is_torch_compile_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_neuroncore_available, is_torch_tensorrt_fx_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_torchdistx_available, is_torchdynamo_available, is_torchvision_available, is_training_run_on_sagemaker, is_vision_available, requires_backends, torch_only_method, ) _lowerCamelCase ="""pytorch_model.bin""" _lowerCamelCase ="""pytorch_model.bin.index.json""" _lowerCamelCase ="""adapter_config.json""" _lowerCamelCase ="""adapter_model.bin""" _lowerCamelCase ="""adapter_model.safetensors""" _lowerCamelCase ="""tf_model.h5""" _lowerCamelCase ="""tf_model.h5.index.json""" _lowerCamelCase ="""model.ckpt""" _lowerCamelCase ="""flax_model.msgpack""" _lowerCamelCase ="""flax_model.msgpack.index.json""" _lowerCamelCase ="""model.safetensors""" _lowerCamelCase ="""model.safetensors.index.json""" _lowerCamelCase ="""config.json""" _lowerCamelCase ="""preprocessor_config.json""" _lowerCamelCase =FEATURE_EXTRACTOR_NAME _lowerCamelCase ="""generation_config.json""" _lowerCamelCase ="""modelcard.json""" _lowerCamelCase ="""▁""" _lowerCamelCase =SENTENCEPIECE_UNDERLINE # Kept for backward compatibility _lowerCamelCase =[ [[0, 1, 0, 1], [1, 0, 0, 1]] ] * 2 # Needs to have 0s and 1s only since XLM uses it for langs too. _lowerCamelCase =[[7, 6, 0, 0, 1], [1, 2, 3, 0, 0], [0, 0, 0, 4, 5]] _lowerCamelCase =[[1, 1, 1, 1, 1], [1, 1, 1, 0, 0], [0, 0, 0, 1, 1]] def _a ( lowerCamelCase ): if version.parse(lowerCamelCase ) < version.parse(lowerCamelCase ): if "dev" in min_version: lowerCamelCase : Optional[int] = ( """This example requires a source install from HuggingFace Transformers (see """ """`https://huggingface.co/docs/transformers/installation#install-from-source`),""" ) else: lowerCamelCase : int = F'''This example requires a minimum version of {min_version},''' error_message += F''' but the version found is {__version__}.\n''' raise ImportError( error_message + """Check out https://github.com/huggingface/transformers/tree/main/examples#important-note for the examples corresponding to other """ """versions of HuggingFace Transformers.""" )
287
1
import os import shutil import sys import tempfile import unittest from pathlib import Path import pytest import transformers from transformers import ( BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoTokenizer, BertConfig, BertTokenizer, BertTokenizerFast, CTRLTokenizer, GPTaTokenizer, GPTaTokenizerFast, PreTrainedTokenizerFast, RobertaTokenizer, RobertaTokenizerFast, is_tokenizers_available, ) from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.auto.tokenization_auto import ( TOKENIZER_MAPPING, get_tokenizer_config, tokenizer_class_from_name, ) from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import ( DUMMY_DIFF_TOKENIZER_IDENTIFIER, DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tokenizers, slow, ) sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = 0 @slow def UpperCamelCase__ ( self ): for model_name in (x for x in BERT_PRETRAINED_CONFIG_ARCHIVE_MAP.keys() if "japanese" not in x): lowerCamelCase : Any = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) self.assertGreater(len(__magic_name__ ) , 0 ) for model_name in GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP.keys(): lowerCamelCase : Tuple = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) self.assertIsInstance(__magic_name__ , (GPTaTokenizer, GPTaTokenizerFast) ) self.assertGreater(len(__magic_name__ ) , 0 ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 1_2 ) def UpperCamelCase__ ( self ): lowerCamelCase : int = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , (RobertaTokenizer, RobertaTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 2_0 ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = AutoConfig.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) # Check that tokenizer_type ≠ model_type lowerCamelCase : int = AutoTokenizer.from_pretrained(__magic_name__ , config=__magic_name__ ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 1_2 ) def UpperCamelCase__ ( self ): with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.txt""" , os.path.join(__magic_name__ , """vocab.txt""" ) ) lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained(__magic_name__ , tokenizer_type="""bert""" , use_fast=__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.json""" , os.path.join(__magic_name__ , """vocab.json""" ) ) shutil.copy("""./tests/fixtures/merges.txt""" , os.path.join(__magic_name__ , """merges.txt""" ) ) lowerCamelCase : List[str] = AutoTokenizer.from_pretrained(__magic_name__ , tokenizer_type="""gpt2""" , use_fast=__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) @require_tokenizers def UpperCamelCase__ ( self ): with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.txt""" , os.path.join(__magic_name__ , """vocab.txt""" ) ) lowerCamelCase : List[str] = AutoTokenizer.from_pretrained(__magic_name__ , tokenizer_type="""bert""" ) self.assertIsInstance(__magic_name__ , __magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy("""./tests/fixtures/vocab.json""" , os.path.join(__magic_name__ , """vocab.json""" ) ) shutil.copy("""./tests/fixtures/merges.txt""" , os.path.join(__magic_name__ , """merges.txt""" ) ) lowerCamelCase : Any = AutoTokenizer.from_pretrained(__magic_name__ , tokenizer_type="""gpt2""" ) self.assertIsInstance(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): with pytest.raises(__magic_name__ ): AutoTokenizer.from_pretrained("""./""" , tokenizer_type="""xxx""" ) @require_tokenizers def UpperCamelCase__ ( self ): for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: lowerCamelCase : Union[str, Any] = tokenizer_class.from_pretrained("""wietsedv/bert-base-dutch-cased""" ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) if isinstance(__magic_name__ , __magic_name__ ): self.assertEqual(tokenizer.basic_tokenizer.do_lower_case , __magic_name__ ) else: self.assertEqual(tokenizer.do_lower_case , __magic_name__ ) self.assertEqual(tokenizer.model_max_length , 5_1_2 ) @require_tokenizers def UpperCamelCase__ ( self ): for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: with self.assertRaisesRegex( __magic_name__ , """julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier""" , ): lowerCamelCase : List[str] = tokenizer_class.from_pretrained("""julien-c/herlolip-not-exists""" ) def UpperCamelCase__ ( self ): # tests: https://github.com/huggingface/transformers/pull/13251 # 1. models with `-`, e.g. xlm-roberta -> xlm_roberta # 2. models that don't remap 1-1 from model-name to model file, e.g., openai-gpt -> openai lowerCamelCase : Tuple = TOKENIZER_MAPPING.values() lowerCamelCase : Tuple = [] for slow_tok, fast_tok in tokenizers: if slow_tok is not None: tokenizer_names.append(slow_tok.__name__ ) if fast_tok is not None: tokenizer_names.append(fast_tok.__name__ ) for tokenizer_name in tokenizer_names: # must find the right class tokenizer_class_from_name(__magic_name__ ) @require_tokenizers def UpperCamelCase__ ( self ): self.assertIsInstance(AutoTokenizer.from_pretrained("""bert-base-cased""" , use_fast=__magic_name__ ) , __magic_name__ ) self.assertIsInstance(AutoTokenizer.from_pretrained("""bert-base-cased""" ) , __magic_name__ ) @require_tokenizers def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained("""distilbert-base-uncased""" , do_lower_case=__magic_name__ ) lowerCamelCase : str = """Hello, world. How are you?""" lowerCamelCase : int = tokenizer.tokenize(__magic_name__ ) self.assertEqual("""[UNK]""" , tokens[0] ) lowerCamelCase : Dict = AutoTokenizer.from_pretrained("""microsoft/mpnet-base""" , do_lower_case=__magic_name__ ) lowerCamelCase : Optional[Any] = tokenizer.tokenize(__magic_name__ ) self.assertEqual("""[UNK]""" , tokens[0] ) @require_tokenizers def UpperCamelCase__ ( self ): lowerCamelCase : Any = AutoTokenizer.from_pretrained("""robot-test/dummy-tokenizer-fast-with-model-config""" ) self.assertEqual(type(__magic_name__ ) , __magic_name__ ) self.assertEqual(tokenizer.model_max_length , 5_1_2 ) self.assertEqual(tokenizer.vocab_size , 3_0_0_0_0 ) self.assertEqual(tokenizer.unk_token , """[UNK]""" ) self.assertEqual(tokenizer.padding_side , """right""" ) self.assertEqual(tokenizer.truncation_side , """right""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , (BertTokenizer, BertTokenizerFast) ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) lowerCamelCase : Any = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , tokenizer.__class__ ) self.assertEqual(tokenizera.vocab_size , 1_2 ) def UpperCamelCase__ ( self ): lowerCamelCase : str = AutoTokenizer.from_pretrained("""ctrl""" ) # There is no fast CTRL so this always gives us a slow tokenizer. self.assertIsInstance(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): # Check we can load the tokenizer config of an online model. lowerCamelCase : Dict = get_tokenizer_config("""bert-base-cased""" ) lowerCamelCase : Optional[Any] = config.pop("""_commit_hash""" , __magic_name__ ) # If we ever update bert-base-cased tokenizer config, this dict here will need to be updated. self.assertEqual(__magic_name__ , {"""do_lower_case""": False} ) # This model does not have a tokenizer_config so we get back an empty dict. lowerCamelCase : Union[str, Any] = get_tokenizer_config(__magic_name__ ) self.assertDictEqual(__magic_name__ , {} ) # A tokenizer saved with `save_pretrained` always creates a tokenizer config. lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained(__magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) lowerCamelCase : List[str] = get_tokenizer_config(__magic_name__ ) # Check the class of the tokenizer was properly saved (note that it always saves the slow class). self.assertEqual(config["""tokenizer_class"""] , """BertTokenizer""" ) def UpperCamelCase__ ( self ): try: AutoConfig.register("""custom""" , __magic_name__ ) AutoTokenizer.register(__magic_name__ , slow_tokenizer_class=__magic_name__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__magic_name__ ): AutoTokenizer.register(__magic_name__ , slow_tokenizer_class=__magic_name__ ) lowerCamelCase : str = CustomTokenizer.from_pretrained(__magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) lowerCamelCase : Any = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] @require_tokenizers def UpperCamelCase__ ( self ): try: AutoConfig.register("""custom""" , __magic_name__ ) # Can register in two steps AutoTokenizer.register(__magic_name__ , slow_tokenizer_class=__magic_name__ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, None) ) AutoTokenizer.register(__magic_name__ , fast_tokenizer_class=__magic_name__ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) del TOKENIZER_MAPPING._extra_content[CustomConfig] # Can register in one step AutoTokenizer.register( __magic_name__ , slow_tokenizer_class=__magic_name__ , fast_tokenizer_class=__magic_name__ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__magic_name__ ): AutoTokenizer.register(__magic_name__ , fast_tokenizer_class=__magic_name__ ) # We pass through a bert tokenizer fast cause there is no converter slow to fast for our new toknizer # and that model does not have a tokenizer.json with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : Dict = BertTokenizerFast.from_pretrained(__magic_name__ ) bert_tokenizer.save_pretrained(__magic_name__ ) lowerCamelCase : Optional[Any] = CustomTokenizerFast.from_pretrained(__magic_name__ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained(__magic_name__ , use_fast=__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def UpperCamelCase__ ( self ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__magic_name__ ): lowerCamelCase : Dict = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(__magic_name__ ): lowerCamelCase : int = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ ) lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ ) self.assertTrue(tokenizer.special_attribute_present ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) lowerCamelCase : Tuple = AutoTokenizer.from_pretrained(__magic_name__ , trust_remote_code=__magic_name__ ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , """NewTokenizerFast""" ) # Test we can also load the slow version lowerCamelCase : int = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__magic_name__ ) lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(__magic_name__ , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) else: self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , """NewTokenizer""" ) @require_tokenizers def UpperCamelCase__ ( self ): class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = False class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : int = NewTokenizer _UpperCAmelCase : List[str] = False try: AutoConfig.register("""custom""" , __magic_name__ ) AutoTokenizer.register(__magic_name__ , slow_tokenizer_class=__magic_name__ ) AutoTokenizer.register(__magic_name__ , fast_tokenizer_class=__magic_name__ ) # If remote code is not set, the default is to use local lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertFalse(tokenizer.special_attribute_present ) lowerCamelCase : List[str] = AutoTokenizer.from_pretrained("""hf-internal-testing/test_dynamic_tokenizer""" , use_fast=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertFalse(tokenizer.special_attribute_present ) # If remote code is disabled, we load the local one. lowerCamelCase : Tuple = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertFalse(tokenizer.special_attribute_present ) lowerCamelCase : Tuple = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertFalse(tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub lowerCamelCase : Dict = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) self.assertTrue(tokenizer.special_attribute_present ) lowerCamelCase : int = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer""" , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) self.assertTrue(tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def UpperCamelCase__ ( self ): lowerCamelCase : Any = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer_legacy""" , trust_remote_code=__magic_name__ ) self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) # Test we can also load the slow version lowerCamelCase : int = AutoTokenizer.from_pretrained( """hf-internal-testing/test_dynamic_tokenizer_legacy""" , trust_remote_code=__magic_name__ , use_fast=__magic_name__ ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) else: self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) def UpperCamelCase__ ( self ): with self.assertRaisesRegex( __magic_name__ , """bert-base is not a local folder and is not a valid model identifier""" ): lowerCamelCase : Tuple = AutoTokenizer.from_pretrained("""bert-base""" ) def UpperCamelCase__ ( self ): with self.assertRaisesRegex( __magic_name__ , r"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): lowerCamelCase : Any = AutoTokenizer.from_pretrained(__magic_name__ , revision="""aaaaaa""" ) def UpperCamelCase__ ( self ): # Make sure we have cached the tokenizer. lowerCamelCase : List[str] = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) with RequestCounter() as counter: lowerCamelCase : List[str] = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
287
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = """camembert""" def __init__( self , __magic_name__=3_0_5_2_2 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , __magic_name__=None , **__magic_name__ , ): super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : int = num_hidden_layers lowerCamelCase : int = num_attention_heads lowerCamelCase : Optional[int] = hidden_act lowerCamelCase : List[Any] = intermediate_size lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Optional[int] = attention_probs_dropout_prob lowerCamelCase : Optional[int] = max_position_embeddings lowerCamelCase : str = type_vocab_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : int = layer_norm_eps lowerCamelCase : Any = position_embedding_type lowerCamelCase : Optional[int] = use_cache lowerCamelCase : Union[str, Any] = classifier_dropout class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): if self.task == "multiple-choice": lowerCamelCase : List[str] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase : List[str] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
287
1
def _a ( lowerCamelCase, lowerCamelCase ): if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) lowerCamelCase : Dict = str(bin(lowerCamelCase ) )[2:] # remove the leading "0b" lowerCamelCase : List[Any] = str(bin(lowerCamelCase ) )[2:] # remove the leading "0b" lowerCamelCase : str = max(len(lowerCamelCase ), len(lowerCamelCase ) ) return "0b" + "".join( str(int(char_a != char_b ) ) for char_a, char_b in zip(a_binary.zfill(lowerCamelCase ), b_binary.zfill(lowerCamelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
287
_lowerCamelCase ={ "joule": 1.0, "kilojoule": 1_0_0_0, "megajoule": 1_0_0_0_0_0_0, "gigajoule": 1_0_0_0_0_0_0_0_0_0, "wattsecond": 1.0, "watthour": 3_6_0_0, "kilowatthour": 3_6_0_0_0_0_0, "newtonmeter": 1.0, "calorie_nutr": 4_1_8_6.8, "kilocalorie_nutr": 4_1_8_6_8_0_0.0_0, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1_0_5_5.0_5_5_8_5, "footpound": 1.355818, } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowerCamelCase : Dict = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {", ".join(lowerCamelCase )}''' ) raise ValueError(lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
287
1
import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def _a ( lowerCamelCase ): if is_torch_version("""<""", """2.0.0""" ) or not hasattr(lowerCamelCase, """_dynamo""" ): return False return isinstance(lowerCamelCase, torch._dynamo.eval_frame.OptimizedModule ) def _a ( lowerCamelCase, lowerCamelCase = True ): lowerCamelCase : Optional[int] = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) lowerCamelCase : Optional[int] = is_compiled_module(lowerCamelCase ) if is_compiled: lowerCamelCase : Union[str, Any] = model lowerCamelCase : Tuple = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(lowerCamelCase, lowerCamelCase ): lowerCamelCase : str = model.module if not keep_fpaa_wrapper: lowerCamelCase : int = getattr(lowerCamelCase, """forward""" ) lowerCamelCase : Optional[Any] = model.__dict__.pop("""_original_forward""", lowerCamelCase ) if original_forward is not None: while hasattr(lowerCamelCase, """__wrapped__""" ): lowerCamelCase : Any = forward.__wrapped__ if forward == original_forward: break lowerCamelCase : Optional[Any] = forward if getattr(lowerCamelCase, """_converted_to_transformer_engine""", lowerCamelCase ): convert_model(lowerCamelCase, to_transformer_engine=lowerCamelCase ) if is_compiled: lowerCamelCase : Any = model lowerCamelCase : List[Any] = compiled_model return model def _a ( ): PartialState().wait_for_everyone() def _a ( lowerCamelCase, lowerCamelCase ): if PartialState().distributed_type == DistributedType.TPU: xm.save(lowerCamelCase, lowerCamelCase ) elif PartialState().local_process_index == 0: torch.save(lowerCamelCase, lowerCamelCase ) @contextmanager def _a ( **lowerCamelCase ): for key, value in kwargs.items(): lowerCamelCase : Optional[Any] = str(lowerCamelCase ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def _a ( lowerCamelCase ): if not hasattr(lowerCamelCase, """__qualname__""" ) and not hasattr(lowerCamelCase, """__name__""" ): lowerCamelCase : List[str] = getattr(lowerCamelCase, """__class__""", lowerCamelCase ) if hasattr(lowerCamelCase, """__qualname__""" ): return obj.__qualname__ if hasattr(lowerCamelCase, """__name__""" ): return obj.__name__ return str(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase ): for key, value in source.items(): if isinstance(lowerCamelCase, lowerCamelCase ): lowerCamelCase : int = destination.setdefault(lowerCamelCase, {} ) merge_dicts(lowerCamelCase, lowerCamelCase ) else: lowerCamelCase : Any = value return destination def _a ( lowerCamelCase = None ): if port is None: lowerCamelCase : List[Any] = 2_9500 with socket.socket(socket.AF_INET, socket.SOCK_STREAM ) as s: return s.connect_ex(("""localhost""", port) ) == 0
287
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=__SCREAMING_SNAKE_CASE) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = field(default="""automatic-speech-recognition""" , metadata={"""include_in_asdict_even_if_is_default""": True}) _UpperCAmelCase : ClassVar[Features] = Features({"""audio""": Audio()}) _UpperCAmelCase : ClassVar[Features] = Features({"""transcription""": Value("""string""")}) _UpperCAmelCase : str = "audio" _UpperCAmelCase : str = "transcription" def UpperCamelCase__ ( self , __magic_name__ ): if self.audio_column not in features: raise ValueError(F'''Column {self.audio_column} is not present in features.''' ) if not isinstance(features[self.audio_column] , __magic_name__ ): raise ValueError(F'''Column {self.audio_column} is not an Audio type.''' ) lowerCamelCase : Optional[Any] = copy.deepcopy(self ) lowerCamelCase : List[Any] = self.input_schema.copy() lowerCamelCase : Tuple = features[self.audio_column] lowerCamelCase : int = input_schema return task_template @property def UpperCamelCase__ ( self ): return {self.audio_column: "audio", self.transcription_column: "transcription"}
287
1
import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : Dict = ["""a""", """b""", """c"""] # Defaults to last layer if both are None lowerCamelCase , lowerCamelCase : List[Any] = get_aligned_output_features_output_indices(__magic_name__ , __magic_name__ , __magic_name__ ) self.assertEqual(__magic_name__ , ["""c"""] ) self.assertEqual(__magic_name__ , [2] ) # Out indices set to match out features lowerCamelCase , lowerCamelCase : Tuple = get_aligned_output_features_output_indices(["""a""", """c"""] , __magic_name__ , __magic_name__ ) self.assertEqual(__magic_name__ , ["""a""", """c"""] ) self.assertEqual(__magic_name__ , [0, 2] ) # Out features set to match out indices lowerCamelCase , lowerCamelCase : Optional[int] = get_aligned_output_features_output_indices(__magic_name__ , [0, 2] , __magic_name__ ) self.assertEqual(__magic_name__ , ["""a""", """c"""] ) self.assertEqual(__magic_name__ , [0, 2] ) # Out features selected from negative indices lowerCamelCase , lowerCamelCase : Union[str, Any] = get_aligned_output_features_output_indices(__magic_name__ , [-3, -1] , __magic_name__ ) self.assertEqual(__magic_name__ , ["""a""", """c"""] ) self.assertEqual(__magic_name__ , [-3, -1] ) def UpperCamelCase__ ( self ): # Stage names must be set with self.assertRaises(__magic_name__ ): verify_out_features_out_indices(["""a""", """b"""] , (0, 1) , __magic_name__ ) # Out features must be a list with self.assertRaises(__magic_name__ ): verify_out_features_out_indices(("""a""", """b""") , (0, 1) , ["""a""", """b"""] ) # Out features must be a subset of stage names with self.assertRaises(__magic_name__ ): verify_out_features_out_indices(["""a""", """b"""] , (0, 1) , ["""a"""] ) # Out indices must be a list or tuple with self.assertRaises(__magic_name__ ): verify_out_features_out_indices(__magic_name__ , 0 , ["""a""", """b"""] ) # Out indices must be a subset of stage names with self.assertRaises(__magic_name__ ): verify_out_features_out_indices(__magic_name__ , (0, 1) , ["""a"""] ) # Out features and out indices must be the same length with self.assertRaises(__magic_name__ ): verify_out_features_out_indices(["""a""", """b"""] , (0,) , ["""a""", """b""", """c"""] ) # Out features should match out indices with self.assertRaises(__magic_name__ ): verify_out_features_out_indices(["""a""", """b"""] , (0, 2) , ["""a""", """b""", """c"""] ) # Out features and out indices should be in order with self.assertRaises(__magic_name__ ): verify_out_features_out_indices(["""b""", """a"""] , (0, 1) , ["""a""", """b"""] ) # Check passes with valid inputs verify_out_features_out_indices(["""a""", """b""", """d"""] , (0, 1, -1) , ["""a""", """b""", """c""", """d"""] ) def UpperCamelCase__ ( self ): lowerCamelCase : str = BackboneMixin() lowerCamelCase : Optional[int] = ["""a""", """b""", """c"""] lowerCamelCase : Tuple = ["""a""", """c"""] lowerCamelCase : Any = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ["""a""", """c"""] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly lowerCamelCase : Any = ["""a""", """b"""] self.assertEqual(backbone.out_features , ["""a""", """b"""] ) self.assertEqual(backbone.out_indices , [0, 1] ) lowerCamelCase : List[str] = [-3, -1] self.assertEqual(backbone.out_features , ["""a""", """c"""] ) self.assertEqual(backbone.out_indices , [-3, -1] )
287
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower_vision_model""" def __init__( self , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=3 , __magic_name__=1_6 , __magic_name__=2_8_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__=True , __magic_name__=False , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : str = num_hidden_layers lowerCamelCase : Optional[int] = num_channels lowerCamelCase : List[str] = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Any = initializer_factor lowerCamelCase : Tuple = layer_norm_eps lowerCamelCase : Tuple = stop_gradient lowerCamelCase : Optional[int] = share_layernorm lowerCamelCase : str = remove_last_layer @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : str = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """bridgetower_text_model""" def __init__( self , __magic_name__=5_0_2_6_5 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=1 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_4 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Any = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : Tuple = hidden_act lowerCamelCase : Optional[int] = initializer_factor lowerCamelCase : Any = intermediate_size lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : str = max_position_embeddings lowerCamelCase : Union[str, Any] = type_vocab_size lowerCamelCase : Optional[int] = layer_norm_eps lowerCamelCase : Optional[int] = position_embedding_type lowerCamelCase : List[str] = use_cache lowerCamelCase : List[str] = pad_token_id lowerCamelCase : List[str] = bos_token_id lowerCamelCase : Optional[int] = eos_token_id @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : Optional[int] = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower""" def __init__( self , __magic_name__=True , __magic_name__="gelu" , __magic_name__=7_6_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__="add" , __magic_name__=1_2 , __magic_name__=6 , __magic_name__=False , __magic_name__=False , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): # TODO: remove this once the Hub files are updated. lowerCamelCase : int = kwargs.pop("""text_config_dict""" , __magic_name__ ) lowerCamelCase : str = kwargs.pop("""vision_config_dict""" , __magic_name__ ) super().__init__(**__magic_name__ ) lowerCamelCase : str = share_cross_modal_transformer_layers lowerCamelCase : Union[str, Any] = hidden_act lowerCamelCase : str = hidden_size lowerCamelCase : Tuple = initializer_factor lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = share_link_tower_layers lowerCamelCase : List[Any] = link_tower_type lowerCamelCase : Tuple = num_attention_heads lowerCamelCase : int = num_hidden_layers lowerCamelCase : Union[str, Any] = tie_word_embeddings lowerCamelCase : Tuple = init_layernorm_from_vision_encoder if text_config is None: lowerCamelCase : Any = {} logger.info("""`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.""" ) if vision_config is None: lowerCamelCase : int = {} logger.info("""`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.""" ) lowerCamelCase : Any = BridgeTowerTextConfig(**__magic_name__ ) lowerCamelCase : Optional[Any] = BridgeTowerVisionConfig(**__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , **__magic_name__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = copy.deepcopy(self.__dict__ ) lowerCamelCase : int = self.text_config.to_dict() lowerCamelCase : Dict = self.vision_config.to_dict() lowerCamelCase : List[str] = self.__class__.model_type return output
287
1
from __future__ import annotations import unittest from transformers import BlenderbotSmallConfig, BlenderbotSmallTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel @require_tf class A__ : _UpperCAmelCase : List[Any] = BlenderbotSmallConfig _UpperCAmelCase : Any = {} _UpperCAmelCase : Tuple = """gelu""" def __init__( self , __magic_name__ , __magic_name__=1_3 , __magic_name__=7 , __magic_name__=True , __magic_name__=False , __magic_name__=9_9 , __magic_name__=3_2 , __magic_name__=2 , __magic_name__=4 , __magic_name__=3_7 , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=2_0 , __magic_name__=2 , __magic_name__=1 , __magic_name__=0 , ): lowerCamelCase : List[str] = parent lowerCamelCase : Optional[int] = batch_size lowerCamelCase : str = seq_length lowerCamelCase : int = is_training lowerCamelCase : Optional[Any] = use_labels lowerCamelCase : List[Any] = vocab_size lowerCamelCase : Optional[Any] = hidden_size lowerCamelCase : Union[str, Any] = num_hidden_layers lowerCamelCase : Optional[int] = num_attention_heads lowerCamelCase : Union[str, Any] = intermediate_size lowerCamelCase : int = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : Optional[Any] = max_position_embeddings lowerCamelCase : str = eos_token_id lowerCamelCase : Dict = pad_token_id lowerCamelCase : Tuple = bos_token_id def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowerCamelCase : Union[str, Any] = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowerCamelCase : List[str] = tf.concat([input_ids, eos_tensor] , axis=1 ) lowerCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase : Dict = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowerCamelCase : Any = prepare_blenderbot_small_inputs_dict(__magic_name__ , __magic_name__ , __magic_name__ ) return config, inputs_dict def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): lowerCamelCase : List[Any] = TFBlenderbotSmallModel(config=__magic_name__ ).get_decoder() lowerCamelCase : List[str] = inputs_dict["""input_ids"""] lowerCamelCase : int = input_ids[:1, :] lowerCamelCase : str = inputs_dict["""attention_mask"""][:1, :] lowerCamelCase : Tuple = inputs_dict["""head_mask"""] lowerCamelCase : int = 1 # first forward pass lowerCamelCase : List[str] = model(__magic_name__ , attention_mask=__magic_name__ , head_mask=__magic_name__ , use_cache=__magic_name__ ) lowerCamelCase , lowerCamelCase : int = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase : str = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCamelCase : Union[str, Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and lowerCamelCase : int = tf.concat([input_ids, next_tokens] , axis=-1 ) lowerCamelCase : Dict = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) lowerCamelCase : int = model(__magic_name__ , attention_mask=__magic_name__ )[0] lowerCamelCase : Optional[int] = model(__magic_name__ , attention_mask=__magic_name__ , past_key_values=__magic_name__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice lowerCamelCase : int = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) lowerCamelCase : Optional[int] = output_from_no_past[:, -3:, random_slice_idx] lowerCamelCase : Optional[int] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__magic_name__ , __magic_name__ , rtol=1e-3 ) def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=None, ): if attention_mask is None: lowerCamelCase : Union[str, Any] = tf.cast(tf.math.not_equal(lowerCamelCase, config.pad_token_id ), tf.inta ) if decoder_attention_mask is None: lowerCamelCase : Any = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape, dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:], config.pad_token_id ), tf.inta ), ], axis=-1, ) if head_mask is None: lowerCamelCase : List[Any] = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCamelCase : Optional[Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowerCamelCase : str = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : str = ( (TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel) if is_tf_available() else () ) _UpperCAmelCase : Any = (TFBlenderbotSmallForConditionalGeneration,) if is_tf_available() else () _UpperCAmelCase : int = ( { """conversational""": TFBlenderbotSmallForConditionalGeneration, """feature-extraction""": TFBlenderbotSmallModel, """summarization""": TFBlenderbotSmallForConditionalGeneration, """text2text-generation""": TFBlenderbotSmallForConditionalGeneration, """translation""": TFBlenderbotSmallForConditionalGeneration, } if is_tf_available() else {} ) _UpperCAmelCase : List[Any] = True _UpperCAmelCase : str = False _UpperCAmelCase : Any = False def UpperCamelCase__ ( self ): lowerCamelCase : int = TFBlenderbotSmallModelTester(self ) lowerCamelCase : List[str] = ConfigTester(self , config_class=__magic_name__ ) def UpperCamelCase__ ( self ): self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__magic_name__ ) @require_tokenizers @require_tf class A__ ( unittest.TestCase): _UpperCAmelCase : int = [ """Social anxiety\nWow, I am never shy. Do you have anxiety?\nYes. I end up sweating and blushing and feel like """ """ i'm going to throw up.\nand why is that?""" ] _UpperCAmelCase : Tuple = """facebook/blenderbot_small-90M""" @cached_property def UpperCamelCase__ ( self ): # use "old" tokenizer here because of bug when downloading new tokenizer return BlenderbotSmallTokenizer.from_pretrained("""facebook/blenderbot-90M""" ) @cached_property def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = self.tokenizer(self.src_text , return_tensors="""tf""" ) lowerCamelCase : Dict = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=__magic_name__ , ) lowerCamelCase : Any = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=__magic_name__ )[0] assert generated_words in ( "i don't know. i just feel like i'm going to throw up. it's not fun.", "i'm not sure. i just feel like i've been feeling like i have to be in a certain place", "i'm not sure. i just feel like i've been in a bad situation.", )
287
def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = set(range(3, lowerCamelCase, 2 ) ) primes.add(2 ) for p in range(3, lowerCamelCase, 2 ): if p not in primes: continue primes.difference_update(set(range(p * p, lowerCamelCase, lowerCamelCase ) ) ) lowerCamelCase : Any = [float(lowerCamelCase ) for n in range(limit + 1 )] for p in primes: for n in range(lowerCamelCase, limit + 1, lowerCamelCase ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'''{solution() = }''')
287
1
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _lowerCamelCase =logging.get_logger(__name__) # pylint: disable=invalid-name class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): @register_to_config def __init__( self , __magic_name__ , __magic_name__ = None , __magic_name__ = None ): super().__init__() lowerCamelCase : int = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" lowerCamelCase : List[Any] = torch.zeros(__magic_name__ , __magic_name__ ) else: lowerCamelCase : List[Any] = None lowerCamelCase : List[Any] = torch.nn.Parameter(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : VQModel _UpperCAmelCase : CLIPTextModel _UpperCAmelCase : CLIPTokenizer _UpperCAmelCase : TransformeraDModel _UpperCAmelCase : LearnedClassifierFreeSamplingEmbeddings _UpperCAmelCase : VQDiffusionScheduler def __init__( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ): super().__init__() self.register_modules( vqvae=__magic_name__ , transformer=__magic_name__ , text_encoder=__magic_name__ , tokenizer=__magic_name__ , scheduler=__magic_name__ , learned_classifier_free_sampling_embeddings=__magic_name__ , ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : int = len(__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else 1 # get prompt text embeddings lowerCamelCase : Tuple = self.tokenizer( __magic_name__ , padding="""max_length""" , max_length=self.tokenizer.model_max_length , return_tensors="""pt""" , ) lowerCamelCase : Dict = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: lowerCamelCase : Union[str, Any] = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( """The following part of your input was truncated because CLIP can only handle sequences up to""" F''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) lowerCamelCase : Union[str, Any] = text_input_ids[:, : self.tokenizer.model_max_length] lowerCamelCase : Any = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 lowerCamelCase : Tuple = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__magic_name__ ) # duplicate text embeddings for each generation per prompt lowerCamelCase : Optional[Any] = prompt_embeds.repeat_interleave(__magic_name__ , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: lowerCamelCase : Any = self.learned_classifier_free_sampling_embeddings.embeddings lowerCamelCase : List[str] = negative_prompt_embeds.unsqueeze(0 ).repeat(__magic_name__ , 1 , 1 ) else: lowerCamelCase : Optional[int] = [""""""] * batch_size lowerCamelCase : int = text_input_ids.shape[-1] lowerCamelCase : Tuple = self.tokenizer( __magic_name__ , padding="""max_length""" , max_length=__magic_name__ , truncation=__magic_name__ , return_tensors="""pt""" , ) lowerCamelCase : Optional[int] = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings lowerCamelCase : Dict = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__magic_name__ ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method lowerCamelCase : Dict = negative_prompt_embeds.shape[1] lowerCamelCase : int = negative_prompt_embeds.repeat(1 , __magic_name__ , 1 ) lowerCamelCase : Tuple = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __magic_name__ , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowerCamelCase : List[str] = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , __magic_name__ , __magic_name__ = 1_0_0 , __magic_name__ = 5.0 , __magic_name__ = 1.0 , __magic_name__ = 1 , __magic_name__ = None , __magic_name__ = None , __magic_name__ = "pil" , __magic_name__ = True , __magic_name__ = None , __magic_name__ = 1 , ): if isinstance(__magic_name__ , __magic_name__ ): lowerCamelCase : List[str] = 1 elif isinstance(__magic_name__ , __magic_name__ ): lowerCamelCase : Dict = len(__magic_name__ ) else: raise ValueError(F'''`prompt` has to be of type `str` or `list` but is {type(__magic_name__ )}''' ) lowerCamelCase : Tuple = batch_size * num_images_per_prompt lowerCamelCase : Dict = guidance_scale > 1.0 lowerCamelCase : Tuple = self._encode_prompt(__magic_name__ , __magic_name__ , __magic_name__ ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__magic_name__ , __magic_name__ ) or callback_steps <= 0) ): raise ValueError( F'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' F''' {type(__magic_name__ )}.''' ) # get the initial completely masked latents unless the user supplied it lowerCamelCase : List[Any] = (batch_size, self.transformer.num_latent_pixels) if latents is None: lowerCamelCase : int = self.transformer.num_vector_embeds - 1 lowerCamelCase : Any = torch.full(__magic_name__ , __magic_name__ ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( """Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,""" F''' {self.transformer.num_vector_embeds - 1} (inclusive).''' ) lowerCamelCase : Union[str, Any] = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__magic_name__ , device=self.device ) lowerCamelCase : Tuple = self.scheduler.timesteps.to(self.device ) lowerCamelCase : Any = latents for i, t in enumerate(self.progress_bar(__magic_name__ ) ): # expand the sample if we are doing classifier free guidance lowerCamelCase : Dict = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` lowerCamelCase : Tuple = self.transformer(__magic_name__ , encoder_hidden_states=__magic_name__ , timestep=__magic_name__ ).sample if do_classifier_free_guidance: lowerCamelCase , lowerCamelCase : Optional[int] = model_output.chunk(2 ) lowerCamelCase : Optional[int] = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__magic_name__ , dim=1 , keepdim=__magic_name__ ) lowerCamelCase : List[str] = self.truncate(__magic_name__ , __magic_name__ ) # remove `log(0)`'s (`-inf`s) lowerCamelCase : List[Any] = model_output.clamp(-7_0 ) # compute the previous noisy sample x_t -> x_t-1 lowerCamelCase : Dict = self.scheduler.step(__magic_name__ , timestep=__magic_name__ , sample=__magic_name__ , generator=__magic_name__ ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__magic_name__ , __magic_name__ , __magic_name__ ) lowerCamelCase : int = self.vqvae.config.vq_embed_dim lowerCamelCase : Optional[Any] = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) lowerCamelCase : Optional[int] = self.vqvae.quantize.get_codebook_entry(__magic_name__ , shape=__magic_name__ ) lowerCamelCase : Any = self.vqvae.decode(__magic_name__ , force_not_quantize=__magic_name__ ).sample lowerCamelCase : Any = (image / 2 + 0.5).clamp(0 , 1 ) lowerCamelCase : int = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCamelCase : List[Any] = self.numpy_to_pil(__magic_name__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): lowerCamelCase , lowerCamelCase : Optional[Any] = torch.sort(__magic_name__ , 1 , descending=__magic_name__ ) lowerCamelCase : List[Any] = torch.exp(__magic_name__ ) lowerCamelCase : Dict = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out lowerCamelCase : Optional[Any] = torch.full_like(keep_mask[:, 0:1, :] , __magic_name__ ) lowerCamelCase : Union[str, Any] = torch.cat((all_true, keep_mask) , dim=1 ) lowerCamelCase : Optional[int] = keep_mask[:, :-1, :] lowerCamelCase : str = keep_mask.gather(1 , indices.argsort(1 ) ) lowerCamelCase : Optional[Any] = log_p_x_0.clone() lowerCamelCase : str = -torch.inf # -inf = log(0) return rv
287
from __future__ import annotations import collections import pprint from pathlib import Path def _a ( lowerCamelCase ): return "".join(sorted(lowerCamelCase ) ) def _a ( lowerCamelCase ): return word_by_signature[signature(lowerCamelCase )] _lowerCamelCase =Path(__file__).parent.joinpath("""words.txt""").read_text(encoding="""utf-8""") _lowerCamelCase =sorted({word.strip().lower() for word in data.splitlines()}) _lowerCamelCase =collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _lowerCamelCase ={word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open("""anagrams.txt""", """w""") as file: file.write("""all_anagrams = \n """) file.write(pprint.pformat(all_anagrams))
287
1
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _lowerCamelCase ="""src/diffusers""" _lowerCamelCase =""".""" # This is to make sure the diffusers module imported is the one in the repo. _lowerCamelCase =importlib.util.spec_from_file_location( """diffusers""", os.path.join(DIFFUSERS_PATH, """__init__.py"""), submodule_search_locations=[DIFFUSERS_PATH], ) _lowerCamelCase =spec.loader.load_module() def _a ( lowerCamelCase, lowerCamelCase ): return line.startswith(lowerCamelCase ) or len(lowerCamelCase ) <= 1 or re.search(R"""^\s*\)(\s*->.*:|:)\s*$""", lowerCamelCase ) is not None def _a ( lowerCamelCase ): lowerCamelCase : List[Any] = object_name.split(""".""" ) lowerCamelCase : Optional[Any] = 0 # First let's find the module where our object lives. lowerCamelCase : Tuple = parts[i] while i < len(lowerCamelCase ) and not os.path.isfile(os.path.join(lowerCamelCase, F'''{module}.py''' ) ): i += 1 if i < len(lowerCamelCase ): lowerCamelCase : str = os.path.join(lowerCamelCase, parts[i] ) if i >= len(lowerCamelCase ): raise ValueError(F'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''' ) with open(os.path.join(lowerCamelCase, F'''{module}.py''' ), """r""", encoding="""utf-8""", newline="""\n""" ) as f: lowerCamelCase : str = f.readlines() # Now let's find the class / func in the code! lowerCamelCase : Dict = """""" lowerCamelCase : Optional[Any] = 0 for name in parts[i + 1 :]: while ( line_index < len(lowerCamelCase ) and re.search(RF'''^{indent}(class|def)\s+{name}(\(|\:)''', lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(lowerCamelCase ): raise ValueError(F''' {object_name} does not match any function or class in {module}.''' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowerCamelCase : Dict = line_index while line_index < len(lowerCamelCase ) and _should_continue(lines[line_index], lowerCamelCase ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowerCamelCase : Union[str, Any] = lines[start_index:line_index] return "".join(lowerCamelCase ) _lowerCamelCase =re.compile(R"""^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)""") _lowerCamelCase =re.compile(R"""^\s*(\S+)->(\S+)(\s+.*|$)""") _lowerCamelCase =re.compile(R"""<FILL\s+[^>]*>""") def _a ( lowerCamelCase ): lowerCamelCase : int = code.split("""\n""" ) lowerCamelCase : Any = 0 while idx < len(lowerCamelCase ) and len(lines[idx] ) == 0: idx += 1 if idx < len(lowerCamelCase ): return re.search(R"""^(\s*)\S""", lines[idx] ).groups()[0] return "" def _a ( lowerCamelCase ): lowerCamelCase : Any = len(get_indent(lowerCamelCase ) ) > 0 if has_indent: lowerCamelCase : Any = F'''class Bla:\n{code}''' lowerCamelCase : Any = black.Mode(target_versions={black.TargetVersion.PYaa}, line_length=119, preview=lowerCamelCase ) lowerCamelCase : Dict = black.format_str(lowerCamelCase, mode=lowerCamelCase ) lowerCamelCase , lowerCamelCase : Union[str, Any] = style_docstrings_in_code(lowerCamelCase ) return result[len("""class Bla:\n""" ) :] if has_indent else result def _a ( lowerCamelCase, lowerCamelCase=False ): with open(lowerCamelCase, """r""", encoding="""utf-8""", newline="""\n""" ) as f: lowerCamelCase : List[Any] = f.readlines() lowerCamelCase : str = [] lowerCamelCase : int = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(lowerCamelCase ): lowerCamelCase : str = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowerCamelCase , lowerCamelCase , lowerCamelCase : Union[str, Any] = search.groups() lowerCamelCase : Optional[int] = find_code_in_diffusers(lowerCamelCase ) lowerCamelCase : Optional[int] = get_indent(lowerCamelCase ) lowerCamelCase : Optional[Any] = line_index + 1 if indent == theoretical_indent else line_index + 2 lowerCamelCase : Optional[Any] = theoretical_indent lowerCamelCase : Dict = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowerCamelCase : Tuple = True while line_index < len(lowerCamelCase ) and should_continue: line_index += 1 if line_index >= len(lowerCamelCase ): break lowerCamelCase : Union[str, Any] = lines[line_index] lowerCamelCase : str = _should_continue(lowerCamelCase, lowerCamelCase ) and re.search(F'''^{indent}# End copy''', lowerCamelCase ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowerCamelCase : Union[str, Any] = lines[start_index:line_index] lowerCamelCase : Tuple = """""".join(lowerCamelCase ) # Remove any nested `Copied from` comments to avoid circular copies lowerCamelCase : Any = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(lowerCamelCase ) is None] lowerCamelCase : Optional[Any] = """\n""".join(lowerCamelCase ) # Before comparing, use the `replace_pattern` on the original code. if len(lowerCamelCase ) > 0: lowerCamelCase : Optional[int] = replace_pattern.replace("""with""", """""" ).split(""",""" ) lowerCamelCase : List[str] = [_re_replace_pattern.search(lowerCamelCase ) for p in patterns] for pattern in patterns: if pattern is None: continue lowerCamelCase , lowerCamelCase , lowerCamelCase : Any = pattern.groups() lowerCamelCase : Optional[Any] = re.sub(lowerCamelCase, lowerCamelCase, lowerCamelCase ) if option.strip() == "all-casing": lowerCamelCase : Optional[Any] = re.sub(obja.lower(), obja.lower(), lowerCamelCase ) lowerCamelCase : Dict = re.sub(obja.upper(), obja.upper(), lowerCamelCase ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowerCamelCase : Any = blackify(lines[start_index - 1] + theoretical_code ) lowerCamelCase : Optional[int] = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: lowerCamelCase : Optional[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] lowerCamelCase : Tuple = start_index + 1 if overwrite and len(lowerCamelCase ) > 0: # Warn the user a file has been modified. print(F'''Detected changes, rewriting {filename}.''' ) with open(lowerCamelCase, """w""", encoding="""utf-8""", newline="""\n""" ) as f: f.writelines(lowerCamelCase ) return diffs def _a ( lowerCamelCase = False ): lowerCamelCase : Dict = glob.glob(os.path.join(lowerCamelCase, """**/*.py""" ), recursive=lowerCamelCase ) lowerCamelCase : Tuple = [] for filename in all_files: lowerCamelCase : List[Any] = is_copy_consistent(lowerCamelCase, lowerCamelCase ) diffs += [F'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(lowerCamelCase ) > 0: lowerCamelCase : Dict = """\n""".join(lowerCamelCase ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": _lowerCamelCase =argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _lowerCamelCase =parser.parse_args() check_copies(args.fix_and_overwrite)
287
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = ["""pixel_values"""] def __init__( self , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , __magic_name__ = PILImageResampling.BILINEAR , __magic_name__ = True , __magic_name__ = 1 / 2_5_5 , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = size if size is not None else {"""shortest_edge""": 3_8_4} lowerCamelCase : Tuple = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Dict = do_resize lowerCamelCase : List[Any] = size # Default value set here for backwards compatibility where the value in config is None lowerCamelCase : Any = crop_pct if crop_pct is not None else 2_2_4 / 2_5_6 lowerCamelCase : Union[str, Any] = resample lowerCamelCase : str = do_rescale lowerCamelCase : Union[str, Any] = rescale_factor lowerCamelCase : Tuple = do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCamelCase : int = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = PILImageResampling.BICUBIC , __magic_name__ = None , **__magic_name__ , ): lowerCamelCase : Union[str, Any] = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) lowerCamelCase : str = size["""shortest_edge"""] if shortest_edge < 3_8_4: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct lowerCamelCase : List[str] = int(shortest_edge / crop_pct ) lowerCamelCase : Optional[Any] = get_resize_output_image_size(__magic_name__ , size=__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Optional[int] = resize(image=__magic_name__ , size=__magic_name__ , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=__magic_name__ , size=(shortest_edge, shortest_edge) , data_format=__magic_name__ , **__magic_name__ ) else: # warping (no cropping) when evaluated at 384 or larger return resize( __magic_name__ , size=(shortest_edge, shortest_edge) , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return rescale(__magic_name__ , scale=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return normalize(__magic_name__ , mean=__magic_name__ , std=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = ChannelDimension.FIRST , **__magic_name__ , ): lowerCamelCase : str = do_resize if do_resize is not None else self.do_resize lowerCamelCase : Optional[Any] = crop_pct if crop_pct is not None else self.crop_pct lowerCamelCase : Optional[int] = resample if resample is not None else self.resample lowerCamelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else self.image_mean lowerCamelCase : Tuple = image_std if image_std is not None else self.image_std lowerCamelCase : Dict = size if size is not None else self.size lowerCamelCase : str = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : List[str] = make_list_of_images(__magic_name__ ) if not valid_images(__magic_name__ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_resize and size["shortest_edge"] < 3_8_4 and crop_pct is None: raise ValueError("""crop_pct must be specified if size < 384.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. lowerCamelCase : Optional[Any] = [to_numpy_array(__magic_name__ ) for image in images] if do_resize: lowerCamelCase : List[Any] = [self.resize(image=__magic_name__ , size=__magic_name__ , crop_pct=__magic_name__ , resample=__magic_name__ ) for image in images] if do_rescale: lowerCamelCase : Union[str, Any] = [self.rescale(image=__magic_name__ , scale=__magic_name__ ) for image in images] if do_normalize: lowerCamelCase : List[Any] = [self.normalize(image=__magic_name__ , mean=__magic_name__ , std=__magic_name__ ) for image in images] lowerCamelCase : Optional[int] = [to_channel_dimension_format(__magic_name__ , __magic_name__ ) for image in images] lowerCamelCase : List[str] = {"""pixel_values""": images} return BatchFeature(data=__magic_name__ , tensor_type=__magic_name__ )
287
1
from __future__ import annotations import unittest from transformers import RoFormerConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerModel, ) from transformers.models.roformer.modeling_tf_roformer import ( TFRoFormerSelfAttention, TFRoFormerSinusoidalPositionalEmbedding, ) class A__ : def __init__( self , __magic_name__ , __magic_name__=1_3 , __magic_name__=7 , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=9_9 , __magic_name__=3_2 , __magic_name__=2 , __magic_name__=4 , __magic_name__=3_7 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=1_6 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=3 , __magic_name__=4 , __magic_name__=None , ): lowerCamelCase : Union[str, Any] = parent lowerCamelCase : List[str] = 1_3 lowerCamelCase : Union[str, Any] = 7 lowerCamelCase : int = True lowerCamelCase : Optional[int] = True lowerCamelCase : Tuple = True lowerCamelCase : List[Any] = True lowerCamelCase : Union[str, Any] = 9_9 lowerCamelCase : List[str] = 3_2 lowerCamelCase : Optional[Any] = 2 lowerCamelCase : Union[str, Any] = 4 lowerCamelCase : str = 3_7 lowerCamelCase : Union[str, Any] = """gelu""" lowerCamelCase : str = 0.1 lowerCamelCase : List[str] = 0.1 lowerCamelCase : str = 5_1_2 lowerCamelCase : int = 1_6 lowerCamelCase : List[str] = 2 lowerCamelCase : Optional[Any] = 0.02 lowerCamelCase : Optional[Any] = 3 lowerCamelCase : List[Any] = 4 lowerCamelCase : Dict = None def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase : Optional[int] = None if self.use_input_mask: lowerCamelCase : Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase : Optional[Any] = None if self.use_token_type_ids: lowerCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase : Optional[Any] = None lowerCamelCase : Any = None lowerCamelCase : int = None if self.use_labels: lowerCamelCase : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase : Tuple = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase : Dict = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=__magic_name__ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Union[str, Any] = TFRoFormerModel(config=__magic_name__ ) lowerCamelCase : Optional[int] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCamelCase : int = [input_ids, input_mask] lowerCamelCase : Any = model(__magic_name__ ) lowerCamelCase : Tuple = model(__magic_name__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : str = True lowerCamelCase : Tuple = TFRoFormerForCausalLM(config=__magic_name__ ) lowerCamelCase : str = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCamelCase : Tuple = model(__magic_name__ )["""logits"""] self.parent.assertListEqual( list(prediction_scores.numpy().shape ) , [self.batch_size, self.seq_length, self.vocab_size] ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Optional[int] = TFRoFormerForMaskedLM(config=__magic_name__ ) lowerCamelCase : Dict = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCamelCase : Dict = model(__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Dict = self.num_labels lowerCamelCase : Optional[int] = TFRoFormerForSequenceClassification(config=__magic_name__ ) lowerCamelCase : str = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCamelCase : Dict = model(__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Any = self.num_choices lowerCamelCase : str = TFRoFormerForMultipleChoice(config=__magic_name__ ) lowerCamelCase : List[str] = tf.tile(tf.expand_dims(__magic_name__ , 1 ) , (1, self.num_choices, 1) ) lowerCamelCase : Dict = tf.tile(tf.expand_dims(__magic_name__ , 1 ) , (1, self.num_choices, 1) ) lowerCamelCase : int = tf.tile(tf.expand_dims(__magic_name__ , 1 ) , (1, self.num_choices, 1) ) lowerCamelCase : Tuple = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } lowerCamelCase : Optional[Any] = model(__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : List[Any] = self.num_labels lowerCamelCase : List[str] = TFRoFormerForTokenClassification(config=__magic_name__ ) lowerCamelCase : Optional[Any] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCamelCase : int = model(__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : str = TFRoFormerForQuestionAnswering(config=__magic_name__ ) lowerCamelCase : Optional[int] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCamelCase : Any = model(__magic_name__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = self.prepare_config_and_inputs() ( ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ) : Optional[int] = config_and_inputs lowerCamelCase : str = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : List[Any] = ( ( TFRoFormerModel, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerForMultipleChoice, ) if is_tf_available() else () ) _UpperCAmelCase : str = ( { """feature-extraction""": TFRoFormerModel, """fill-mask""": TFRoFormerForMaskedLM, """question-answering""": TFRoFormerForQuestionAnswering, """text-classification""": TFRoFormerForSequenceClassification, """text-generation""": TFRoFormerForCausalLM, """token-classification""": TFRoFormerForTokenClassification, """zero-shot""": TFRoFormerForSequenceClassification, } if is_tf_available() else {} ) _UpperCAmelCase : Tuple = False _UpperCAmelCase : str = False def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): if pipeline_test_casse_name == "TextGenerationPipelineTests": return True return False def UpperCamelCase__ ( self ): lowerCamelCase : str = TFRoFormerModelTester(self ) lowerCamelCase : Optional[int] = ConfigTester(self , config_class=__magic_name__ , hidden_size=3_7 ) def UpperCamelCase__ ( self ): self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__magic_name__ ) @slow def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = TFRoFormerModel.from_pretrained("""junnyu/roformer_chinese_base""" ) self.assertIsNotNone(__magic_name__ ) @require_tf class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : Any = TFRoFormerForMaskedLM.from_pretrained("""junnyu/roformer_chinese_base""" ) lowerCamelCase : Optional[int] = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowerCamelCase : Optional[Any] = model(__magic_name__ )[0] # TODO Replace vocab size lowerCamelCase : Optional[Any] = 5_0_0_0_0 lowerCamelCase : Union[str, Any] = [1, 6, vocab_size] self.assertEqual(output.shape , __magic_name__ ) print(output[:, :3, :3] ) # TODO Replace values below with what was printed above. lowerCamelCase : Union[str, Any] = tf.constant( [ [ [-0.12_053_341, -1.0_264_901, 0.29_221_946], [-1.5_133_783, 0.197_433, 0.15_190_607], [-5.0_135_403, -3.900_256, -0.84_038_764], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , __magic_name__ , atol=1e-4 ) @require_tf class A__ ( unittest.TestCase): _UpperCAmelCase : List[str] = 1E-4 def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = tf.constant([[4, 1_0]] ) lowerCamelCase : Optional[int] = TFRoFormerSinusoidalPositionalEmbedding(num_positions=6 , embedding_dim=6 ) lowerCamelCase : Optional[Any] = emba(input_ids.shape ) lowerCamelCase : Tuple = tf.constant( [[0.0_000, 0.0_000, 0.0_000, 1.0_000, 1.0_000, 1.0_000], [0.8_415, 0.0_464, 0.0_022, 0.5_403, 0.9_989, 1.0_000]] ) tf.debugging.assert_near(__magic_name__ , __magic_name__ , atol=self.tolerance ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = tf.constant( [ [0.0_000, 0.0_000, 0.0_000, 0.0_000, 0.0_000], [0.8_415, 0.8_219, 0.8_020, 0.7_819, 0.7_617], [0.9_093, 0.9_364, 0.9_581, 0.9_749, 0.9_870], ] ) lowerCamelCase : Dict = TFRoFormerSinusoidalPositionalEmbedding(num_positions=5_1_2 , embedding_dim=5_1_2 ) emba([2, 1_6, 5_1_2] ) lowerCamelCase : List[str] = emba.weight[:3, :5] tf.debugging.assert_near(__magic_name__ , __magic_name__ , atol=self.tolerance ) @require_tf class A__ ( unittest.TestCase): _UpperCAmelCase : Optional[int] = 1E-4 def UpperCamelCase__ ( self ): # 2,12,16,64 lowerCamelCase : Union[str, Any] = tf.reshape(tf.range(2 * 1_2 * 1_6 * 6_4 , dtype=tf.floataa ) , shape=(2, 1_2, 1_6, 6_4) ) / 1_0_0 lowerCamelCase : Dict = -tf.reshape(tf.range(2 * 1_2 * 1_6 * 6_4 , dtype=tf.floataa ) , shape=(2, 1_2, 1_6, 6_4) ) / 1_0_0 lowerCamelCase : Union[str, Any] = TFRoFormerSinusoidalPositionalEmbedding(num_positions=3_2 , embedding_dim=6_4 ) lowerCamelCase : List[str] = embed_positions([2, 1_6, 7_6_8] )[None, None, :, :] lowerCamelCase , lowerCamelCase : Any = TFRoFormerSelfAttention.apply_rotary_position_embeddings( __magic_name__ , __magic_name__ , __magic_name__ ) lowerCamelCase : Union[str, Any] = tf.constant( [ [0.0_000, 0.0_100, 0.0_200, 0.0_300, 0.0_400, 0.0_500, 0.0_600, 0.0_700], [-0.2_012, 0.8_897, 0.0_263, 0.9_401, 0.2_074, 0.9_463, 0.3_481, 0.9_343], [-1.7_057, 0.6_271, -1.2_145, 1.3_897, -0.6_303, 1.7_647, -0.1_173, 1.8_985], [-2.1_731, -1.6_397, -2.7_358, 0.2_854, -2.1_840, 1.7_183, -1.3_018, 2.4_871], [0.2_717, -3.6_173, -2.9_206, -2.1_988, -3.6_638, 0.3_858, -2.9_155, 2.2_980], [3.9_859, -2.1_580, -0.7_984, -4.4_904, -4.1_181, -2.0_252, -4.4_782, 1.1_253], ] ) lowerCamelCase : str = tf.constant( [ [0.0_000, -0.0_100, -0.0_200, -0.0_300, -0.0_400, -0.0_500, -0.0_600, -0.0_700], [0.2_012, -0.8_897, -0.0_263, -0.9_401, -0.2_074, -0.9_463, -0.3_481, -0.9_343], [1.7_057, -0.6_271, 1.2_145, -1.3_897, 0.6_303, -1.7_647, 0.1_173, -1.8_985], [2.1_731, 1.6_397, 2.7_358, -0.2_854, 2.1_840, -1.7_183, 1.3_018, -2.4_871], [-0.2_717, 3.6_173, 2.9_206, 2.1_988, 3.6_638, -0.3_858, 2.9_155, -2.2_980], [-3.9_859, 2.1_580, 0.7_984, 4.4_904, 4.1_181, 2.0_252, 4.4_782, -1.1_253], ] ) tf.debugging.assert_near(query_layer[0, 0, :6, :8] , __magic_name__ , atol=self.tolerance ) tf.debugging.assert_near(key_layer[0, 0, :6, :8] , __magic_name__ , atol=self.tolerance )
287
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 _lowerCamelCase ={ """return_dict""": False, """output_hidden_states""": True, """output_attentions""": True, """torchscript""": True, """torch_dtype""": """float16""", """use_bfloat16""": True, """tf_legacy_loss""": True, """pruned_heads""": {"""a""": 1}, """tie_word_embeddings""": False, """is_decoder""": True, """cross_attention_hidden_size""": 1_2_8, """add_cross_attention""": True, """tie_encoder_decoder""": True, """max_length""": 5_0, """min_length""": 3, """do_sample""": True, """early_stopping""": True, """num_beams""": 3, """num_beam_groups""": 3, """diversity_penalty""": 0.5, """temperature""": 2.0, """top_k""": 1_0, """top_p""": 0.7, """typical_p""": 0.2, """repetition_penalty""": 0.8, """length_penalty""": 0.8, """no_repeat_ngram_size""": 5, """encoder_no_repeat_ngram_size""": 5, """bad_words_ids""": [1, 2, 3], """num_return_sequences""": 3, """chunk_size_feed_forward""": 5, """output_scores""": True, """return_dict_in_generate""": True, """forced_bos_token_id""": 2, """forced_eos_token_id""": 3, """remove_invalid_values""": True, """architectures""": ["""BertModel"""], """finetuning_task""": """translation""", """id2label""": {0: """label"""}, """label2id""": {"""label""": """0"""}, """tokenizer_class""": """BertTokenizerFast""", """prefix""": """prefix""", """bos_token_id""": 6, """pad_token_id""": 7, """eos_token_id""": 8, """sep_token_id""": 9, """decoder_start_token_id""": 1_0, """exponential_decay_length_penalty""": (5, 1.01), """suppress_tokens""": [0, 1], """begin_suppress_tokens""": 2, """task_specific_params""": {"""translation""": """some_params"""}, """problem_type""": """regression""", } @is_staging_test class A__ ( unittest.TestCase): @classmethod def UpperCamelCase__ ( cls ): lowerCamelCase : int = TOKEN HfFolder.save_token(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls ): try: delete_repo(token=cls._token , repo_id="""test-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-config-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-config""" ) except HTTPError: pass def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""test-config""" , use_auth_token=self._token ) lowerCamelCase : Any = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__magic_name__ , repo_id="""test-config""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : Optional[Any] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""valid_org/test-config-org""" , use_auth_token=self._token ) lowerCamelCase : Optional[int] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( __magic_name__ , repo_id="""valid_org/test-config-org""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : List[str] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): CustomConfig.register_for_auto_class() lowerCamelCase : Optional[Any] = CustomConfig(attribute=4_2 ) config.push_to_hub("""test-dynamic-config""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"""AutoConfig""": """custom_configuration.CustomConfig"""} ) lowerCamelCase : List[str] = AutoConfig.from_pretrained(F'''{USER}/test-dynamic-config''' , trust_remote_code=__magic_name__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , """CustomConfig""" ) self.assertEqual(new_config.attribute , 4_2 ) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : str = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated lowerCamelCase : Optional[int] = c.n_embd + 1 # int lowerCamelCase : Optional[int] = c.resid_pdrop + 1.0 # float lowerCamelCase : Tuple = not c.scale_attn_weights # bool lowerCamelCase : Any = c.summary_type + """foo""" # str c.update_from_string( F'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(__magic_name__ , c.n_embd , """mismatch for key: n_embd""" ) self.assertEqual(__magic_name__ , c.resid_pdrop , """mismatch for key: resid_pdrop""" ) self.assertEqual(__magic_name__ , c.scale_attn_weights , """mismatch for key: scale_attn_weights""" ) self.assertEqual(__magic_name__ , c.summary_type , """mismatch for key: summary_type""" ) def UpperCamelCase__ ( self ): lowerCamelCase : str = PretrainedConfig() lowerCamelCase : int = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( __magic_name__ , ["""is_encoder_decoder""", """_name_or_path""", """_commit_hash""", """transformers_version"""] ) lowerCamelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(__magic_name__ , __magic_name__ )] if len(__magic_name__ ) > 0: raise ValueError( """The following keys are set with the default values in""" """ `test_configuration_common.config_common_kwargs` pick another value for them:""" F''' {", ".join(__magic_name__ )}.''' ) def UpperCamelCase__ ( self ): with self.assertRaises(__magic_name__ ): # config is in subfolder, the following should not work without specifying the subfolder lowerCamelCase : Dict = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" ) lowerCamelCase : str = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" , subfolder="""bert""" ) self.assertIsNotNone(__magic_name__ ) def UpperCamelCase__ ( self ): # A mock response for an HTTP head request to emulate server down lowerCamelCase : Dict = mock.Mock() lowerCamelCase : Optional[int] = 5_0_0 lowerCamelCase : List[Any] = {} lowerCamelCase : Tuple = HTTPError lowerCamelCase : Union[str, Any] = {} # Download this model to make sure it's in the cache. lowerCamelCase : List[str] = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=__magic_name__ ) as mock_head: lowerCamelCase : Any = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase__ ( self ): # This test is for deprecated behavior and can be removed in v5 lowerCamelCase : List[str] = BertConfig.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = AutoConfig.from_pretrained("""bert-base-cased""" ) lowerCamelCase : Optional[Any] = ["""config.4.0.0.json"""] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(__magic_name__ ) lowerCamelCase : str = 2 json.dump(configuration.to_dict() , open(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , """w""" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 lowerCamelCase : Any = ["""config.42.0.0.json"""] lowerCamelCase : Optional[Any] = 7_6_8 configuration.save_pretrained(__magic_name__ ) shutil.move(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , os.path.join(__magic_name__ , """config.42.0.0.json""" ) ) lowerCamelCase : int = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 7_6_8 ) def UpperCamelCase__ ( self ): # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. lowerCamelCase : str = """hf-internal-testing/test-two-configs""" import transformers as new_transformers lowerCamelCase : Tuple = """v4.0.0""" lowerCamelCase , lowerCamelCase : Optional[int] = new_transformers.models.auto.AutoConfig.from_pretrained( __magic_name__ , return_unused_kwargs=__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(__magic_name__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers lowerCamelCase : Tuple = """v3.0.0""" lowerCamelCase : Any = old_transformers.models.auto.AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(old_configuration.hidden_size , 7_6_8 )
287
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowerCamelCase ={ """configuration_owlvit""": [ """OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """OwlViTConfig""", """OwlViTOnnxConfig""", """OwlViTTextConfig""", """OwlViTVisionConfig""", ], """processing_owlvit""": ["""OwlViTProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =["""OwlViTFeatureExtractor"""] _lowerCamelCase =["""OwlViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =[ """OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """OwlViTModel""", """OwlViTPreTrainedModel""", """OwlViTTextModel""", """OwlViTVisionModel""", """OwlViTForObjectDetection""", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _lowerCamelCase =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
287
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) lowerCamelCase : Any = tf.convert_to_tensor( [[5, 1_2_1, 1_1, 6_6_0, 1_6, 7_3_0, 2_5_5_4_3, 1_1_0, 8_3, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowerCamelCase : str = model(__magic_name__ )["""last_hidden_state"""] lowerCamelCase : Union[str, Any] = tf.TensorShape((1, 1_0, 7_6_8) ) self.assertEqual(output.shape , __magic_name__ ) # compare the actual values for a slice. lowerCamelCase : Dict = tf.convert_to_tensor( [[[-0.0_254, 0.0_235, 0.1_027], [0.0_606, -0.1_811, -0.0_418], [-0.1_561, -0.1_127, 0.2_687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
287
1
from collections import defaultdict from math import gcd def _a ( lowerCamelCase = 150_0000 ): lowerCamelCase : defaultdict = defaultdict(lowerCamelCase ) lowerCamelCase : str = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1, lowerCamelCase, 2 ): if gcd(lowerCamelCase, lowerCamelCase ) > 1: continue lowerCamelCase : Dict = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(lowerCamelCase, limit + 1, lowerCamelCase ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(f'''{solution() = }''')
287
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _lowerCamelCase =get_logger(__name__) class A__ : def __init__( self , __magic_name__ = None ): lowerCamelCase : Dict = ( os.path.join(__magic_name__ , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) lowerCamelCase : List[str] = Extractor def UpperCamelCase__ ( self , __magic_name__ ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" lowerCamelCase : int = os.path.abspath(__magic_name__ ) return os.path.join(self.extract_dir , hash_url_to_filename(__magic_name__ ) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): return force_extract or ( not os.path.isfile(__magic_name__ ) and not (os.path.isdir(__magic_name__ ) and os.listdir(__magic_name__ )) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = False ): lowerCamelCase : Union[str, Any] = self.extractor.infer_extractor_format(__magic_name__ ) if not extractor_format: return input_path lowerCamelCase : int = self._get_output_path(__magic_name__ ) if self._do_extract(__magic_name__ , __magic_name__ ): self.extractor.extract(__magic_name__ , __magic_name__ , __magic_name__ ) return output_path class A__ ( __SCREAMING_SNAKE_CASE): @classmethod @abstractmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): ... @staticmethod @abstractmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): ... class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[bytes] = [] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with open(__magic_name__ , """rb""" ) as f: return f.read(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if not magic_number: lowerCamelCase : Optional[Any] = max(len(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) try: lowerCamelCase : Tuple = cls.read_magic_number(__magic_name__ , __magic_name__ ) except OSError: return False return any(magic_number.startswith(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) class A__ ( __SCREAMING_SNAKE_CASE): @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): return tarfile.is_tarfile(__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): def resolved(__magic_name__ ) -> str: return os.path.realpath(os.path.abspath(__magic_name__ ) ) def badpath(__magic_name__ , __magic_name__ ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__magic_name__ , __magic_name__ ) ).startswith(__magic_name__ ) def badlink(__magic_name__ , __magic_name__ ) -> bool: # Links are interpreted relative to the directory containing the link lowerCamelCase : List[str] = resolved(os.path.join(__magic_name__ , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=__magic_name__ ) lowerCamelCase : Optional[Any] = resolved(__magic_name__ ) for finfo in members: if badpath(finfo.name , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Dict = tarfile.open(__magic_name__ ) tar_file.extractall(__magic_name__ , members=TarExtractor.safemembers(__magic_name__ , __magic_name__ ) ) tar_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = [B"""\x1F\x8B"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with gzip.open(__magic_name__ , """rb""" ) as gzip_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[int] = [ B"""PK\x03\x04""", B"""PK\x05\x06""", # empty archive B"""PK\x07\x08""", # spanned archive ] @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if super().is_extractable(__magic_name__ , magic_number=__magic_name__ ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__magic_name__ , """rb""" ) as fp: lowerCamelCase : List[str] = _EndRecData(__magic_name__ ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: lowerCamelCase : List[Any] = fp.read(__magic_name__ ) # CD is where we expect it to be if len(__magic_name__ ) == sizeCentralDir: lowerCamelCase : str = struct.unpack(__magic_name__ , __magic_name__ ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with zipfile.ZipFile(__magic_name__ , """r""" ) as zip_file: zip_file.extractall(__magic_name__ ) zip_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[str] = [B"""\xFD\x37\x7A\x58\x5A\x00"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with lzma.open(__magic_name__ ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""Rar!\x1a\x07\x00""", B"""Rar!\x1a\x07\x01\x00"""] # RAR_ID # RAR5_ID @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Union[str, Any] = rarfile.RarFile(__magic_name__ ) rf.extractall(__magic_name__ ) rf.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = [B"""\x28\xb5\x2F\xFD"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd lowerCamelCase : int = zstd.ZstdDecompressor() with open(__magic_name__ , """rb""" ) as ifh, open(__magic_name__ , """wb""" ) as ofh: dctx.copy_stream(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""\x42\x5A\x68"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with bza.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x37\x7A\xBC\xAF\x27\x1C"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with pyazr.SevenZipFile(__magic_name__ , """r""" ) as archive: archive.extractall(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x04\x22\x4D\x18"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) _UpperCAmelCase : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def UpperCamelCase__ ( cls ): return max( len(__magic_name__ ) for extractor in cls.extractors.values() if issubclass(__magic_name__ , __magic_name__ ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): try: return MagicNumberBaseExtractor.read_magic_number(__magic_name__ , magic_number_length=__magic_name__ ) except OSError: return b"" @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = False ): warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = cls.infer_extractor_format(__magic_name__ ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): # <Added version="2.4.0"/> lowerCamelCase : Dict = cls._get_magic_number_max_length() lowerCamelCase : Optional[Any] = cls._read_magic_number(__magic_name__ , __magic_name__ ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__magic_name__ , magic_number=__magic_name__ ): return extractor_format @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = "deprecated" , ): os.makedirs(os.path.dirname(__magic_name__ ) , exist_ok=__magic_name__ ) # Prevent parallel extractions lowerCamelCase : Tuple = str(Path(__magic_name__ ).with_suffix(""".lock""" ) ) with FileLock(__magic_name__ ): shutil.rmtree(__magic_name__ , ignore_errors=__magic_name__ ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__magic_name__ , __magic_name__ ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = extractor if extractor != """deprecated""" else extractor_format else: lowerCamelCase : Optional[int] = cls.extractors[extractor_format] return extractor.extract(__magic_name__ , __magic_name__ ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=__magic_name__ , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__magic_name__ ): return extractor.extract(__magic_name__ , __magic_name__ )
287
1
from __future__ import annotations from math import gcd def _a ( lowerCamelCase, lowerCamelCase = 2, lowerCamelCase = 1, lowerCamelCase = 3, ): # A value less than 2 can cause an infinite loop in the algorithm. if num < 2: raise ValueError("""The input value cannot be less than 2""" ) # Because of the relationship between ``f(f(x))`` and ``f(x)``, this # algorithm struggles to find factors that are divisible by two. # As a workaround, we specifically check for two and even inputs. # See: https://math.stackexchange.com/a/2856214/165820 if num > 2 and num % 2 == 0: return 2 # Pollard's Rho algorithm requires a function that returns pseudorandom # values between 0 <= X < ``num``. It doesn't need to be random in the # sense that the output value is cryptographically secure or difficult # to calculate, it only needs to be random in the sense that all output # values should be equally likely to appear. # For this reason, Pollard suggested using ``f(x) = (x**2 - 1) % num`` # However, the success of Pollard's algorithm isn't guaranteed and is # determined in part by the initial seed and the chosen random function. # To make retries easier, we will instead use ``f(x) = (x**2 + C) % num`` # where ``C`` is a value that we can modify between each attempt. def rand_fn(lowerCamelCase, lowerCamelCase, lowerCamelCase ) -> int: return (pow(lowerCamelCase, 2 ) + step) % modulus for _ in range(lowerCamelCase ): # These track the position within the cycle detection logic. lowerCamelCase : List[str] = seed lowerCamelCase : List[str] = seed while True: # At each iteration, the tortoise moves one step and the hare moves two. lowerCamelCase : int = rand_fn(lowerCamelCase, lowerCamelCase, lowerCamelCase ) lowerCamelCase : List[str] = rand_fn(lowerCamelCase, lowerCamelCase, lowerCamelCase ) lowerCamelCase : int = rand_fn(lowerCamelCase, lowerCamelCase, lowerCamelCase ) # At some point both the tortoise and the hare will enter a cycle whose # length ``p`` is a divisor of ``num``. Once in that cycle, at some point # the tortoise and hare will end up on the same value modulo ``p``. # We can detect when this happens because the position difference between # the tortoise and the hare will share a common divisor with ``num``. lowerCamelCase : Any = gcd(hare - tortoise, lowerCamelCase ) if divisor == 1: # No common divisor yet, just keep searching. continue else: # We found a common divisor! if divisor == num: # Unfortunately, the divisor is ``num`` itself and is useless. break else: # The divisor is a nontrivial factor of ``num``! return divisor # If we made it here, then this attempt failed. # We need to pick a new starting seed for the tortoise and hare # in addition to a new step value for the random function. # To keep this example implementation deterministic, the # new values will be generated based on currently available # values instead of using something like ``random.randint``. # We can use the hare's position as the new seed. # This is actually what Richard Brent's the "optimized" variant does. lowerCamelCase : Tuple = hare # The new step value for the random function can just be incremented. # At first the results will be similar to what the old function would # have produced, but the value will quickly diverge after a bit. step += 1 # We haven't found a divisor within the requested number of attempts. # We were unlucky or ``num`` itself is actually prime. return None if __name__ == "__main__": import argparse _lowerCamelCase =argparse.ArgumentParser() parser.add_argument( """num""", type=int, help="""The value to find a divisor of""", ) parser.add_argument( """--attempts""", type=int, default=3, help="""The number of attempts before giving up""", ) _lowerCamelCase =parser.parse_args() _lowerCamelCase =pollard_rho(args.num, attempts=args.attempts) if divisor is None: print(f'''{args.num} is probably prime''') else: _lowerCamelCase =args.num // divisor print(f'''{args.num} = {divisor} * {quotient}''')
287
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration _lowerCamelCase =5_0_0_0_0_0 _lowerCamelCase , _lowerCamelCase =os.path.split(__file__) _lowerCamelCase =os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.map(**lowerCamelCase ) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.filter(**lowerCamelCase ) def _a ( ): lowerCamelCase : Optional[Any] = {"""num examples""": SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : Any = datasets.Features({"""text""": datasets.Value("""string""" ), """numbers""": datasets.Value("""float32""" )} ) lowerCamelCase : Tuple = generate_example_dataset( os.path.join(lowerCamelCase, """dataset.arrow""" ), lowerCamelCase, num_examples=lowerCamelCase ) lowerCamelCase : Tuple = transformers.AutoTokenizer.from_pretrained("""bert-base-cased""", use_fast=lowerCamelCase ) def tokenize(lowerCamelCase ): return tokenizer(examples["""text"""] ) lowerCamelCase : List[str] = map(lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""numpy""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""pandas""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""torch""", columns="""numbers""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""tensorflow""", columns="""numbers""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : Union[str, Any] = filter(lowerCamelCase ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(lowerCamelCase, """wb""" ) as f: f.write(json.dumps(lowerCamelCase ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
287
1
_lowerCamelCase =8.3144598 def _a ( lowerCamelCase, lowerCamelCase ): if temperature < 0: raise Exception("""Temperature cannot be less than 0 K""" ) if molar_mass <= 0: raise Exception("""Molar mass cannot be less than or equal to 0 kg/mol""" ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _lowerCamelCase =3_0_0 _lowerCamelCase =2_8 _lowerCamelCase =rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
287
def _a ( lowerCamelCase ): if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True lowerCamelCase : Any = 4 lowerCamelCase : List[str] = (1 << p) - 1 for _ in range(p - 2 ): lowerCamelCase : List[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(1_1))
287
1
from sklearn.metrics import fa_score import datasets _lowerCamelCase =""" The F1 score is the harmonic mean of the precision and recall. It can be computed with the equation: F1 = 2 * (precision * recall) / (precision + recall) """ _lowerCamelCase =""" Args: predictions (`list` of `int`): Predicted labels. references (`list` of `int`): Ground truth labels. labels (`list` of `int`): The set of labels to include when `average` is not set to `'binary'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None. pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1. average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `'binary'`. - 'binary': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary. - 'micro': Calculate metrics globally by counting the total true positives, false negatives and false positives. - 'macro': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - 'weighted': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `'macro'` to account for label imbalance. This option can result in an F-score that is not between precision and recall. - 'samples': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). sample_weight (`list` of `float`): Sample weights Defaults to None. Returns: f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better. Examples: Example 1-A simple binary example >>> f1_metric = datasets.load_metric(\"f1\") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0]) >>> print(results) {'f1': 0.5} Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`. >>> f1_metric = datasets.load_metric(\"f1\") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0) >>> print(round(results['f1'], 2)) 0.67 Example 3-The same simple binary example as in Example 1, but with `sample_weight` included. >>> f1_metric = datasets.load_metric(\"f1\") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3]) >>> print(round(results['f1'], 2)) 0.35 Example 4-A multiclass example, with different values for the `average` input. >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = f1_metric.compute(predictions=predictions, references=references, average=\"macro\") >>> print(round(results['f1'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average=\"micro\") >>> print(round(results['f1'], 2)) 0.33 >>> results = f1_metric.compute(predictions=predictions, references=references, average=\"weighted\") >>> print(round(results['f1'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {'f1': array([0.8, 0. , 0. ])} """ _lowerCamelCase =""" @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class A__ ( datasets.Metric): def UpperCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""int32""" ) ), """references""": datasets.Sequence(datasets.Value("""int32""" ) ), } if self.config_name == """multilabel""" else { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=["""https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html"""] , ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__=None , __magic_name__=1 , __magic_name__="binary" , __magic_name__=None ): lowerCamelCase : Dict = fa_score( __magic_name__ , __magic_name__ , labels=__magic_name__ , pos_label=__magic_name__ , average=__magic_name__ , sample_weight=__magic_name__ ) return {"f1": float(__magic_name__ ) if score.size == 1 else score}
287
import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , *__magic_name__ , **__magic_name__ ): warnings.warn( """The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use OwlViTImageProcessor instead.""" , __magic_name__ , ) super().__init__(*__magic_name__ , **__magic_name__ )
287
1
from cva import destroyAllWindows, imread, imshow, waitKey def _a ( lowerCamelCase ): # getting number of pixels in the image lowerCamelCase , lowerCamelCase : Union[str, Any] = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(lowerCamelCase ): for j in range(lowerCamelCase ): lowerCamelCase : Optional[int] = [255, 255, 255] - img[i][j] return img if __name__ == "__main__": # read original image _lowerCamelCase =imread("""image_data/lena.jpg""", 1) # convert to its negative _lowerCamelCase =convert_to_negative(img) # show result image imshow("""negative of original image""", img) waitKey(0) destroyAllWindows()
287
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """nvidia/segformer-b0-finetuned-ade-512-512""": ( """https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json""" ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """segformer""" def __init__( self , __magic_name__=3 , __magic_name__=4 , __magic_name__=[2, 2, 2, 2] , __magic_name__=[8, 4, 2, 1] , __magic_name__=[3_2, 6_4, 1_6_0, 2_5_6] , __magic_name__=[7, 3, 3, 3] , __magic_name__=[4, 2, 2, 2] , __magic_name__=[1, 2, 5, 8] , __magic_name__=[4, 4, 4, 4] , __magic_name__="gelu" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=0.02 , __magic_name__=0.1 , __magic_name__=1e-6 , __magic_name__=2_5_6 , __magic_name__=2_5_5 , **__magic_name__ , ): super().__init__(**__magic_name__ ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( """Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be""" """ removed, as the behaviour will default to that of reshape_last_stage = True.""" , __magic_name__ , ) lowerCamelCase : Optional[Any] = num_channels lowerCamelCase : str = num_encoder_blocks lowerCamelCase : Any = depths lowerCamelCase : List[Any] = sr_ratios lowerCamelCase : int = hidden_sizes lowerCamelCase : Union[str, Any] = patch_sizes lowerCamelCase : Optional[Any] = strides lowerCamelCase : Dict = mlp_ratios lowerCamelCase : str = num_attention_heads lowerCamelCase : Any = hidden_act lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase : Dict = classifier_dropout_prob lowerCamelCase : Tuple = initializer_range lowerCamelCase : Dict = drop_path_rate lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : Any = decoder_hidden_size lowerCamelCase : str = kwargs.get("""reshape_last_stage""" , __magic_name__ ) lowerCamelCase : Dict = semantic_loss_ignore_index class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-4 @property def UpperCamelCase__ ( self ): return 1_2
287
1
class A__ : def __init__( self ): lowerCamelCase : Optional[int] = """""" lowerCamelCase : Dict = """""" lowerCamelCase : Optional[int] = [] def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): if m == -1: return n + 1 elif n == -1: return m + 1 elif self.dp[m][n] > -1: return self.dp[m][n] else: if self.worda[m] == self.worda[n]: lowerCamelCase : Dict = self.__min_dist_top_down_dp(m - 1 , n - 1 ) else: lowerCamelCase : str = self.__min_dist_top_down_dp(__magic_name__ , n - 1 ) lowerCamelCase : Dict = self.__min_dist_top_down_dp(m - 1 , __magic_name__ ) lowerCamelCase : Tuple = self.__min_dist_top_down_dp(m - 1 , n - 1 ) lowerCamelCase : int = 1 + min(__magic_name__ , __magic_name__ , __magic_name__ ) return self.dp[m][n] def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): lowerCamelCase : Union[str, Any] = worda lowerCamelCase : Optional[int] = worda lowerCamelCase : Tuple = [[-1 for _ in range(len(__magic_name__ ) )] for _ in range(len(__magic_name__ ) )] return self.__min_dist_top_down_dp(len(__magic_name__ ) - 1 , len(__magic_name__ ) - 1 ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): lowerCamelCase : Tuple = worda lowerCamelCase : List[str] = worda lowerCamelCase : str = len(__magic_name__ ) lowerCamelCase : str = len(__magic_name__ ) lowerCamelCase : List[str] = [[0 for _ in range(n + 1 )] for _ in range(m + 1 )] for i in range(m + 1 ): for j in range(n + 1 ): if i == 0: # first string is empty lowerCamelCase : List[str] = j elif j == 0: # second string is empty lowerCamelCase : Union[str, Any] = i elif worda[i - 1] == worda[j - 1]: # last characters are equal lowerCamelCase : Any = self.dp[i - 1][j - 1] else: lowerCamelCase : List[str] = self.dp[i][j - 1] lowerCamelCase : Any = self.dp[i - 1][j] lowerCamelCase : Optional[int] = self.dp[i - 1][j - 1] lowerCamelCase : int = 1 + min(__magic_name__ , __magic_name__ , __magic_name__ ) return self.dp[m][n] if __name__ == "__main__": _lowerCamelCase =EditDistance() print("""****************** Testing Edit Distance DP Algorithm ******************""") print() _lowerCamelCase =input("""Enter the first string: """).strip() _lowerCamelCase =input("""Enter the second string: """).strip() print() print(f'''The minimum edit distance is: {solver.min_dist_top_down(Sa, Sa)}''') print(f'''The minimum edit distance is: {solver.min_dist_bottom_up(Sa, Sa)}''') print() print("""*************** End of Testing Edit Distance DP Algorithm ***************""")
287
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = """gpt_neo""" _UpperCAmelCase : Union[str, Any] = ["""past_key_values"""] _UpperCAmelCase : List[Any] = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , __magic_name__=5_0_2_5_7 , __magic_name__=2_0_4_8 , __magic_name__=2_0_4_8 , __magic_name__=2_4 , __magic_name__=[[["global", "local"], 1_2]] , __magic_name__=1_6 , __magic_name__=None , __magic_name__=2_5_6 , __magic_name__="gelu_new" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=1e-5 , __magic_name__=0.02 , __magic_name__=True , __magic_name__=5_0_2_5_6 , __magic_name__=5_0_2_5_6 , **__magic_name__ , ): lowerCamelCase : List[Any] = vocab_size lowerCamelCase : str = max_position_embeddings lowerCamelCase : str = hidden_size lowerCamelCase : Optional[int] = num_layers lowerCamelCase : str = num_heads lowerCamelCase : Optional[Any] = intermediate_size lowerCamelCase : List[Any] = window_size lowerCamelCase : int = activation_function lowerCamelCase : Union[str, Any] = resid_dropout lowerCamelCase : List[Any] = embed_dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Dict = classifier_dropout lowerCamelCase : Any = layer_norm_epsilon lowerCamelCase : Dict = initializer_range lowerCamelCase : Dict = use_cache lowerCamelCase : Optional[Any] = bos_token_id lowerCamelCase : int = eos_token_id lowerCamelCase : List[Any] = attention_types lowerCamelCase : Optional[Any] = self.expand_attention_types_params(__magic_name__ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.attention_layers)` == `config.num_layers` """ F'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' """`config.attention_layers` is prepared using `config.attention_types`. """ """Please verify the value of `config.attention_types` argument.""" ) super().__init__(bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ ): lowerCamelCase : Optional[int] = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : Any = input.size() lowerCamelCase : List[Any] = len(lowerCamelCase ) lowerCamelCase : Optional[Any] = shape[dimension] lowerCamelCase : Optional[int] = torch.arange(0, lowerCamelCase, lowerCamelCase ) lowerCamelCase : Dict = torch.div(sizedim - size, lowerCamelCase, rounding_mode="""floor""" ) + 1 lowerCamelCase : int = torch.arange(lowerCamelCase ) + low_indices[:min_length][:, None] lowerCamelCase : str = [slice(lowerCamelCase )] * rank lowerCamelCase : List[str] = indices lowerCamelCase : Dict = input[s] lowerCamelCase : Any = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : List[Any] = torch.arange(1, lowerCamelCase ) lowerCamelCase : Optional[int] = torch.remainder(lowerCamelCase, lowerCamelCase ) lowerCamelCase : List[Any] = remainders == 0 lowerCamelCase : List[Any] = candidates[divisor_indices] lowerCamelCase : Optional[Any] = torch.max(lowerCamelCase ) return largest_divisor, torch.div(lowerCamelCase, lowerCamelCase, rounding_mode="""floor""" ) class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): lowerCamelCase : str = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(__magic_name__ , direction="""inputs""" ) lowerCamelCase : int = {0: """batch""", 1: """past_sequence + sequence"""} else: lowerCamelCase : Tuple = {0: """batch""", 1: """sequence"""} return common_inputs @property def UpperCamelCase__ ( self ): return self._config.num_heads def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ): lowerCamelCase : Optional[int] = super(__magic_name__ , self ).generate_dummy_inputs( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) # We need to order the input in the way they appears in the forward() lowerCamelCase : int = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch lowerCamelCase , lowerCamelCase : Optional[Any] = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values lowerCamelCase : Optional[int] = seqlen + 2 lowerCamelCase : List[Any] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowerCamelCase : str = [ (torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) for _ in range(self.num_layers ) ] lowerCamelCase : Tuple = common_inputs["""attention_mask"""] if self.use_past: lowerCamelCase : str = ordered_inputs["""attention_mask"""].dtype lowerCamelCase : Any = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(__magic_name__ , __magic_name__ , dtype=__magic_name__ )] , dim=1 ) return ordered_inputs @property def UpperCamelCase__ ( self ): return 1_3
287
1
import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _lowerCamelCase =parse(importlib.metadata.version("""torch""")) def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(F'''`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}''' ) lowerCamelCase : Optional[Any] = STR_OPERATION_TO_FUNC[operation] if isinstance(lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[Any] = parse(importlib.metadata.version(lowerCamelCase ) ) return operation(lowerCamelCase, parse(lowerCamelCase ) ) def _a ( lowerCamelCase, lowerCamelCase ): return compare_versions(lowerCamelCase, lowerCamelCase, lowerCamelCase )
287
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
287
1
# Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar _lowerCamelCase =TypeVar("""T""") class A__ ( Generic[T]): def __init__( self , __magic_name__ = True ): lowerCamelCase : dict[T, list[T]] = {} # dictionary of lists lowerCamelCase : Dict = directed def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(__magic_name__ ) self.adj_list[destination_vertex].append(__magic_name__ ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(__magic_name__ ) lowerCamelCase : Any = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(__magic_name__ ) lowerCamelCase : Optional[Any] = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: lowerCamelCase : Any = [destination_vertex] lowerCamelCase : Tuple = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(__magic_name__ ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(__magic_name__ ) lowerCamelCase : List[str] = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: lowerCamelCase : List[str] = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: lowerCamelCase : Optional[Any] = [destination_vertex] lowerCamelCase : Tuple = [] return self def __repr__( self ): return pformat(self.adj_list )
287
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): for param, grad_param in zip(model_a.parameters(), model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=True ): model.train() lowerCamelCase : Dict = model(lowerCamelCase ) lowerCamelCase : Any = F.mse_loss(lowerCamelCase, target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase=False ): set_seed(42 ) lowerCamelCase : Tuple = RegressionModel() lowerCamelCase : Any = deepcopy(lowerCamelCase ) lowerCamelCase : Any = RegressionDataset(length=80 ) lowerCamelCase : Dict = DataLoader(lowerCamelCase, batch_size=16 ) model.to(accelerator.device ) if sched: lowerCamelCase : int = AdamW(params=model.parameters(), lr=1e-3 ) lowerCamelCase : Optional[Any] = AdamW(params=ddp_model.parameters(), lr=1e-3 ) lowerCamelCase : str = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) lowerCamelCase : Tuple = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) # Make a copy of `model` if sched: lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = accelerator.prepare(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: lowerCamelCase , lowerCamelCase : List[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def _a ( lowerCamelCase ): # Test when on a single CPU or GPU that the context manager does nothing lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : int = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad, ddp_param.grad ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : List[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase ): # Test on distributed setup that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : int = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Any = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : str = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Optional[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : Any = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : Any = get_training_setup(lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : str = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(lowerCamelCase ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Any = ddp_input[torch.randperm(len(lowerCamelCase ) )] GradientState._reset_state() def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : List[Any] = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase, lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(lowerCamelCase )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]["lr"]}\nDDP opt: {ddp_opt.param_groups[0]["lr"]}\n''' lowerCamelCase : Union[str, Any] = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(lowerCamelCase )) if accelerator.num_processes > 1: check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def _a ( ): lowerCamelCase : int = Accelerator() lowerCamelCase : Optional[Any] = RegressionDataset(length=80 ) lowerCamelCase : List[str] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase : int = RegressionDataset(length=96 ) lowerCamelCase : Optional[int] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if iteration < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if batch_num < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def _a ( ): lowerCamelCase : List[Any] = Accelerator() lowerCamelCase : int = accelerator.state if state.local_process_index == 0: print("""**Test `accumulate` gradient accumulation with dataloader break**""" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("""**Test NOOP `no_sync` context manager**""" ) test_noop_sync(lowerCamelCase ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("""**Test Distributed `no_sync` context manager**""" ) test_distributed_sync(lowerCamelCase ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation(lowerCamelCase, lowerCamelCase ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("""<""", """2.0""" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, """`split_batches=False`, `dispatch_batches=False`**""", ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation_with_opt_and_scheduler(lowerCamelCase, lowerCamelCase ) def _a ( lowerCamelCase ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
287
1
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = ["""image_processor"""] _UpperCAmelCase : List[str] = """SamImageProcessor""" def __init__( self , __magic_name__ ): super().__init__(__magic_name__ ) lowerCamelCase : Optional[int] = self.image_processor lowerCamelCase : Tuple = -1_0 lowerCamelCase : Tuple = self.image_processor.size["""longest_edge"""] def __call__( self , __magic_name__=None , __magic_name__=None , __magic_name__=None , __magic_name__=None , __magic_name__ = None , **__magic_name__ , ): lowerCamelCase : int = self.image_processor( __magic_name__ , return_tensors=__magic_name__ , **__magic_name__ , ) # pop arguments that are not used in the foward but used nevertheless lowerCamelCase : List[str] = encoding_image_processor["""original_sizes"""] if hasattr(__magic_name__ , """numpy""" ): # Checks if Torch or TF tensor lowerCamelCase : Optional[Any] = original_sizes.numpy() lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[Any] = self._check_and_preprocess_points( input_points=__magic_name__ , input_labels=__magic_name__ , input_boxes=__magic_name__ , ) lowerCamelCase : str = self._normalize_and_convert( __magic_name__ , __magic_name__ , input_points=__magic_name__ , input_labels=__magic_name__ , input_boxes=__magic_name__ , return_tensors=__magic_name__ , ) return encoding_image_processor def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__=None , __magic_name__=None , __magic_name__=None , __magic_name__="pt" , ): if input_points is not None: if len(__magic_name__ ) != len(__magic_name__ ): lowerCamelCase : Tuple = [ self._normalize_coordinates(self.target_size , __magic_name__ , original_sizes[0] ) for point in input_points ] else: lowerCamelCase : Optional[int] = [ self._normalize_coordinates(self.target_size , __magic_name__ , __magic_name__ ) for point, original_size in zip(__magic_name__ , __magic_name__ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: lowerCamelCase , lowerCamelCase : Optional[int] = self._pad_points_and_labels(__magic_name__ , __magic_name__ ) lowerCamelCase : Dict = np.array(__magic_name__ ) if input_labels is not None: lowerCamelCase : Optional[int] = np.array(__magic_name__ ) if input_boxes is not None: if len(__magic_name__ ) != len(__magic_name__ ): lowerCamelCase : List[Any] = [ self._normalize_coordinates(self.target_size , __magic_name__ , original_sizes[0] , is_bounding_box=__magic_name__ ) for box in input_boxes ] else: lowerCamelCase : List[str] = [ self._normalize_coordinates(self.target_size , __magic_name__ , __magic_name__ , is_bounding_box=__magic_name__ ) for box, original_size in zip(__magic_name__ , __magic_name__ ) ] lowerCamelCase : Dict = np.array(__magic_name__ ) if input_boxes is not None: if return_tensors == "pt": lowerCamelCase : Optional[int] = torch.from_numpy(__magic_name__ ) # boxes batch size of 1 by default lowerCamelCase : List[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": lowerCamelCase : int = tf.convert_to_tensor(__magic_name__ ) # boxes batch size of 1 by default lowerCamelCase : List[str] = tf.expand_dims(__magic_name__ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"""input_boxes""": input_boxes} ) if input_points is not None: if return_tensors == "pt": lowerCamelCase : Tuple = torch.from_numpy(__magic_name__ ) # point batch size of 1 by default lowerCamelCase : str = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": lowerCamelCase : Any = tf.convert_to_tensor(__magic_name__ ) # point batch size of 1 by default lowerCamelCase : int = tf.expand_dims(__magic_name__ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"""input_points""": input_points} ) if input_labels is not None: if return_tensors == "pt": lowerCamelCase : Tuple = torch.from_numpy(__magic_name__ ) # point batch size of 1 by default lowerCamelCase : Union[str, Any] = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": lowerCamelCase : Any = tf.convert_to_tensor(__magic_name__ ) # point batch size of 1 by default lowerCamelCase : Union[str, Any] = tf.expand_dims(__magic_name__ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"""input_labels""": input_labels} ) return encoding_image_processor def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): lowerCamelCase : List[str] = max([point.shape[0] for point in input_points] ) lowerCamelCase : int = [] for i, point in enumerate(__magic_name__ ): if point.shape[0] != expected_nb_points: lowerCamelCase : Dict = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) lowerCamelCase : Tuple = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(__magic_name__ ) lowerCamelCase : Union[str, Any] = processed_input_points return input_points, input_labels def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__=False ): lowerCamelCase , lowerCamelCase : Optional[int] = original_size lowerCamelCase , lowerCamelCase : Union[str, Any] = self.image_processor._get_preprocess_shape(__magic_name__ , longest_edge=__magic_name__ ) lowerCamelCase : List[Any] = deepcopy(__magic_name__ ).astype(__magic_name__ ) if is_bounding_box: lowerCamelCase : Optional[Any] = coords.reshape(-1 , 2 , 2 ) lowerCamelCase : Tuple = coords[..., 0] * (new_w / old_w) lowerCamelCase : Dict = coords[..., 1] * (new_h / old_h) if is_bounding_box: lowerCamelCase : Union[str, Any] = coords.reshape(-1 , 4 ) return coords def UpperCamelCase__ ( self , __magic_name__=None , __magic_name__=None , __magic_name__=None , ): if input_points is not None: if hasattr(__magic_name__ , """numpy""" ): # Checks for TF or Torch tensor lowerCamelCase : List[Any] = input_points.numpy().tolist() if not isinstance(__magic_name__ , __magic_name__ ) or not isinstance(input_points[0] , __magic_name__ ): raise ValueError("""Input points must be a list of list of floating points.""" ) lowerCamelCase : List[Any] = [np.array(__magic_name__ ) for input_point in input_points] else: lowerCamelCase : str = None if input_labels is not None: if hasattr(__magic_name__ , """numpy""" ): lowerCamelCase : int = input_labels.numpy().tolist() if not isinstance(__magic_name__ , __magic_name__ ) or not isinstance(input_labels[0] , __magic_name__ ): raise ValueError("""Input labels must be a list of list integers.""" ) lowerCamelCase : Optional[int] = [np.array(__magic_name__ ) for label in input_labels] else: lowerCamelCase : Dict = None if input_boxes is not None: if hasattr(__magic_name__ , """numpy""" ): lowerCamelCase : Union[str, Any] = input_boxes.numpy().tolist() if ( not isinstance(__magic_name__ , __magic_name__ ) or not isinstance(input_boxes[0] , __magic_name__ ) or not isinstance(input_boxes[0][0] , __magic_name__ ) ): raise ValueError("""Input boxes must be a list of list of list of floating points.""" ) lowerCamelCase : Dict = [np.array(__magic_name__ ).astype(np.floataa ) for box in input_boxes] else: lowerCamelCase : List[Any] = None return input_points, input_labels, input_boxes @property def UpperCamelCase__ ( self ): lowerCamelCase : Optional[Any] = self.image_processor.model_input_names return list(dict.fromkeys(__magic_name__ ) ) def UpperCamelCase__ ( self , *__magic_name__ , **__magic_name__ ): return self.image_processor.post_process_masks(*__magic_name__ , **__magic_name__ )
287
from scipy.stats import pearsonr import datasets _lowerCamelCase =""" Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ _lowerCamelCase =""" Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ _lowerCamelCase =""" @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class A__ ( datasets.Metric): def UpperCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"""] , ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__=False ): if return_pvalue: lowerCamelCase : Optional[Any] = pearsonr(__magic_name__ , __magic_name__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(__magic_name__ , __magic_name__ )[0] )}
287
1
_lowerCamelCase ={ """A""": ["""B""", """C""", """E"""], """B""": ["""A""", """D""", """E"""], """C""": ["""A""", """F""", """G"""], """D""": ["""B"""], """E""": ["""A""", """B""", """D"""], """F""": ["""C"""], """G""": ["""C"""], } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowerCamelCase : str = set() # keep track of all the paths to be checked lowerCamelCase : Any = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue lowerCamelCase : Any = queue.pop(0 ) # get the last node from the path lowerCamelCase : Union[str, Any] = path[-1] if node not in explored: lowerCamelCase : List[Any] = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: lowerCamelCase : str = list(lowerCamelCase ) new_path.append(lowerCamelCase ) queue.append(lowerCamelCase ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(lowerCamelCase ) # in case there's no path between the 2 nodes return [] def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 lowerCamelCase : Optional[Any] = [start] lowerCamelCase : List[Any] = set(lowerCamelCase ) # Keep tab on distances from `start` node. lowerCamelCase : int = {start: 0, target: -1} while queue: lowerCamelCase : List[Any] = queue.pop(0 ) if node == target: lowerCamelCase : Dict = ( dist[node] if dist[target] == -1 else min(dist[target], dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(lowerCamelCase ) queue.append(lowerCamelCase ) lowerCamelCase : Dict = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, """G""", """D""")) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, """G""", """D""")) # returns 4
287
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """microsoft/conditional-detr-resnet-50""": ( """https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """conditional_detr""" _UpperCAmelCase : Optional[int] = ["""past_key_values"""] _UpperCAmelCase : Optional[Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , __magic_name__=True , __magic_name__=None , __magic_name__=3 , __magic_name__=3_0_0 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=True , __magic_name__="relu" , __magic_name__=2_5_6 , __magic_name__=0.1 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.02 , __magic_name__=1.0 , __magic_name__=False , __magic_name__="sine" , __magic_name__="resnet50" , __magic_name__=True , __magic_name__=False , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=1 , __magic_name__=1 , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=0.25 , **__magic_name__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) lowerCamelCase : Optional[int] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(__magic_name__ , __magic_name__ ): lowerCamelCase : List[Any] = backbone_config.get("""model_type""" ) lowerCamelCase : Dict = CONFIG_MAPPING[backbone_model_type] lowerCamelCase : str = config_class.from_dict(__magic_name__ ) lowerCamelCase : Dict = use_timm_backbone lowerCamelCase : str = backbone_config lowerCamelCase : Tuple = num_channels lowerCamelCase : Dict = num_queries lowerCamelCase : Any = d_model lowerCamelCase : Optional[Any] = encoder_ffn_dim lowerCamelCase : List[str] = encoder_layers lowerCamelCase : Union[str, Any] = encoder_attention_heads lowerCamelCase : Any = decoder_ffn_dim lowerCamelCase : Dict = decoder_layers lowerCamelCase : Union[str, Any] = decoder_attention_heads lowerCamelCase : Dict = dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Union[str, Any] = activation_dropout lowerCamelCase : Optional[int] = activation_function lowerCamelCase : int = init_std lowerCamelCase : str = init_xavier_std lowerCamelCase : Tuple = encoder_layerdrop lowerCamelCase : str = decoder_layerdrop lowerCamelCase : Tuple = encoder_layers lowerCamelCase : Optional[int] = auxiliary_loss lowerCamelCase : Optional[Any] = position_embedding_type lowerCamelCase : Optional[int] = backbone lowerCamelCase : Union[str, Any] = use_pretrained_backbone lowerCamelCase : str = dilation # Hungarian matcher lowerCamelCase : Optional[Any] = class_cost lowerCamelCase : Dict = bbox_cost lowerCamelCase : Tuple = giou_cost # Loss coefficients lowerCamelCase : Union[str, Any] = mask_loss_coefficient lowerCamelCase : Dict = dice_loss_coefficient lowerCamelCase : Optional[int] = cls_loss_coefficient lowerCamelCase : Optional[int] = bbox_loss_coefficient lowerCamelCase : Optional[int] = giou_loss_coefficient lowerCamelCase : Optional[int] = focal_alpha super().__init__(is_encoder_decoder=__magic_name__ , **__magic_name__ ) @property def UpperCamelCase__ ( self ): return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): return self.d_model def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: lowerCamelCase : Optional[int] = self.backbone_config.to_dict() lowerCamelCase : Optional[Any] = self.__class__.model_type return output class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-5 @property def UpperCamelCase__ ( self ): return 1_2
287
1
def _a ( lowerCamelCase = 10, lowerCamelCase = 1000, lowerCamelCase = True ): assert ( isinstance(lowerCamelCase, lowerCamelCase ) and isinstance(lowerCamelCase, lowerCamelCase ) and isinstance(lowerCamelCase, lowerCamelCase ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("""Invalid value for min_val or max_val (min_value < max_value)""" ) return min_val if option else max_val def _a ( lowerCamelCase, lowerCamelCase ): return int((number_a + number_a) / 2 ) def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): assert ( isinstance(lowerCamelCase, lowerCamelCase ) and isinstance(lowerCamelCase, lowerCamelCase ) and isinstance(lowerCamelCase, lowerCamelCase ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("""argument value for lower and higher must be(lower > higher)""" ) if not lower < to_guess < higher: raise ValueError( """guess value must be within the range of lower and higher value""" ) def answer(lowerCamelCase ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("""started...""" ) lowerCamelCase : Optional[Any] = lower lowerCamelCase : str = higher lowerCamelCase : List[str] = [] while True: lowerCamelCase : Tuple = get_avg(lowerCamelCase, lowerCamelCase ) last_numbers.append(lowerCamelCase ) if answer(lowerCamelCase ) == "low": lowerCamelCase : int = number elif answer(lowerCamelCase ) == "high": lowerCamelCase : Any = number else: break print(F'''guess the number : {last_numbers[-1]}''' ) print(F'''details : {last_numbers!s}''' ) def _a ( ): lowerCamelCase : int = int(input("""Enter lower value : """ ).strip() ) lowerCamelCase : Any = int(input("""Enter high value : """ ).strip() ) lowerCamelCase : Union[str, Any] = int(input("""Enter value to guess : """ ).strip() ) guess_the_number(lowerCamelCase, lowerCamelCase, lowerCamelCase ) if __name__ == "__main__": main()
287
import json import sys def _a ( lowerCamelCase, lowerCamelCase ): with open(lowerCamelCase, encoding="""utf-8""" ) as f: lowerCamelCase : List[Any] = json.load(lowerCamelCase ) lowerCamelCase : Optional[Any] = ["""<details>""", """<summary>Show updated benchmarks!</summary>""", """ """] for benchmark_name in sorted(lowerCamelCase ): lowerCamelCase : List[Any] = results[benchmark_name] lowerCamelCase : Union[str, Any] = benchmark_name.split("""/""" )[-1] output_md.append(F'''### Benchmark: {benchmark_file_name}''' ) lowerCamelCase : Any = """| metric |""" lowerCamelCase : str = """|--------|""" lowerCamelCase : List[Any] = """| new / old (diff) |""" for metric_name in sorted(lowerCamelCase ): lowerCamelCase : List[Any] = benchmark_res[metric_name] lowerCamelCase : Tuple = metric_vals["""new"""] lowerCamelCase : int = metric_vals.get("""old""", lowerCamelCase ) lowerCamelCase : Dict = metric_vals.get("""diff""", lowerCamelCase ) lowerCamelCase : Dict = F''' {new_val:f}''' if isinstance(lowerCamelCase, (int, float) ) else """None""" if old_val is not None: val_str += F''' / {old_val:f}''' if isinstance(lowerCamelCase, (int, float) ) else "None" if dif_val is not None: val_str += F''' ({dif_val:f})''' if isinstance(lowerCamelCase, (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append("""</details>""" ) with open(lowerCamelCase, """w""", encoding="""utf-8""" ) as f: f.writelines("""\n""".join(lowerCamelCase ) ) if __name__ == "__main__": _lowerCamelCase =sys.argv[1] _lowerCamelCase =sys.argv[2] format_json_to_md(input_json_file, output_md_file)
287
1
def _a ( lowerCamelCase, lowerCamelCase = 0 ): lowerCamelCase : Tuple = length or len(lowerCamelCase ) lowerCamelCase : List[Any] = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: lowerCamelCase , lowerCamelCase : Any = list_data[i + 1], list_data[i] lowerCamelCase : Any = True return list_data if not swapped else bubble_sort(lowerCamelCase, length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
287
def _a ( lowerCamelCase ): return " ".join( """""".join(word[::-1] ) if len(lowerCamelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("""Hey wollef sroirraw"""))
287
1
from __future__ import annotations def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : Optional[Any] = sorted(numsa + numsa ) lowerCamelCase , lowerCamelCase : List[Any] = divmod(len(lowerCamelCase ), 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() _lowerCamelCase =[float(x) for x in input("""Enter the elements of first array: """).split()] _lowerCamelCase =[float(x) for x in input("""Enter the elements of second array: """).split()] print(f'''The median of two arrays is: {median_of_two_arrays(array_a, array_a)}''')
287
# Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from packaging import version from .. import __version__ from .constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD from .doc import ( add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, copy_func, replace_return_docstrings, ) from .generic import ( ContextManagers, ExplicitEnum, ModelOutput, PaddingStrategy, TensorType, add_model_info_to_auto_map, cached_property, can_return_loss, expand_dims, find_labels, flatten_dict, infer_framework, is_jax_tensor, is_numpy_array, is_tensor, is_tf_symbolic_tensor, is_tf_tensor, is_torch_device, is_torch_dtype, is_torch_tensor, reshape, squeeze, strtobool, tensor_size, to_numpy, to_py_obj, transpose, working_or_temp_dir, ) from .hub import ( CLOUDFRONT_DISTRIB_PREFIX, DISABLE_TELEMETRY, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, EntryNotFoundError, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, cached_file, default_cache_path, define_sagemaker_information, download_url, extract_commit_hash, get_cached_models, get_file_from_repo, get_full_repo_name, has_file, http_user_agent, is_offline_mode, is_remote_url, move_cache, send_example_telemetry, try_to_load_from_cache, ) from .import_utils import ( ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, TORCH_FX_REQUIRED_VERSION, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, ccl_version, direct_transformers_import, get_torch_version, is_accelerate_available, is_apex_available, is_bitsandbytes_available, is_bsa_available, is_coloredlogs_available, is_cython_available, is_datasets_available, is_decord_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_jieba_available, is_jumanpp_available, is_kenlm_available, is_keras_nlp_available, is_librosa_available, is_natten_available, is_ninja_available, is_onnx_available, is_openai_available, is_optimum_available, is_pandas_available, is_peft_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytest_available, is_pytorch_quantization_available, is_rjieba_available, is_sacremoses_available, is_safetensors_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_sudachi_available, is_tensorflow_probability_available, is_tensorflow_text_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_bfaa_cpu_available, is_torch_bfaa_gpu_available, is_torch_compile_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_neuroncore_available, is_torch_tensorrt_fx_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_torchdistx_available, is_torchdynamo_available, is_torchvision_available, is_training_run_on_sagemaker, is_vision_available, requires_backends, torch_only_method, ) _lowerCamelCase ="""pytorch_model.bin""" _lowerCamelCase ="""pytorch_model.bin.index.json""" _lowerCamelCase ="""adapter_config.json""" _lowerCamelCase ="""adapter_model.bin""" _lowerCamelCase ="""adapter_model.safetensors""" _lowerCamelCase ="""tf_model.h5""" _lowerCamelCase ="""tf_model.h5.index.json""" _lowerCamelCase ="""model.ckpt""" _lowerCamelCase ="""flax_model.msgpack""" _lowerCamelCase ="""flax_model.msgpack.index.json""" _lowerCamelCase ="""model.safetensors""" _lowerCamelCase ="""model.safetensors.index.json""" _lowerCamelCase ="""config.json""" _lowerCamelCase ="""preprocessor_config.json""" _lowerCamelCase =FEATURE_EXTRACTOR_NAME _lowerCamelCase ="""generation_config.json""" _lowerCamelCase ="""modelcard.json""" _lowerCamelCase ="""▁""" _lowerCamelCase =SENTENCEPIECE_UNDERLINE # Kept for backward compatibility _lowerCamelCase =[ [[0, 1, 0, 1], [1, 0, 0, 1]] ] * 2 # Needs to have 0s and 1s only since XLM uses it for langs too. _lowerCamelCase =[[7, 6, 0, 0, 1], [1, 2, 3, 0, 0], [0, 0, 0, 4, 5]] _lowerCamelCase =[[1, 1, 1, 1, 1], [1, 1, 1, 0, 0], [0, 0, 0, 1, 1]] def _a ( lowerCamelCase ): if version.parse(lowerCamelCase ) < version.parse(lowerCamelCase ): if "dev" in min_version: lowerCamelCase : Optional[int] = ( """This example requires a source install from HuggingFace Transformers (see """ """`https://huggingface.co/docs/transformers/installation#install-from-source`),""" ) else: lowerCamelCase : int = F'''This example requires a minimum version of {min_version},''' error_message += F''' but the version found is {__version__}.\n''' raise ImportError( error_message + """Check out https://github.com/huggingface/transformers/tree/main/examples#important-note for the examples corresponding to other """ """versions of HuggingFace Transformers.""" )
287
1
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging _lowerCamelCase =logging.get_logger(__name__) def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowerCamelCase : int = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(lowerCamelCase ) == len(lowerCamelCase ), F'''{len(lowerCamelCase )} != {len(lowerCamelCase )}''' dest_layers.load_state_dict(layers_to_copy.state_dict() ) _lowerCamelCase ={ # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 1_2: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 1_1], 4: [0, 4, 8, 1_1], 6: [0, 2, 4, 7, 9, 1_1], 9: [0, 1, 2, 4, 5, 7, 9, 1_0, 1_1], 1_2: list(range(1_2)), }, 1_6: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 1_5], 3: [0, 8, 1_5], 4: [0, 5, 1_0, 1_5], 6: [0, 3, 6, 9, 1_2, 1_5], 8: [0, 2, 4, 6, 8, 1_0, 1_2, 1_5], 9: [0, 1, 3, 5, 7, 9, 1_1, 1_3, 1_5], 1_2: [0, 1, 2, 3, 4, 5, 6, 7, 9, 1_1, 1_3, 1_5], 1_6: list(range(1_6)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } _lowerCamelCase ={ # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 1_2: {1: [1_1], 2: [5, 1_1], 3: [3, 7, 1_1], 6: [1, 3, 5, 8, 1_0, 1_1]}, 1_6: {1: [1_5], 4: [4, 9, 1_2, 1_5], 8: [1, 3, 5, 7, 9, 1_1, 1_3, 1_5]}, } def _a ( lowerCamelCase, lowerCamelCase ): try: lowerCamelCase : List[str] = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F'''no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first''' F''' {n_student}''' ) return list(range(lowerCamelCase ) ) def _a ( lowerCamelCase, lowerCamelCase ): if n_student > n_teacher: raise ValueError(F'''Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}''' ) elif n_teacher == n_student: return list(range(lowerCamelCase ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def _a ( lowerCamelCase, lowerCamelCase = "student", lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase=False, lowerCamelCase=None, lowerCamelCase=None, **lowerCamelCase, ): lowerCamelCase : Optional[Any] = """encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.""" assert (e is not None) or (d is not None), _msg if isinstance(lowerCamelCase, lowerCamelCase ): AutoTokenizer.from_pretrained(lowerCamelCase ).save_pretrained(lowerCamelCase ) # purely for convenience lowerCamelCase : int = AutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase ).eval() else: assert isinstance(lowerCamelCase, lowerCamelCase ), F'''teacher must be a model or string got type {type(lowerCamelCase )}''' lowerCamelCase : Tuple = teacher.config.to_diff_dict() try: lowerCamelCase , lowerCamelCase : List[str] = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: lowerCamelCase : List[Any] = teacher_e if d is None: lowerCamelCase : List[Any] = teacher_d init_kwargs.update({"""encoder_layers""": e, """decoder_layers""": d} ) except AttributeError: # T5 if hasattr(teacher.config, """num_encoder_layers""" ): lowerCamelCase , lowerCamelCase : Dict = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: lowerCamelCase , lowerCamelCase : int = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: lowerCamelCase : Any = teacher_e if d is None: lowerCamelCase : List[str] = teacher_d if hasattr(teacher.config, """num_encoder_layers""" ): init_kwargs.update({"""num_encoder_layers""": e, """num_decoder_layers""": d} ) else: init_kwargs.update({"""num_layers""": e, """num_decoder_layers""": d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(lowerCamelCase ) # Copy weights lowerCamelCase : Tuple = teacher.config_class(**lowerCamelCase ) lowerCamelCase : Optional[Any] = AutoModelForSeqaSeqLM.from_config(lowerCamelCase ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. lowerCamelCase : Any = student.load_state_dict(teacher.state_dict(), strict=lowerCamelCase ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save lowerCamelCase , lowerCamelCase : List[str] = list(range(lowerCamelCase ) ), list(range(lowerCamelCase ) ) logger.info( F'''Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to''' F''' {save_path}''' ) student.save_pretrained(lowerCamelCase ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: lowerCamelCase : List[int] = pick_layers_to_copy(lowerCamelCase, lowerCamelCase ) if d_layers_to_copy is None: lowerCamelCase : List[int] = pick_layers_to_copy(lowerCamelCase, lowerCamelCase ) try: if hasattr( lowerCamelCase, """prophetnet""" ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers, student.prophetnet.encoder.layers, lowerCamelCase ) copy_layers(teacher.prophetnet.decoder.layers, student.prophetnet.decoder.layers, lowerCamelCase ) else: copy_layers(teacher.model.encoder.layers, student.model.encoder.layers, lowerCamelCase ) copy_layers(teacher.model.decoder.layers, student.model.decoder.layers, lowerCamelCase ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block, student.encoder.block, lowerCamelCase ) copy_layers(teacher.decoder.block, student.decoder.block, lowerCamelCase ) logger.info( F'''Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}''' ) lowerCamelCase : Any = { """teacher_type""": teacher.config.model_type, """copied_encoder_layers""": e_layers_to_copy, """copied_decoder_layers""": d_layers_to_copy, } student.save_pretrained(lowerCamelCase ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
287
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = """camembert""" def __init__( self , __magic_name__=3_0_5_2_2 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , __magic_name__=None , **__magic_name__ , ): super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : int = num_hidden_layers lowerCamelCase : int = num_attention_heads lowerCamelCase : Optional[int] = hidden_act lowerCamelCase : List[Any] = intermediate_size lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Optional[int] = attention_probs_dropout_prob lowerCamelCase : Optional[int] = max_position_embeddings lowerCamelCase : str = type_vocab_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : int = layer_norm_eps lowerCamelCase : Any = position_embedding_type lowerCamelCase : Optional[int] = use_cache lowerCamelCase : Union[str, Any] = classifier_dropout class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): if self.task == "multiple-choice": lowerCamelCase : List[str] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase : List[str] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
287
1
import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={"""vocab_file""": """spiece.model"""} _lowerCamelCase ={ """vocab_file""": { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model""", """google/bigbird-roberta-large""": ( """https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model""" ), """google/bigbird-base-trivia-itc""": ( """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model""" ), } } _lowerCamelCase ={ """google/bigbird-roberta-base""": 4_0_9_6, """google/bigbird-roberta-large""": 4_0_9_6, """google/bigbird-base-trivia-itc""": 4_0_9_6, } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : int = VOCAB_FILES_NAMES _UpperCAmelCase : int = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : Tuple = ["""input_ids""", """attention_mask"""] _UpperCAmelCase : List[int] = [] def __init__( self , __magic_name__ , __magic_name__="<unk>" , __magic_name__="<s>" , __magic_name__="</s>" , __magic_name__="<pad>" , __magic_name__="[SEP]" , __magic_name__="[MASK]" , __magic_name__="[CLS]" , __magic_name__ = None , **__magic_name__ , ): lowerCamelCase : Tuple = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else bos_token lowerCamelCase : Optional[int] = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else eos_token lowerCamelCase : str = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else unk_token lowerCamelCase : int = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else pad_token lowerCamelCase : Optional[Any] = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else cls_token lowerCamelCase : Any = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase : int = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else mask_token lowerCamelCase : List[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__magic_name__ , eos_token=__magic_name__ , unk_token=__magic_name__ , pad_token=__magic_name__ , sep_token=__magic_name__ , mask_token=__magic_name__ , cls_token=__magic_name__ , sp_model_kwargs=self.sp_model_kwargs , **__magic_name__ , ) lowerCamelCase : Any = vocab_file lowerCamelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__magic_name__ ) @property def UpperCamelCase__ ( self ): return self.sp_model.get_piece_size() def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = {self.convert_ids_to_tokens(__magic_name__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowerCamelCase : str = self.__dict__.copy() lowerCamelCase : Optional[int] = None return state def __setstate__( self , __magic_name__ ): lowerCamelCase : Union[str, Any] = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): lowerCamelCase : Union[str, Any] = {} lowerCamelCase : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCamelCase__ ( self , __magic_name__ ): return self.sp_model.encode(__magic_name__ , out_type=__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ ): return self.sp_model.piece_to_id(__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : List[str] = self.sp_model.IdToPiece(__magic_name__ ) return token def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : Any = [] lowerCamelCase : Dict = """""" lowerCamelCase : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__magic_name__ ) + token lowerCamelCase : Union[str, Any] = True lowerCamelCase : List[str] = [] else: current_sub_tokens.append(__magic_name__ ) lowerCamelCase : Dict = False out_string += self.sp_model.decode(__magic_name__ ) return out_string.strip() def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = False , __magic_name__ = None , __magic_name__ = True , **__magic_name__ , ): lowerCamelCase : Optional[int] = kwargs.pop("""use_source_tokenizer""" , __magic_name__ ) lowerCamelCase : Any = self.convert_ids_to_tokens(__magic_name__ , skip_special_tokens=__magic_name__ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowerCamelCase : Any = [] lowerCamelCase : Tuple = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(__magic_name__ ) ) lowerCamelCase : str = [] sub_texts.append(__magic_name__ ) else: current_sub_text.append(__magic_name__ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(__magic_name__ ) ) # Mimic the behavior of the Rust tokenizer: # No space before [MASK] and [SEP] if spaces_between_special_tokens: lowerCamelCase : Union[str, Any] = re.sub(r""" (\[(MASK|SEP)\])""" , r"""\1""" , """ """.join(__magic_name__ ) ) else: lowerCamelCase : int = """""".join(__magic_name__ ) lowerCamelCase : int = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowerCamelCase : Optional[int] = self.clean_up_tokenization(__magic_name__ ) return clean_text else: return text def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): if not os.path.isdir(__magic_name__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCamelCase : List[str] = os.path.join( __magic_name__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__magic_name__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __magic_name__ ) elif not os.path.isfile(self.vocab_file ): with open(__magic_name__ , """wb""" ) as fi: lowerCamelCase : str = self.sp_model.serialized_model_proto() fi.write(__magic_name__ ) return (out_vocab_file,) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCamelCase : int = [self.cls_token_id] lowerCamelCase : Optional[int] = [self.sep_token_id] return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None , __magic_name__ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__magic_name__ , token_ids_a=__magic_name__ , already_has_special_tokens=__magic_name__ ) if token_ids_a is None: return [1] + ([0] * len(__magic_name__ )) + [1] return [1] + ([0] * len(__magic_name__ )) + [1] + ([0] * len(__magic_name__ )) + [1] def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): lowerCamelCase : str = [self.sep_token_id] lowerCamelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1]
287
_lowerCamelCase ={ "joule": 1.0, "kilojoule": 1_0_0_0, "megajoule": 1_0_0_0_0_0_0, "gigajoule": 1_0_0_0_0_0_0_0_0_0, "wattsecond": 1.0, "watthour": 3_6_0_0, "kilowatthour": 3_6_0_0_0_0_0, "newtonmeter": 1.0, "calorie_nutr": 4_1_8_6.8, "kilocalorie_nutr": 4_1_8_6_8_0_0.0_0, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1_0_5_5.0_5_5_8_5, "footpound": 1.355818, } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowerCamelCase : Dict = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {", ".join(lowerCamelCase )}''' ) raise ValueError(lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
287
1
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 _lowerCamelCase ={ """return_dict""": False, """output_hidden_states""": True, """output_attentions""": True, """torchscript""": True, """torch_dtype""": """float16""", """use_bfloat16""": True, """tf_legacy_loss""": True, """pruned_heads""": {"""a""": 1}, """tie_word_embeddings""": False, """is_decoder""": True, """cross_attention_hidden_size""": 1_2_8, """add_cross_attention""": True, """tie_encoder_decoder""": True, """max_length""": 5_0, """min_length""": 3, """do_sample""": True, """early_stopping""": True, """num_beams""": 3, """num_beam_groups""": 3, """diversity_penalty""": 0.5, """temperature""": 2.0, """top_k""": 1_0, """top_p""": 0.7, """typical_p""": 0.2, """repetition_penalty""": 0.8, """length_penalty""": 0.8, """no_repeat_ngram_size""": 5, """encoder_no_repeat_ngram_size""": 5, """bad_words_ids""": [1, 2, 3], """num_return_sequences""": 3, """chunk_size_feed_forward""": 5, """output_scores""": True, """return_dict_in_generate""": True, """forced_bos_token_id""": 2, """forced_eos_token_id""": 3, """remove_invalid_values""": True, """architectures""": ["""BertModel"""], """finetuning_task""": """translation""", """id2label""": {0: """label"""}, """label2id""": {"""label""": """0"""}, """tokenizer_class""": """BertTokenizerFast""", """prefix""": """prefix""", """bos_token_id""": 6, """pad_token_id""": 7, """eos_token_id""": 8, """sep_token_id""": 9, """decoder_start_token_id""": 1_0, """exponential_decay_length_penalty""": (5, 1.01), """suppress_tokens""": [0, 1], """begin_suppress_tokens""": 2, """task_specific_params""": {"""translation""": """some_params"""}, """problem_type""": """regression""", } @is_staging_test class A__ ( unittest.TestCase): @classmethod def UpperCamelCase__ ( cls ): lowerCamelCase : int = TOKEN HfFolder.save_token(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls ): try: delete_repo(token=cls._token , repo_id="""test-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-config-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-config""" ) except HTTPError: pass def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""test-config""" , use_auth_token=self._token ) lowerCamelCase : Any = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__magic_name__ , repo_id="""test-config""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : Optional[Any] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""valid_org/test-config-org""" , use_auth_token=self._token ) lowerCamelCase : Optional[int] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( __magic_name__ , repo_id="""valid_org/test-config-org""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : List[str] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): CustomConfig.register_for_auto_class() lowerCamelCase : Optional[Any] = CustomConfig(attribute=4_2 ) config.push_to_hub("""test-dynamic-config""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"""AutoConfig""": """custom_configuration.CustomConfig"""} ) lowerCamelCase : List[str] = AutoConfig.from_pretrained(F'''{USER}/test-dynamic-config''' , trust_remote_code=__magic_name__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , """CustomConfig""" ) self.assertEqual(new_config.attribute , 4_2 ) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : str = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated lowerCamelCase : Optional[int] = c.n_embd + 1 # int lowerCamelCase : Optional[int] = c.resid_pdrop + 1.0 # float lowerCamelCase : Tuple = not c.scale_attn_weights # bool lowerCamelCase : Any = c.summary_type + """foo""" # str c.update_from_string( F'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(__magic_name__ , c.n_embd , """mismatch for key: n_embd""" ) self.assertEqual(__magic_name__ , c.resid_pdrop , """mismatch for key: resid_pdrop""" ) self.assertEqual(__magic_name__ , c.scale_attn_weights , """mismatch for key: scale_attn_weights""" ) self.assertEqual(__magic_name__ , c.summary_type , """mismatch for key: summary_type""" ) def UpperCamelCase__ ( self ): lowerCamelCase : str = PretrainedConfig() lowerCamelCase : int = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( __magic_name__ , ["""is_encoder_decoder""", """_name_or_path""", """_commit_hash""", """transformers_version"""] ) lowerCamelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(__magic_name__ , __magic_name__ )] if len(__magic_name__ ) > 0: raise ValueError( """The following keys are set with the default values in""" """ `test_configuration_common.config_common_kwargs` pick another value for them:""" F''' {", ".join(__magic_name__ )}.''' ) def UpperCamelCase__ ( self ): with self.assertRaises(__magic_name__ ): # config is in subfolder, the following should not work without specifying the subfolder lowerCamelCase : Dict = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" ) lowerCamelCase : str = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" , subfolder="""bert""" ) self.assertIsNotNone(__magic_name__ ) def UpperCamelCase__ ( self ): # A mock response for an HTTP head request to emulate server down lowerCamelCase : Dict = mock.Mock() lowerCamelCase : Optional[int] = 5_0_0 lowerCamelCase : List[Any] = {} lowerCamelCase : Tuple = HTTPError lowerCamelCase : Union[str, Any] = {} # Download this model to make sure it's in the cache. lowerCamelCase : List[str] = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=__magic_name__ ) as mock_head: lowerCamelCase : Any = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase__ ( self ): # This test is for deprecated behavior and can be removed in v5 lowerCamelCase : List[str] = BertConfig.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = AutoConfig.from_pretrained("""bert-base-cased""" ) lowerCamelCase : Optional[Any] = ["""config.4.0.0.json"""] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(__magic_name__ ) lowerCamelCase : str = 2 json.dump(configuration.to_dict() , open(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , """w""" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 lowerCamelCase : Any = ["""config.42.0.0.json"""] lowerCamelCase : Optional[Any] = 7_6_8 configuration.save_pretrained(__magic_name__ ) shutil.move(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , os.path.join(__magic_name__ , """config.42.0.0.json""" ) ) lowerCamelCase : int = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 7_6_8 ) def UpperCamelCase__ ( self ): # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. lowerCamelCase : str = """hf-internal-testing/test-two-configs""" import transformers as new_transformers lowerCamelCase : Tuple = """v4.0.0""" lowerCamelCase , lowerCamelCase : Optional[int] = new_transformers.models.auto.AutoConfig.from_pretrained( __magic_name__ , return_unused_kwargs=__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(__magic_name__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers lowerCamelCase : Tuple = """v3.0.0""" lowerCamelCase : Any = old_transformers.models.auto.AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(old_configuration.hidden_size , 7_6_8 )
287
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=__SCREAMING_SNAKE_CASE) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = field(default="""automatic-speech-recognition""" , metadata={"""include_in_asdict_even_if_is_default""": True}) _UpperCAmelCase : ClassVar[Features] = Features({"""audio""": Audio()}) _UpperCAmelCase : ClassVar[Features] = Features({"""transcription""": Value("""string""")}) _UpperCAmelCase : str = "audio" _UpperCAmelCase : str = "transcription" def UpperCamelCase__ ( self , __magic_name__ ): if self.audio_column not in features: raise ValueError(F'''Column {self.audio_column} is not present in features.''' ) if not isinstance(features[self.audio_column] , __magic_name__ ): raise ValueError(F'''Column {self.audio_column} is not an Audio type.''' ) lowerCamelCase : Optional[Any] = copy.deepcopy(self ) lowerCamelCase : List[Any] = self.input_schema.copy() lowerCamelCase : Tuple = features[self.audio_column] lowerCamelCase : int = input_schema return task_template @property def UpperCamelCase__ ( self ): return {self.audio_column: "audio", self.transcription_column: "transcription"}
287
1
import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html _lowerCamelCase ="""platform""" import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=None, lowerCamelCase=None, ): if attention_mask is None: lowerCamelCase : Union[str, Any] = np.where(input_ids != config.pad_token_id, 1, 0 ) if decoder_attention_mask is None: lowerCamelCase : Optional[Any] = np.where(decoder_input_ids != config.pad_token_id, 1, 0 ) if head_mask is None: lowerCamelCase : str = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCamelCase : Optional[Any] = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowerCamelCase : Tuple = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class A__ : def __init__( self , __magic_name__ , __magic_name__=1_3 , __magic_name__=7 , __magic_name__=True , __magic_name__=False , __magic_name__=9_9 , __magic_name__=1_6 , __magic_name__=2 , __magic_name__=4 , __magic_name__=4 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=3_2 , __magic_name__=2 , __magic_name__=1 , __magic_name__=0 , __magic_name__=0.02 , ): lowerCamelCase : Optional[Any] = parent lowerCamelCase : Tuple = batch_size lowerCamelCase : Dict = seq_length lowerCamelCase : List[Any] = is_training lowerCamelCase : Any = use_labels lowerCamelCase : Optional[int] = vocab_size lowerCamelCase : Dict = hidden_size lowerCamelCase : Optional[int] = num_hidden_layers lowerCamelCase : List[Any] = num_attention_heads lowerCamelCase : Optional[int] = intermediate_size lowerCamelCase : Optional[Any] = hidden_act lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : int = attention_probs_dropout_prob lowerCamelCase : List[Any] = max_position_embeddings lowerCamelCase : Any = eos_token_id lowerCamelCase : Any = pad_token_id lowerCamelCase : Tuple = bos_token_id lowerCamelCase : List[str] = initializer_range def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) lowerCamelCase : List[Any] = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) lowerCamelCase : Dict = shift_tokens_right(__magic_name__ , 1 , 2 ) lowerCamelCase : str = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=__magic_name__ , ) lowerCamelCase : List[str] = prepare_blenderbot_inputs_dict(__magic_name__ , __magic_name__ , __magic_name__ ) return config, inputs_dict def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Any = self.prepare_config_and_inputs() return config, inputs_dict def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Dict = 2_0 lowerCamelCase : Tuple = model_class_name(__magic_name__ ) lowerCamelCase : Union[str, Any] = model.encode(inputs_dict["""input_ids"""] ) lowerCamelCase , lowerCamelCase : List[str] = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) lowerCamelCase : int = model.init_cache(decoder_input_ids.shape[0] , __magic_name__ , __magic_name__ ) lowerCamelCase : Dict = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="""i4""" ) lowerCamelCase : int = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowerCamelCase : List[str] = model.decode( decoder_input_ids[:, :-1] , __magic_name__ , decoder_attention_mask=__magic_name__ , past_key_values=__magic_name__ , decoder_position_ids=__magic_name__ , ) lowerCamelCase : str = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) lowerCamelCase : Optional[int] = model.decode( decoder_input_ids[:, -1:] , __magic_name__ , decoder_attention_mask=__magic_name__ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=__magic_name__ , ) lowerCamelCase : int = model.decode(__magic_name__ , __magic_name__ ) lowerCamelCase : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=F'''Max diff is {diff}''' ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : str = 2_0 lowerCamelCase : Tuple = model_class_name(__magic_name__ ) lowerCamelCase : List[str] = model.encode(inputs_dict["""input_ids"""] ) lowerCamelCase , lowerCamelCase : int = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) lowerCamelCase : Dict = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) lowerCamelCase : int = model.init_cache(decoder_input_ids.shape[0] , __magic_name__ , __magic_name__ ) lowerCamelCase : Any = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowerCamelCase : List[str] = model.decode( decoder_input_ids[:, :-1] , __magic_name__ , decoder_attention_mask=__magic_name__ , past_key_values=__magic_name__ , decoder_position_ids=__magic_name__ , ) lowerCamelCase : Any = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) lowerCamelCase : Optional[Any] = model.decode( decoder_input_ids[:, -1:] , __magic_name__ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=__magic_name__ , decoder_position_ids=__magic_name__ , ) lowerCamelCase : str = model.decode(__magic_name__ , __magic_name__ , decoder_attention_mask=__magic_name__ ) lowerCamelCase : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=F'''Max diff is {diff}''' ) @require_flax class A__ ( unittest.TestCase): _UpperCAmelCase : str = 99 def UpperCamelCase__ ( self ): lowerCamelCase : Dict = np.array( [ [7_1, 8_2, 1_8, 3_3, 4_6, 9_1, 2], [6_8, 3_4, 2_6, 5_8, 3_0, 8_2, 2], [5, 9_7, 1_7, 3_9, 9_4, 4_0, 2], [7_6, 8_3, 9_4, 2_5, 7_0, 7_8, 2], [8_7, 5_9, 4_1, 3_5, 4_8, 6_6, 2], [5_5, 1_3, 1_6, 5_8, 5, 2, 1], # note padding [6_4, 2_7, 3_1, 5_1, 1_2, 7_5, 2], [5_2, 6_4, 8_6, 1_7, 8_3, 3_9, 2], [4_8, 6_1, 9, 2_4, 7_1, 8_2, 2], [2_6, 1, 6_0, 4_8, 2_2, 1_3, 2], [2_1, 5, 6_2, 2_8, 1_4, 7_6, 2], [4_5, 9_8, 3_7, 8_6, 5_9, 4_8, 2], [7_0, 7_0, 5_0, 9, 2_8, 0, 2], ] , dtype=np.intaa , ) lowerCamelCase : Dict = input_ids.shape[0] lowerCamelCase : Optional[Any] = BlenderbotConfig( vocab_size=self.vocab_size , d_model=2_4 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=3_2 , decoder_ffn_dim=3_2 , max_position_embeddings=4_8 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase : int = self._get_config_and_data() lowerCamelCase : Optional[Any] = FlaxBlenderbotForConditionalGeneration(__magic_name__ ) lowerCamelCase : List[Any] = lm_model(input_ids=__magic_name__ ) lowerCamelCase : Optional[Any] = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["""logits"""].shape , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = BlenderbotConfig( vocab_size=self.vocab_size , d_model=1_4 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=4_8 , ) lowerCamelCase : Union[str, Any] = FlaxBlenderbotForConditionalGeneration(__magic_name__ ) lowerCamelCase : str = np.array([[7_1, 8_2, 1_8, 3_3, 4_6, 9_1, 2], [6_8, 3_4, 2_6, 5_8, 3_0, 2, 1]] , dtype=np.intaa ) lowerCamelCase : int = np.array([[8_2, 7_1, 8_2, 1_8, 2], [5_8, 6_8, 2, 1, 1]] , dtype=np.intaa ) lowerCamelCase : Optional[int] = lm_model(input_ids=__magic_name__ , decoder_input_ids=__magic_name__ ) lowerCamelCase : Tuple = (*summary.shape, config.vocab_size) self.assertEqual(outputs["""logits"""].shape , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = np.array([[7_1, 8_2, 1_8, 3_3, 2, 1, 1], [6_8, 3_4, 2_6, 5_8, 3_0, 8_2, 2]] , dtype=np.intaa ) lowerCamelCase : Any = shift_tokens_right(__magic_name__ , 1 , 2 ) lowerCamelCase : Optional[Any] = np.equal(__magic_name__ , 1 ).astype(np.floataa ).sum() lowerCamelCase : int = np.equal(__magic_name__ , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(__magic_name__ , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase , __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[Any] = True _UpperCAmelCase : Union[str, Any] = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) _UpperCAmelCase : Optional[Any] = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = FlaxBlenderbotModelTester(self ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(__magic_name__ , __magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(__magic_name__ , __magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowerCamelCase : str = self._prepare_for_class(__magic_name__ , __magic_name__ ) lowerCamelCase : Union[str, Any] = model_class(__magic_name__ ) @jax.jit def encode_jitted(__magic_name__ , __magic_name__=None , **__magic_name__ ): return model.encode(input_ids=__magic_name__ , attention_mask=__magic_name__ ) with self.subTest("""JIT Enabled""" ): lowerCamelCase : Union[str, Any] = encode_jitted(**__magic_name__ ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): lowerCamelCase : List[str] = encode_jitted(**__magic_name__ ).to_tuple() self.assertEqual(len(__magic_name__ ) , len(__magic_name__ ) ) for jitted_output, output in zip(__magic_name__ , __magic_name__ ): self.assertEqual(jitted_output.shape , output.shape ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowerCamelCase : Any = model_class(__magic_name__ ) lowerCamelCase : Dict = model.encode(inputs_dict["""input_ids"""] , inputs_dict["""attention_mask"""] ) lowerCamelCase : Optional[int] = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(__magic_name__ , __magic_name__ , __magic_name__ ): return model.decode( decoder_input_ids=__magic_name__ , decoder_attention_mask=__magic_name__ , encoder_outputs=__magic_name__ , ) with self.subTest("""JIT Enabled""" ): lowerCamelCase : Union[str, Any] = decode_jitted(**__magic_name__ ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): lowerCamelCase : Union[str, Any] = decode_jitted(**__magic_name__ ).to_tuple() self.assertEqual(len(__magic_name__ ) , len(__magic_name__ ) ) for jitted_output, output in zip(__magic_name__ , __magic_name__ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def UpperCamelCase__ ( self ): for model_class_name in self.all_model_classes: lowerCamelCase : Union[str, Any] = model_class_name.from_pretrained("""facebook/blenderbot-400M-distill""" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids lowerCamelCase : Any = np.ones((1, 1) ) * model.config.eos_token_id lowerCamelCase : Tuple = model(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) @unittest.skipUnless(jax_device != """cpu""" , """3B test too slow on CPU.""" ) @slow def UpperCamelCase__ ( self ): lowerCamelCase : Any = {"""num_beams""": 1, """early_stopping""": True, """min_length""": 1_5, """max_length""": 2_5} lowerCamelCase : Tuple = {"""skip_special_tokens""": True, """clean_up_tokenization_spaces""": True} lowerCamelCase : Dict = FlaxBlenderbotForConditionalGeneration.from_pretrained("""facebook/blenderbot-3B""" , from_pt=__magic_name__ ) lowerCamelCase : List[Any] = BlenderbotTokenizer.from_pretrained("""facebook/blenderbot-3B""" ) lowerCamelCase : Union[str, Any] = ["""Sam"""] lowerCamelCase : Optional[Any] = tokenizer(__magic_name__ , return_tensors="""jax""" ) lowerCamelCase : Optional[Any] = model.generate(**__magic_name__ , **__magic_name__ ) lowerCamelCase : Optional[int] = """Sam is a great name. It means \"sun\" in Gaelic.""" lowerCamelCase : Tuple = tokenizer.batch_decode(__magic_name__ , **__magic_name__ ) assert generated_txt[0].strip() == tgt_text
287
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower_vision_model""" def __init__( self , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=3 , __magic_name__=1_6 , __magic_name__=2_8_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__=True , __magic_name__=False , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : str = num_hidden_layers lowerCamelCase : Optional[int] = num_channels lowerCamelCase : List[str] = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Any = initializer_factor lowerCamelCase : Tuple = layer_norm_eps lowerCamelCase : Tuple = stop_gradient lowerCamelCase : Optional[int] = share_layernorm lowerCamelCase : str = remove_last_layer @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : str = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """bridgetower_text_model""" def __init__( self , __magic_name__=5_0_2_6_5 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=1 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_4 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Any = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : Tuple = hidden_act lowerCamelCase : Optional[int] = initializer_factor lowerCamelCase : Any = intermediate_size lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : str = max_position_embeddings lowerCamelCase : Union[str, Any] = type_vocab_size lowerCamelCase : Optional[int] = layer_norm_eps lowerCamelCase : Optional[int] = position_embedding_type lowerCamelCase : List[str] = use_cache lowerCamelCase : List[str] = pad_token_id lowerCamelCase : List[str] = bos_token_id lowerCamelCase : Optional[int] = eos_token_id @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : Optional[int] = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower""" def __init__( self , __magic_name__=True , __magic_name__="gelu" , __magic_name__=7_6_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__="add" , __magic_name__=1_2 , __magic_name__=6 , __magic_name__=False , __magic_name__=False , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): # TODO: remove this once the Hub files are updated. lowerCamelCase : int = kwargs.pop("""text_config_dict""" , __magic_name__ ) lowerCamelCase : str = kwargs.pop("""vision_config_dict""" , __magic_name__ ) super().__init__(**__magic_name__ ) lowerCamelCase : str = share_cross_modal_transformer_layers lowerCamelCase : Union[str, Any] = hidden_act lowerCamelCase : str = hidden_size lowerCamelCase : Tuple = initializer_factor lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = share_link_tower_layers lowerCamelCase : List[Any] = link_tower_type lowerCamelCase : Tuple = num_attention_heads lowerCamelCase : int = num_hidden_layers lowerCamelCase : Union[str, Any] = tie_word_embeddings lowerCamelCase : Tuple = init_layernorm_from_vision_encoder if text_config is None: lowerCamelCase : Any = {} logger.info("""`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.""" ) if vision_config is None: lowerCamelCase : int = {} logger.info("""`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.""" ) lowerCamelCase : Any = BridgeTowerTextConfig(**__magic_name__ ) lowerCamelCase : Optional[Any] = BridgeTowerVisionConfig(**__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , **__magic_name__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = copy.deepcopy(self.__dict__ ) lowerCamelCase : int = self.text_config.to_dict() lowerCamelCase : Dict = self.vision_config.to_dict() lowerCamelCase : List[str] = self.__class__.model_type return output
287
1
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """ClapFeatureExtractor""" _UpperCAmelCase : List[str] = ("""RobertaTokenizer""", """RobertaTokenizerFast""") def __init__( self , __magic_name__ , __magic_name__ ): super().__init__(__magic_name__ , __magic_name__ ) def __call__( self , __magic_name__=None , __magic_name__=None , __magic_name__=None , **__magic_name__ ): lowerCamelCase : List[Any] = kwargs.pop("""sampling_rate""" , __magic_name__ ) if text is None and audios is None: raise ValueError("""You have to specify either text or audios. Both cannot be none.""" ) if text is not None: lowerCamelCase : Union[str, Any] = self.tokenizer(__magic_name__ , return_tensors=__magic_name__ , **__magic_name__ ) if audios is not None: lowerCamelCase : List[Any] = self.feature_extractor( __magic_name__ , sampling_rate=__magic_name__ , return_tensors=__magic_name__ , **__magic_name__ ) if text is not None and audios is not None: lowerCamelCase : List[str] = audio_features.input_features return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__magic_name__ ) , tensor_type=__magic_name__ ) def UpperCamelCase__ ( self , *__magic_name__ , **__magic_name__ ): return self.tokenizer.batch_decode(*__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , *__magic_name__ , **__magic_name__ ): return self.tokenizer.decode(*__magic_name__ , **__magic_name__ ) @property def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = self.tokenizer.model_input_names lowerCamelCase : Any = self.feature_extractor.model_input_names return list(dict.fromkeys(tokenizer_input_names + feature_extractor_input_names ) )
287
def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = set(range(3, lowerCamelCase, 2 ) ) primes.add(2 ) for p in range(3, lowerCamelCase, 2 ): if p not in primes: continue primes.difference_update(set(range(p * p, lowerCamelCase, lowerCamelCase ) ) ) lowerCamelCase : Any = [float(lowerCamelCase ) for n in range(limit + 1 )] for p in primes: for n in range(lowerCamelCase, limit + 1, lowerCamelCase ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'''{solution() = }''')
287
1
import re from filelock import FileLock try: import nltk _lowerCamelCase =True except (ImportError, ModuleNotFoundError): _lowerCamelCase =False if NLTK_AVAILABLE: with FileLock(""".lock""") as lock: nltk.download("""punkt""", quiet=True) def _a ( lowerCamelCase ): re.sub("""<n>""", """""", lowerCamelCase ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(lowerCamelCase ) )
287
from __future__ import annotations import collections import pprint from pathlib import Path def _a ( lowerCamelCase ): return "".join(sorted(lowerCamelCase ) ) def _a ( lowerCamelCase ): return word_by_signature[signature(lowerCamelCase )] _lowerCamelCase =Path(__file__).parent.joinpath("""words.txt""").read_text(encoding="""utf-8""") _lowerCamelCase =sorted({word.strip().lower() for word in data.splitlines()}) _lowerCamelCase =collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _lowerCamelCase ={word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open("""anagrams.txt""", """w""") as file: file.write("""all_anagrams = \n """) file.write(pprint.pformat(all_anagrams))
287
1
def _a ( lowerCamelCase ): # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("""The given input must be positive""" ) # get the generated string sequence lowerCamelCase : List[Any] = gray_code_sequence_string(lowerCamelCase ) # # convert them to integers for i in range(len(lowerCamelCase ) ): lowerCamelCase : str = int(sequence[i], 2 ) return sequence def _a ( lowerCamelCase ): # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] lowerCamelCase : Dict = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits lowerCamelCase : Dict = gray_code_sequence_string(bit_count - 1 ) lowerCamelCase : Dict = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): lowerCamelCase : List[Any] = """0""" + smaller_sequence[i] sequence.append(lowerCamelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): lowerCamelCase : Dict = """1""" + smaller_sequence[i] sequence.append(lowerCamelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
287
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = ["""pixel_values"""] def __init__( self , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , __magic_name__ = PILImageResampling.BILINEAR , __magic_name__ = True , __magic_name__ = 1 / 2_5_5 , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = size if size is not None else {"""shortest_edge""": 3_8_4} lowerCamelCase : Tuple = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Dict = do_resize lowerCamelCase : List[Any] = size # Default value set here for backwards compatibility where the value in config is None lowerCamelCase : Any = crop_pct if crop_pct is not None else 2_2_4 / 2_5_6 lowerCamelCase : Union[str, Any] = resample lowerCamelCase : str = do_rescale lowerCamelCase : Union[str, Any] = rescale_factor lowerCamelCase : Tuple = do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCamelCase : int = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = PILImageResampling.BICUBIC , __magic_name__ = None , **__magic_name__ , ): lowerCamelCase : Union[str, Any] = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) lowerCamelCase : str = size["""shortest_edge"""] if shortest_edge < 3_8_4: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct lowerCamelCase : List[str] = int(shortest_edge / crop_pct ) lowerCamelCase : Optional[Any] = get_resize_output_image_size(__magic_name__ , size=__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Optional[int] = resize(image=__magic_name__ , size=__magic_name__ , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=__magic_name__ , size=(shortest_edge, shortest_edge) , data_format=__magic_name__ , **__magic_name__ ) else: # warping (no cropping) when evaluated at 384 or larger return resize( __magic_name__ , size=(shortest_edge, shortest_edge) , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return rescale(__magic_name__ , scale=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return normalize(__magic_name__ , mean=__magic_name__ , std=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = ChannelDimension.FIRST , **__magic_name__ , ): lowerCamelCase : str = do_resize if do_resize is not None else self.do_resize lowerCamelCase : Optional[Any] = crop_pct if crop_pct is not None else self.crop_pct lowerCamelCase : Optional[int] = resample if resample is not None else self.resample lowerCamelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else self.image_mean lowerCamelCase : Tuple = image_std if image_std is not None else self.image_std lowerCamelCase : Dict = size if size is not None else self.size lowerCamelCase : str = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : List[str] = make_list_of_images(__magic_name__ ) if not valid_images(__magic_name__ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_resize and size["shortest_edge"] < 3_8_4 and crop_pct is None: raise ValueError("""crop_pct must be specified if size < 384.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. lowerCamelCase : Optional[Any] = [to_numpy_array(__magic_name__ ) for image in images] if do_resize: lowerCamelCase : List[Any] = [self.resize(image=__magic_name__ , size=__magic_name__ , crop_pct=__magic_name__ , resample=__magic_name__ ) for image in images] if do_rescale: lowerCamelCase : Union[str, Any] = [self.rescale(image=__magic_name__ , scale=__magic_name__ ) for image in images] if do_normalize: lowerCamelCase : List[Any] = [self.normalize(image=__magic_name__ , mean=__magic_name__ , std=__magic_name__ ) for image in images] lowerCamelCase : Optional[int] = [to_channel_dimension_format(__magic_name__ , __magic_name__ ) for image in images] lowerCamelCase : List[str] = {"""pixel_values""": images} return BatchFeature(data=__magic_name__ , tensor_type=__magic_name__ )
287
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """alibaba-damo/mgp-str-base""": """https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json""", } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = """mgp-str""" def __init__( self , __magic_name__=[3_2, 1_2_8] , __magic_name__=4 , __magic_name__=3 , __magic_name__=2_7 , __magic_name__=3_8 , __magic_name__=5_0_2_5_7 , __magic_name__=3_0_5_2_2 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=4.0 , __magic_name__=True , __magic_name__=False , __magic_name__=1e-5 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=False , __magic_name__=0.02 , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Tuple = image_size lowerCamelCase : List[Any] = patch_size lowerCamelCase : List[Any] = num_channels lowerCamelCase : Dict = max_token_length lowerCamelCase : Any = num_character_labels lowerCamelCase : str = num_bpe_labels lowerCamelCase : Dict = num_wordpiece_labels lowerCamelCase : int = hidden_size lowerCamelCase : List[str] = num_hidden_layers lowerCamelCase : Any = num_attention_heads lowerCamelCase : Union[str, Any] = mlp_ratio lowerCamelCase : Optional[Any] = distilled lowerCamelCase : Tuple = layer_norm_eps lowerCamelCase : Optional[int] = drop_rate lowerCamelCase : List[str] = qkv_bias lowerCamelCase : Dict = attn_drop_rate lowerCamelCase : List[Any] = drop_path_rate lowerCamelCase : List[str] = output_aa_attentions lowerCamelCase : Tuple = initializer_range
287
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 _lowerCamelCase ={ """return_dict""": False, """output_hidden_states""": True, """output_attentions""": True, """torchscript""": True, """torch_dtype""": """float16""", """use_bfloat16""": True, """tf_legacy_loss""": True, """pruned_heads""": {"""a""": 1}, """tie_word_embeddings""": False, """is_decoder""": True, """cross_attention_hidden_size""": 1_2_8, """add_cross_attention""": True, """tie_encoder_decoder""": True, """max_length""": 5_0, """min_length""": 3, """do_sample""": True, """early_stopping""": True, """num_beams""": 3, """num_beam_groups""": 3, """diversity_penalty""": 0.5, """temperature""": 2.0, """top_k""": 1_0, """top_p""": 0.7, """typical_p""": 0.2, """repetition_penalty""": 0.8, """length_penalty""": 0.8, """no_repeat_ngram_size""": 5, """encoder_no_repeat_ngram_size""": 5, """bad_words_ids""": [1, 2, 3], """num_return_sequences""": 3, """chunk_size_feed_forward""": 5, """output_scores""": True, """return_dict_in_generate""": True, """forced_bos_token_id""": 2, """forced_eos_token_id""": 3, """remove_invalid_values""": True, """architectures""": ["""BertModel"""], """finetuning_task""": """translation""", """id2label""": {0: """label"""}, """label2id""": {"""label""": """0"""}, """tokenizer_class""": """BertTokenizerFast""", """prefix""": """prefix""", """bos_token_id""": 6, """pad_token_id""": 7, """eos_token_id""": 8, """sep_token_id""": 9, """decoder_start_token_id""": 1_0, """exponential_decay_length_penalty""": (5, 1.01), """suppress_tokens""": [0, 1], """begin_suppress_tokens""": 2, """task_specific_params""": {"""translation""": """some_params"""}, """problem_type""": """regression""", } @is_staging_test class A__ ( unittest.TestCase): @classmethod def UpperCamelCase__ ( cls ): lowerCamelCase : int = TOKEN HfFolder.save_token(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls ): try: delete_repo(token=cls._token , repo_id="""test-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-config-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-config""" ) except HTTPError: pass def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""test-config""" , use_auth_token=self._token ) lowerCamelCase : Any = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__magic_name__ , repo_id="""test-config""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : Optional[Any] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""valid_org/test-config-org""" , use_auth_token=self._token ) lowerCamelCase : Optional[int] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( __magic_name__ , repo_id="""valid_org/test-config-org""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : List[str] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): CustomConfig.register_for_auto_class() lowerCamelCase : Optional[Any] = CustomConfig(attribute=4_2 ) config.push_to_hub("""test-dynamic-config""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"""AutoConfig""": """custom_configuration.CustomConfig"""} ) lowerCamelCase : List[str] = AutoConfig.from_pretrained(F'''{USER}/test-dynamic-config''' , trust_remote_code=__magic_name__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , """CustomConfig""" ) self.assertEqual(new_config.attribute , 4_2 ) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : str = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated lowerCamelCase : Optional[int] = c.n_embd + 1 # int lowerCamelCase : Optional[int] = c.resid_pdrop + 1.0 # float lowerCamelCase : Tuple = not c.scale_attn_weights # bool lowerCamelCase : Any = c.summary_type + """foo""" # str c.update_from_string( F'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(__magic_name__ , c.n_embd , """mismatch for key: n_embd""" ) self.assertEqual(__magic_name__ , c.resid_pdrop , """mismatch for key: resid_pdrop""" ) self.assertEqual(__magic_name__ , c.scale_attn_weights , """mismatch for key: scale_attn_weights""" ) self.assertEqual(__magic_name__ , c.summary_type , """mismatch for key: summary_type""" ) def UpperCamelCase__ ( self ): lowerCamelCase : str = PretrainedConfig() lowerCamelCase : int = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( __magic_name__ , ["""is_encoder_decoder""", """_name_or_path""", """_commit_hash""", """transformers_version"""] ) lowerCamelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(__magic_name__ , __magic_name__ )] if len(__magic_name__ ) > 0: raise ValueError( """The following keys are set with the default values in""" """ `test_configuration_common.config_common_kwargs` pick another value for them:""" F''' {", ".join(__magic_name__ )}.''' ) def UpperCamelCase__ ( self ): with self.assertRaises(__magic_name__ ): # config is in subfolder, the following should not work without specifying the subfolder lowerCamelCase : Dict = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" ) lowerCamelCase : str = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" , subfolder="""bert""" ) self.assertIsNotNone(__magic_name__ ) def UpperCamelCase__ ( self ): # A mock response for an HTTP head request to emulate server down lowerCamelCase : Dict = mock.Mock() lowerCamelCase : Optional[int] = 5_0_0 lowerCamelCase : List[Any] = {} lowerCamelCase : Tuple = HTTPError lowerCamelCase : Union[str, Any] = {} # Download this model to make sure it's in the cache. lowerCamelCase : List[str] = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=__magic_name__ ) as mock_head: lowerCamelCase : Any = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase__ ( self ): # This test is for deprecated behavior and can be removed in v5 lowerCamelCase : List[str] = BertConfig.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = AutoConfig.from_pretrained("""bert-base-cased""" ) lowerCamelCase : Optional[Any] = ["""config.4.0.0.json"""] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(__magic_name__ ) lowerCamelCase : str = 2 json.dump(configuration.to_dict() , open(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , """w""" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 lowerCamelCase : Any = ["""config.42.0.0.json"""] lowerCamelCase : Optional[Any] = 7_6_8 configuration.save_pretrained(__magic_name__ ) shutil.move(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , os.path.join(__magic_name__ , """config.42.0.0.json""" ) ) lowerCamelCase : int = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 7_6_8 ) def UpperCamelCase__ ( self ): # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. lowerCamelCase : str = """hf-internal-testing/test-two-configs""" import transformers as new_transformers lowerCamelCase : Tuple = """v4.0.0""" lowerCamelCase , lowerCamelCase : Optional[int] = new_transformers.models.auto.AutoConfig.from_pretrained( __magic_name__ , return_unused_kwargs=__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(__magic_name__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers lowerCamelCase : Tuple = """v3.0.0""" lowerCamelCase : Any = old_transformers.models.auto.AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(old_configuration.hidden_size , 7_6_8 )
287
1
from __future__ import annotations import requests _lowerCamelCase =set( """approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports""".split() ) def _a ( lowerCamelCase, lowerCamelCase = 1, lowerCamelCase = "new", lowerCamelCase = None ): lowerCamelCase : Tuple = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(lowerCamelCase ) - valid_terms ) ): lowerCamelCase : Tuple = F'''Invalid search term: {invalid_search_terms}''' raise ValueError(lowerCamelCase ) lowerCamelCase : Tuple = requests.get( F'''https://reddit.com/r/{subreddit}/{age}.json?limit={limit}''', headers={"""User-agent""": """A random string"""}, ) if response.status_code == 429: raise requests.HTTPError lowerCamelCase : List[Any] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(lowerCamelCase )} lowerCamelCase : Dict = {} for id_ in range(lowerCamelCase ): lowerCamelCase : Dict = { item: data["""data"""]["""children"""][id_]["""data"""][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("""learnpython""", wanted_data=["""title""", """url""", """selftext"""]))
287
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) lowerCamelCase : Any = tf.convert_to_tensor( [[5, 1_2_1, 1_1, 6_6_0, 1_6, 7_3_0, 2_5_5_4_3, 1_1_0, 8_3, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowerCamelCase : str = model(__magic_name__ )["""last_hidden_state"""] lowerCamelCase : Union[str, Any] = tf.TensorShape((1, 1_0, 7_6_8) ) self.assertEqual(output.shape , __magic_name__ ) # compare the actual values for a slice. lowerCamelCase : Dict = tf.convert_to_tensor( [[[-0.0_254, 0.0_235, 0.1_027], [0.0_606, -0.1_811, -0.0_418], [-0.1_561, -0.1_127, 0.2_687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
287
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase ={ """configuration_mgp_str""": ["""MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MgpstrConfig"""], """processing_mgp_str""": ["""MgpstrProcessor"""], """tokenization_mgp_str""": ["""MgpstrTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =[ """MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST""", """MgpstrModel""", """MgpstrPreTrainedModel""", """MgpstrForSceneTextRecognition""", ] if TYPE_CHECKING: from .configuration_mgp_str import MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP, MgpstrConfig from .processing_mgp_str import MgpstrProcessor from .tokenization_mgp_str import MgpstrTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mgp_str import ( MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST, MgpstrForSceneTextRecognition, MgpstrModel, MgpstrPreTrainedModel, ) else: import sys _lowerCamelCase =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
287
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _lowerCamelCase =get_logger(__name__) class A__ : def __init__( self , __magic_name__ = None ): lowerCamelCase : Dict = ( os.path.join(__magic_name__ , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) lowerCamelCase : List[str] = Extractor def UpperCamelCase__ ( self , __magic_name__ ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" lowerCamelCase : int = os.path.abspath(__magic_name__ ) return os.path.join(self.extract_dir , hash_url_to_filename(__magic_name__ ) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): return force_extract or ( not os.path.isfile(__magic_name__ ) and not (os.path.isdir(__magic_name__ ) and os.listdir(__magic_name__ )) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = False ): lowerCamelCase : Union[str, Any] = self.extractor.infer_extractor_format(__magic_name__ ) if not extractor_format: return input_path lowerCamelCase : int = self._get_output_path(__magic_name__ ) if self._do_extract(__magic_name__ , __magic_name__ ): self.extractor.extract(__magic_name__ , __magic_name__ , __magic_name__ ) return output_path class A__ ( __SCREAMING_SNAKE_CASE): @classmethod @abstractmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): ... @staticmethod @abstractmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): ... class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[bytes] = [] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with open(__magic_name__ , """rb""" ) as f: return f.read(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if not magic_number: lowerCamelCase : Optional[Any] = max(len(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) try: lowerCamelCase : Tuple = cls.read_magic_number(__magic_name__ , __magic_name__ ) except OSError: return False return any(magic_number.startswith(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) class A__ ( __SCREAMING_SNAKE_CASE): @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): return tarfile.is_tarfile(__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): def resolved(__magic_name__ ) -> str: return os.path.realpath(os.path.abspath(__magic_name__ ) ) def badpath(__magic_name__ , __magic_name__ ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__magic_name__ , __magic_name__ ) ).startswith(__magic_name__ ) def badlink(__magic_name__ , __magic_name__ ) -> bool: # Links are interpreted relative to the directory containing the link lowerCamelCase : List[str] = resolved(os.path.join(__magic_name__ , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=__magic_name__ ) lowerCamelCase : Optional[Any] = resolved(__magic_name__ ) for finfo in members: if badpath(finfo.name , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Dict = tarfile.open(__magic_name__ ) tar_file.extractall(__magic_name__ , members=TarExtractor.safemembers(__magic_name__ , __magic_name__ ) ) tar_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = [B"""\x1F\x8B"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with gzip.open(__magic_name__ , """rb""" ) as gzip_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[int] = [ B"""PK\x03\x04""", B"""PK\x05\x06""", # empty archive B"""PK\x07\x08""", # spanned archive ] @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if super().is_extractable(__magic_name__ , magic_number=__magic_name__ ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__magic_name__ , """rb""" ) as fp: lowerCamelCase : List[str] = _EndRecData(__magic_name__ ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: lowerCamelCase : List[Any] = fp.read(__magic_name__ ) # CD is where we expect it to be if len(__magic_name__ ) == sizeCentralDir: lowerCamelCase : str = struct.unpack(__magic_name__ , __magic_name__ ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with zipfile.ZipFile(__magic_name__ , """r""" ) as zip_file: zip_file.extractall(__magic_name__ ) zip_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[str] = [B"""\xFD\x37\x7A\x58\x5A\x00"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with lzma.open(__magic_name__ ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""Rar!\x1a\x07\x00""", B"""Rar!\x1a\x07\x01\x00"""] # RAR_ID # RAR5_ID @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Union[str, Any] = rarfile.RarFile(__magic_name__ ) rf.extractall(__magic_name__ ) rf.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = [B"""\x28\xb5\x2F\xFD"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd lowerCamelCase : int = zstd.ZstdDecompressor() with open(__magic_name__ , """rb""" ) as ifh, open(__magic_name__ , """wb""" ) as ofh: dctx.copy_stream(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""\x42\x5A\x68"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with bza.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x37\x7A\xBC\xAF\x27\x1C"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with pyazr.SevenZipFile(__magic_name__ , """r""" ) as archive: archive.extractall(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x04\x22\x4D\x18"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) _UpperCAmelCase : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def UpperCamelCase__ ( cls ): return max( len(__magic_name__ ) for extractor in cls.extractors.values() if issubclass(__magic_name__ , __magic_name__ ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): try: return MagicNumberBaseExtractor.read_magic_number(__magic_name__ , magic_number_length=__magic_name__ ) except OSError: return b"" @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = False ): warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = cls.infer_extractor_format(__magic_name__ ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): # <Added version="2.4.0"/> lowerCamelCase : Dict = cls._get_magic_number_max_length() lowerCamelCase : Optional[Any] = cls._read_magic_number(__magic_name__ , __magic_name__ ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__magic_name__ , magic_number=__magic_name__ ): return extractor_format @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = "deprecated" , ): os.makedirs(os.path.dirname(__magic_name__ ) , exist_ok=__magic_name__ ) # Prevent parallel extractions lowerCamelCase : Tuple = str(Path(__magic_name__ ).with_suffix(""".lock""" ) ) with FileLock(__magic_name__ ): shutil.rmtree(__magic_name__ , ignore_errors=__magic_name__ ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__magic_name__ , __magic_name__ ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = extractor if extractor != """deprecated""" else extractor_format else: lowerCamelCase : Optional[int] = cls.extractors[extractor_format] return extractor.extract(__magic_name__ , __magic_name__ ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=__magic_name__ , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__magic_name__ ): return extractor.extract(__magic_name__ , __magic_name__ )
287
1
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase): _UpperCAmelCase : Union[str, Any] = ShapEPipeline _UpperCAmelCase : str = ["""prompt"""] _UpperCAmelCase : Dict = ["""prompt"""] _UpperCAmelCase : Tuple = [ """num_images_per_prompt""", """num_inference_steps""", """generator""", """latents""", """guidance_scale""", """frame_size""", """output_type""", """return_dict""", ] _UpperCAmelCase : List[str] = False @property def UpperCamelCase__ ( self ): return 3_2 @property def UpperCamelCase__ ( self ): return 3_2 @property def UpperCamelCase__ ( self ): return self.time_input_dim * 4 @property def UpperCamelCase__ ( self ): return 8 @property def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) return tokenizer @property def UpperCamelCase__ ( self ): torch.manual_seed(0 ) lowerCamelCase : List[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(__magic_name__ ) @property def UpperCamelCase__ ( self ): torch.manual_seed(0 ) lowerCamelCase : Optional[int] = { """num_attention_heads""": 2, """attention_head_dim""": 1_6, """embedding_dim""": self.time_input_dim, """num_embeddings""": 3_2, """embedding_proj_dim""": self.text_embedder_hidden_size, """time_embed_dim""": self.time_embed_dim, """num_layers""": 1, """clip_embed_dim""": self.time_input_dim * 2, """additional_embeddings""": 0, """time_embed_act_fn""": """gelu""", """norm_in_type""": """layer""", """encoder_hid_proj_type""": None, """added_emb_type""": None, } lowerCamelCase : Dict = PriorTransformer(**__magic_name__ ) return model @property def UpperCamelCase__ ( self ): torch.manual_seed(0 ) lowerCamelCase : str = { """param_shapes""": ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), """d_latent""": self.time_input_dim, """d_hidden""": self.renderer_dim, """n_output""": 1_2, """background""": ( 0.1, 0.1, 0.1, ), } lowerCamelCase : List[Any] = ShapERenderer(**__magic_name__ ) return model def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = self.dummy_prior lowerCamelCase : str = self.dummy_text_encoder lowerCamelCase : int = self.dummy_tokenizer lowerCamelCase : int = self.dummy_renderer lowerCamelCase : Tuple = HeunDiscreteScheduler( beta_schedule="""exp""" , num_train_timesteps=1_0_2_4 , prediction_type="""sample""" , use_karras_sigmas=__magic_name__ , clip_sample=__magic_name__ , clip_sample_range=1.0 , ) lowerCamelCase : Optional[int] = { """prior""": prior, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """renderer""": renderer, """scheduler""": scheduler, } return components def UpperCamelCase__ ( self , __magic_name__ , __magic_name__=0 ): if str(__magic_name__ ).startswith("""mps""" ): lowerCamelCase : List[str] = torch.manual_seed(__magic_name__ ) else: lowerCamelCase : List[Any] = torch.Generator(device=__magic_name__ ).manual_seed(__magic_name__ ) lowerCamelCase : int = { """prompt""": """horse""", """generator""": generator, """num_inference_steps""": 1, """frame_size""": 3_2, """output_type""": """np""", } return inputs def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = """cpu""" lowerCamelCase : Union[str, Any] = self.get_dummy_components() lowerCamelCase : Optional[Any] = self.pipeline_class(**__magic_name__ ) lowerCamelCase : List[str] = pipe.to(__magic_name__ ) pipe.set_progress_bar_config(disable=__magic_name__ ) lowerCamelCase : Optional[Any] = pipe(**self.get_dummy_inputs(__magic_name__ ) ) lowerCamelCase : Optional[Any] = output.images[0] lowerCamelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) lowerCamelCase : List[str] = np.array( [ 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase__ ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def UpperCamelCase__ ( self ): lowerCamelCase : str = torch_device == """cpu""" lowerCamelCase : str = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=__magic_name__ , relax_max_difference=__magic_name__ , ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = self.get_dummy_components() lowerCamelCase : int = self.pipeline_class(**__magic_name__ ) lowerCamelCase : Optional[Any] = pipe.to(__magic_name__ ) pipe.set_progress_bar_config(disable=__magic_name__ ) lowerCamelCase : Optional[int] = 1 lowerCamelCase : Dict = 2 lowerCamelCase : Optional[Any] = self.get_dummy_inputs(__magic_name__ ) for key in inputs.keys(): if key in self.batch_params: lowerCamelCase : List[Any] = batch_size * [inputs[key]] lowerCamelCase : int = pipe(**__magic_name__ , num_images_per_prompt=__magic_name__ )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/shap_e/test_shap_e_np_out.npy""" ) lowerCamelCase : List[str] = ShapEPipeline.from_pretrained("""openai/shap-e""" ) lowerCamelCase : str = pipe.to(__magic_name__ ) pipe.set_progress_bar_config(disable=__magic_name__ ) lowerCamelCase : List[Any] = torch.Generator(device=__magic_name__ ).manual_seed(0 ) lowerCamelCase : Optional[int] = pipe( """a shark""" , generator=__magic_name__ , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type="""np""" , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(__magic_name__ , __magic_name__ )
287
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration _lowerCamelCase =5_0_0_0_0_0 _lowerCamelCase , _lowerCamelCase =os.path.split(__file__) _lowerCamelCase =os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.map(**lowerCamelCase ) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.filter(**lowerCamelCase ) def _a ( ): lowerCamelCase : Optional[Any] = {"""num examples""": SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : Any = datasets.Features({"""text""": datasets.Value("""string""" ), """numbers""": datasets.Value("""float32""" )} ) lowerCamelCase : Tuple = generate_example_dataset( os.path.join(lowerCamelCase, """dataset.arrow""" ), lowerCamelCase, num_examples=lowerCamelCase ) lowerCamelCase : Tuple = transformers.AutoTokenizer.from_pretrained("""bert-base-cased""", use_fast=lowerCamelCase ) def tokenize(lowerCamelCase ): return tokenizer(examples["""text"""] ) lowerCamelCase : List[str] = map(lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""numpy""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""pandas""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""torch""", columns="""numbers""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""tensorflow""", columns="""numbers""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : Union[str, Any] = filter(lowerCamelCase ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(lowerCamelCase, """wb""" ) as f: f.write(json.dumps(lowerCamelCase ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
287
1
import mpmath # for roots of unity import numpy as np class A__ : def __init__( self , __magic_name__=None , __magic_name__=None ): # Input as list lowerCamelCase : str = list(poly_a or [0] )[:] lowerCamelCase : Union[str, Any] = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() lowerCamelCase : List[Any] = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() lowerCamelCase : Union[str, Any] = len(self.polyB ) # Add 0 to make lengths equal a power of 2 lowerCamelCase : List[str] = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform lowerCamelCase : Optional[int] = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product lowerCamelCase : Union[str, Any] = self.__multiply() def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : Optional[int] = [[x] for x in self.polyA] if which == """A""" else [[x] for x in self.polyB] # Corner case if len(__magic_name__ ) <= 1: return dft[0] # lowerCamelCase : List[Any] = self.c_max_length // 2 while next_ncol > 0: lowerCamelCase : Optional[Any] = [[] for i in range(__magic_name__ )] lowerCamelCase : int = self.root**next_ncol # First half of next step lowerCamelCase : Tuple = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(__magic_name__ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step lowerCamelCase : Optional[Any] = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(__magic_name__ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update lowerCamelCase : List[str] = new_dft lowerCamelCase : Dict = next_ncol // 2 return dft[0] def UpperCamelCase__ ( self ): lowerCamelCase : str = self.__dft("""A""" ) lowerCamelCase : Optional[int] = self.__dft("""B""" ) lowerCamelCase : Optional[int] = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT lowerCamelCase : str = 2 while next_ncol <= self.c_max_length: lowerCamelCase : Tuple = [[] for i in range(__magic_name__ )] lowerCamelCase : List[Any] = self.root ** (next_ncol // 2) lowerCamelCase : Optional[Any] = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update lowerCamelCase : Tuple = new_inverse_c next_ncol *= 2 # Unpack lowerCamelCase : int = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self ): lowerCamelCase : Optional[Any] = """A = """ + """ + """.join( F'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) lowerCamelCase : Any = """B = """ + """ + """.join( F'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) lowerCamelCase : List[Any] = """A*B = """ + """ + """.join( F'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return F'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
287
def _a ( lowerCamelCase ): if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True lowerCamelCase : Any = 4 lowerCamelCase : List[str] = (1 << p) - 1 for _ in range(p - 2 ): lowerCamelCase : List[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(1_1))
287
1
from math import factorial def _a ( lowerCamelCase = 100 ): return sum(map(lowerCamelCase, str(factorial(lowerCamelCase ) ) ) ) if __name__ == "__main__": print(solution(int(input("""Enter the Number: """).strip())))
287
import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , *__magic_name__ , **__magic_name__ ): warnings.warn( """The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use OwlViTImageProcessor instead.""" , __magic_name__ , ) super().__init__(*__magic_name__ , **__magic_name__ )
287
1
from argparse import ArgumentParser from . import BaseTransformersCLICommand def _a ( lowerCamelCase ): return DownloadCommand(args.model, args.cache_dir, args.force, args.trust_remote_code ) class A__ ( __SCREAMING_SNAKE_CASE): @staticmethod def UpperCamelCase__ ( __magic_name__ ): lowerCamelCase : List[Any] = parser.add_parser("""download""" ) download_parser.add_argument( """--cache-dir""" , type=__magic_name__ , default=__magic_name__ , help="""Path to location to store the models""" ) download_parser.add_argument( """--force""" , action="""store_true""" , help="""Force the model to be download even if already in cache-dir""" ) download_parser.add_argument( """--trust-remote-code""" , action="""store_true""" , help="""Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you've reviewed the code as it will execute on your local machine""" , ) download_parser.add_argument("""model""" , type=__magic_name__ , help="""Name of the model to download""" ) download_parser.set_defaults(func=__magic_name__ ) def __init__( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): lowerCamelCase : Optional[Any] = model lowerCamelCase : List[Any] = cache lowerCamelCase : Optional[int] = force lowerCamelCase : str = trust_remote_code def UpperCamelCase__ ( self ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
287
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """nvidia/segformer-b0-finetuned-ade-512-512""": ( """https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json""" ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """segformer""" def __init__( self , __magic_name__=3 , __magic_name__=4 , __magic_name__=[2, 2, 2, 2] , __magic_name__=[8, 4, 2, 1] , __magic_name__=[3_2, 6_4, 1_6_0, 2_5_6] , __magic_name__=[7, 3, 3, 3] , __magic_name__=[4, 2, 2, 2] , __magic_name__=[1, 2, 5, 8] , __magic_name__=[4, 4, 4, 4] , __magic_name__="gelu" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=0.02 , __magic_name__=0.1 , __magic_name__=1e-6 , __magic_name__=2_5_6 , __magic_name__=2_5_5 , **__magic_name__ , ): super().__init__(**__magic_name__ ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( """Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be""" """ removed, as the behaviour will default to that of reshape_last_stage = True.""" , __magic_name__ , ) lowerCamelCase : Optional[Any] = num_channels lowerCamelCase : str = num_encoder_blocks lowerCamelCase : Any = depths lowerCamelCase : List[Any] = sr_ratios lowerCamelCase : int = hidden_sizes lowerCamelCase : Union[str, Any] = patch_sizes lowerCamelCase : Optional[Any] = strides lowerCamelCase : Dict = mlp_ratios lowerCamelCase : str = num_attention_heads lowerCamelCase : Any = hidden_act lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Union[str, Any] = attention_probs_dropout_prob lowerCamelCase : Dict = classifier_dropout_prob lowerCamelCase : Tuple = initializer_range lowerCamelCase : Dict = drop_path_rate lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : Any = decoder_hidden_size lowerCamelCase : str = kwargs.get("""reshape_last_stage""" , __magic_name__ ) lowerCamelCase : Dict = semantic_loss_ignore_index class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-4 @property def UpperCamelCase__ ( self ): return 1_2
287
1
import math def _a ( lowerCamelCase ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5, int(math.sqrt(lowerCamelCase ) + 1 ), 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _a ( lowerCamelCase = 0.1 ): lowerCamelCase : Optional[int] = 3 lowerCamelCase : Dict = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1, (j + 2) * (j + 2), j + 1 ): primes += is_prime(lowerCamelCase ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
287
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = """gpt_neo""" _UpperCAmelCase : Union[str, Any] = ["""past_key_values"""] _UpperCAmelCase : List[Any] = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , __magic_name__=5_0_2_5_7 , __magic_name__=2_0_4_8 , __magic_name__=2_0_4_8 , __magic_name__=2_4 , __magic_name__=[[["global", "local"], 1_2]] , __magic_name__=1_6 , __magic_name__=None , __magic_name__=2_5_6 , __magic_name__="gelu_new" , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.1 , __magic_name__=1e-5 , __magic_name__=0.02 , __magic_name__=True , __magic_name__=5_0_2_5_6 , __magic_name__=5_0_2_5_6 , **__magic_name__ , ): lowerCamelCase : List[Any] = vocab_size lowerCamelCase : str = max_position_embeddings lowerCamelCase : str = hidden_size lowerCamelCase : Optional[int] = num_layers lowerCamelCase : str = num_heads lowerCamelCase : Optional[Any] = intermediate_size lowerCamelCase : List[Any] = window_size lowerCamelCase : int = activation_function lowerCamelCase : Union[str, Any] = resid_dropout lowerCamelCase : List[Any] = embed_dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Dict = classifier_dropout lowerCamelCase : Any = layer_norm_epsilon lowerCamelCase : Dict = initializer_range lowerCamelCase : Dict = use_cache lowerCamelCase : Optional[Any] = bos_token_id lowerCamelCase : int = eos_token_id lowerCamelCase : List[Any] = attention_types lowerCamelCase : Optional[Any] = self.expand_attention_types_params(__magic_name__ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.attention_layers)` == `config.num_layers` """ F'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' """`config.attention_layers` is prepared using `config.attention_types`. """ """Please verify the value of `config.attention_types` argument.""" ) super().__init__(bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ ): lowerCamelCase : Optional[int] = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : Any = input.size() lowerCamelCase : List[Any] = len(lowerCamelCase ) lowerCamelCase : Optional[Any] = shape[dimension] lowerCamelCase : Optional[int] = torch.arange(0, lowerCamelCase, lowerCamelCase ) lowerCamelCase : Dict = torch.div(sizedim - size, lowerCamelCase, rounding_mode="""floor""" ) + 1 lowerCamelCase : int = torch.arange(lowerCamelCase ) + low_indices[:min_length][:, None] lowerCamelCase : str = [slice(lowerCamelCase )] * rank lowerCamelCase : List[str] = indices lowerCamelCase : Dict = input[s] lowerCamelCase : Any = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase ): import torch lowerCamelCase : List[Any] = torch.arange(1, lowerCamelCase ) lowerCamelCase : Optional[int] = torch.remainder(lowerCamelCase, lowerCamelCase ) lowerCamelCase : List[Any] = remainders == 0 lowerCamelCase : List[Any] = candidates[divisor_indices] lowerCamelCase : Optional[Any] = torch.max(lowerCamelCase ) return largest_divisor, torch.div(lowerCamelCase, lowerCamelCase, rounding_mode="""floor""" ) class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): lowerCamelCase : str = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(__magic_name__ , direction="""inputs""" ) lowerCamelCase : int = {0: """batch""", 1: """past_sequence + sequence"""} else: lowerCamelCase : Tuple = {0: """batch""", 1: """sequence"""} return common_inputs @property def UpperCamelCase__ ( self ): return self._config.num_heads def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = -1 , __magic_name__ = -1 , __magic_name__ = False , __magic_name__ = None , ): lowerCamelCase : Optional[int] = super(__magic_name__ , self ).generate_dummy_inputs( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) # We need to order the input in the way they appears in the forward() lowerCamelCase : int = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch lowerCamelCase , lowerCamelCase : Optional[Any] = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values lowerCamelCase : Optional[int] = seqlen + 2 lowerCamelCase : List[Any] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowerCamelCase : str = [ (torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) for _ in range(self.num_layers ) ] lowerCamelCase : Tuple = common_inputs["""attention_mask"""] if self.use_past: lowerCamelCase : str = ordered_inputs["""attention_mask"""].dtype lowerCamelCase : Any = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(__magic_name__ , __magic_name__ , dtype=__magic_name__ )] , dim=1 ) return ordered_inputs @property def UpperCamelCase__ ( self ): return 1_3
287
1
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=__SCREAMING_SNAKE_CASE) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = field(default="""audio-classification""" , metadata={"""include_in_asdict_even_if_is_default""": True}) _UpperCAmelCase : ClassVar[Features] = Features({"""audio""": Audio()}) _UpperCAmelCase : ClassVar[Features] = Features({"""labels""": ClassLabel}) _UpperCAmelCase : str = "audio" _UpperCAmelCase : str = "labels" def UpperCamelCase__ ( self , __magic_name__ ): if self.label_column not in features: raise ValueError(F'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] , __magic_name__ ): raise ValueError(F'''Column {self.label_column} is not a ClassLabel.''' ) lowerCamelCase : int = copy.deepcopy(self ) lowerCamelCase : Tuple = self.label_schema.copy() lowerCamelCase : Optional[int] = features[self.label_column] lowerCamelCase : Dict = label_schema return task_template @property def UpperCamelCase__ ( self ): return { self.audio_column: "audio", self.label_column: "labels", }
287
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
287
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_camembert import CamembertTokenizer else: _lowerCamelCase =None _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} _lowerCamelCase ={ """vocab_file""": { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model""", }, """tokenizer_file""": { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/tokenizer.json""", }, } _lowerCamelCase ={ """camembert-base""": 5_1_2, } _lowerCamelCase ="""▁""" class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = VOCAB_FILES_NAMES _UpperCAmelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : int = ["""input_ids""", """attention_mask"""] _UpperCAmelCase : List[Any] = CamembertTokenizer def __init__( self , __magic_name__=None , __magic_name__=None , __magic_name__="<s>" , __magic_name__="</s>" , __magic_name__="</s>" , __magic_name__="<s>" , __magic_name__="<unk>" , __magic_name__="<pad>" , __magic_name__="<mask>" , __magic_name__=["<s>NOTUSED", "</s>NOTUSED"] , **__magic_name__ , ): # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase : Optional[Any] = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else mask_token super().__init__( __magic_name__ , tokenizer_file=__magic_name__ , bos_token=__magic_name__ , eos_token=__magic_name__ , sep_token=__magic_name__ , cls_token=__magic_name__ , unk_token=__magic_name__ , pad_token=__magic_name__ , mask_token=__magic_name__ , additional_special_tokens=__magic_name__ , **__magic_name__ , ) lowerCamelCase : Tuple = vocab_file lowerCamelCase : Dict = False if not self.vocab_file else True def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCamelCase : Any = [self.cls_token_id] lowerCamelCase : List[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): lowerCamelCase : List[str] = [self.sep_token_id] lowerCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(__magic_name__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCamelCase : Any = os.path.join( __magic_name__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__magic_name__ ): copyfile(self.vocab_file , __magic_name__ ) return (out_vocab_file,)
287
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): for param, grad_param in zip(model_a.parameters(), model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, grad_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=True ): model.train() lowerCamelCase : Dict = model(lowerCamelCase ) lowerCamelCase : Any = F.mse_loss(lowerCamelCase, target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase=False ): set_seed(42 ) lowerCamelCase : Tuple = RegressionModel() lowerCamelCase : Any = deepcopy(lowerCamelCase ) lowerCamelCase : Any = RegressionDataset(length=80 ) lowerCamelCase : Dict = DataLoader(lowerCamelCase, batch_size=16 ) model.to(accelerator.device ) if sched: lowerCamelCase : int = AdamW(params=model.parameters(), lr=1e-3 ) lowerCamelCase : Optional[Any] = AdamW(params=ddp_model.parameters(), lr=1e-3 ) lowerCamelCase : str = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) lowerCamelCase : Tuple = LambdaLR(lowerCamelCase, lr_lambda=lambda lowerCamelCase : epoch**0.6_5 ) # Make a copy of `model` if sched: lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = accelerator.prepare(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: lowerCamelCase , lowerCamelCase : List[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def _a ( lowerCamelCase ): # Test when on a single CPU or GPU that the context manager does nothing lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : int = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad, ddp_param.grad ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : List[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase ): # Test on distributed setup that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : int = get_training_setup(lowerCamelCase ) # Use a single batch lowerCamelCase , lowerCamelCase : Union[str, Any] = next(iter(lowerCamelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Any = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : str = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) else: # Sync grads step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Optional[Any] = ddp_input[torch.randperm(len(lowerCamelCase ) )] def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : Any = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase : Any = get_training_setup(lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : str = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Do "gradient accumulation" (noop) with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters(), ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(lowerCamelCase ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad, ddp_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowerCamelCase : Any = ddp_input[torch.randperm(len(lowerCamelCase ) )] GradientState._reset_state() def _a ( lowerCamelCase=False, lowerCamelCase=False ): lowerCamelCase : List[Any] = Accelerator( split_batches=lowerCamelCase, dispatch_batches=lowerCamelCase, gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = get_training_setup(lowerCamelCase, lowerCamelCase ) for iteration, batch in enumerate(lowerCamelCase ): lowerCamelCase , lowerCamelCase : Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.gather((ddp_input, ddp_target) ) lowerCamelCase , lowerCamelCase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(lowerCamelCase )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(lowerCamelCase ): step_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]["lr"]}\nDDP opt: {ddp_opt.param_groups[0]["lr"]}\n''' lowerCamelCase : Union[str, Any] = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(lowerCamelCase )) if accelerator.num_processes > 1: check_model_parameters(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def _a ( ): lowerCamelCase : int = Accelerator() lowerCamelCase : Optional[Any] = RegressionDataset(length=80 ) lowerCamelCase : List[str] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase : int = RegressionDataset(length=96 ) lowerCamelCase : Optional[int] = DataLoader(lowerCamelCase, batch_size=16 ) lowerCamelCase , lowerCamelCase : Optional[Any] = accelerator.prepare(lowerCamelCase, lowerCamelCase ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if iteration < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(lowerCamelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase ) if batch_num < len(lowerCamelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def _a ( ): lowerCamelCase : List[Any] = Accelerator() lowerCamelCase : int = accelerator.state if state.local_process_index == 0: print("""**Test `accumulate` gradient accumulation with dataloader break**""" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("""**Test NOOP `no_sync` context manager**""" ) test_noop_sync(lowerCamelCase ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("""**Test Distributed `no_sync` context manager**""" ) test_distributed_sync(lowerCamelCase ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation(lowerCamelCase, lowerCamelCase ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("""<""", """2.0""" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, """`split_batches=False`, `dispatch_batches=False`**""", ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """, F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''', ) test_gradient_accumulation_with_opt_and_scheduler(lowerCamelCase, lowerCamelCase ) def _a ( lowerCamelCase ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
287
1
from __future__ import annotations from random import choice def _a ( lowerCamelCase ): return choice(lowerCamelCase ) def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : Any = random_pivot(lowerCamelCase ) # partition based on pivot # linear time lowerCamelCase : Union[str, Any] = [e for e in lst if e < pivot] lowerCamelCase : Optional[Any] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(lowerCamelCase ) == k - 1: return pivot # pivot is in elements bigger than k elif len(lowerCamelCase ) < k - 1: return kth_number(lowerCamelCase, k - len(lowerCamelCase ) - 1 ) # pivot is in elements smaller than k else: return kth_number(lowerCamelCase, lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
287
from scipy.stats import pearsonr import datasets _lowerCamelCase =""" Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ _lowerCamelCase =""" Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ _lowerCamelCase =""" @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class A__ ( datasets.Metric): def UpperCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"""] , ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__=False ): if return_pvalue: lowerCamelCase : Optional[Any] = pearsonr(__magic_name__ , __magic_name__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(__magic_name__ , __magic_name__ )[0] )}
287
1
import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel _lowerCamelCase ={ """text_branch""": """text_model""", """audio_branch""": """audio_model.audio_encoder""", """attn""": """attention.self""", """self.proj""": """output.dense""", """attention.self_mask""": """attn_mask""", """mlp.fc1""": """intermediate.dense""", """mlp.fc2""": """output.dense""", """norm1""": """layernorm_before""", """norm2""": """layernorm_after""", """bn0""": """batch_norm""", } _lowerCamelCase =AutoFeatureExtractor.from_pretrained("""laion/clap-htsat-unfused""", truncation="""rand_trunc""") def _a ( lowerCamelCase, lowerCamelCase=False ): lowerCamelCase , lowerCamelCase : int = create_model( """HTSAT-tiny""", """roberta""", lowerCamelCase, precision="""fp32""", device="""cuda:0""" if torch.cuda.is_available() else """cpu""", enable_fusion=lowerCamelCase, fusion_type="""aff_2d""" if enable_fusion else None, ) return model, model_cfg def _a ( lowerCamelCase ): lowerCamelCase : List[Any] = {} lowerCamelCase : str = R""".*sequential.(\d+).*""" lowerCamelCase : str = R""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCamelCase : int = key.replace(lowerCamelCase, lowerCamelCase ) if re.match(lowerCamelCase, lowerCamelCase ): # replace sequential layers with list lowerCamelCase : str = re.match(lowerCamelCase, lowerCamelCase ).group(1 ) lowerCamelCase : Optional[int] = key.replace(F'''sequential.{sequential_layer}.''', F'''layers.{int(lowerCamelCase )//3}.linear.''' ) elif re.match(lowerCamelCase, lowerCamelCase ): lowerCamelCase : Tuple = int(re.match(lowerCamelCase, lowerCamelCase ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... lowerCamelCase : List[Any] = 1 if projecton_layer == 0 else 2 lowerCamelCase : Optional[Any] = key.replace(F'''_projection.{projecton_layer}.''', F'''_projection.linear{transformers_projection_layer}.''' ) if "audio" and "qkv" in key: # split qkv into query key and value lowerCamelCase : Dict = value lowerCamelCase : int = mixed_qkv.size(0 ) // 3 lowerCamelCase : Dict = mixed_qkv[:qkv_dim] lowerCamelCase : List[str] = mixed_qkv[qkv_dim : qkv_dim * 2] lowerCamelCase : Tuple = mixed_qkv[qkv_dim * 2 :] lowerCamelCase : List[str] = query_layer lowerCamelCase : List[str] = key_layer lowerCamelCase : Union[str, Any] = value_layer else: lowerCamelCase : Optional[Any] = value return model_state_dict def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=False ): lowerCamelCase , lowerCamelCase : List[Any] = init_clap(lowerCamelCase, enable_fusion=lowerCamelCase ) clap_model.eval() lowerCamelCase : List[Any] = clap_model.state_dict() lowerCamelCase : str = rename_state_dict(lowerCamelCase ) lowerCamelCase : Union[str, Any] = ClapConfig() lowerCamelCase : List[Any] = enable_fusion lowerCamelCase : Tuple = ClapModel(lowerCamelCase ) # ignore the spectrogram embedding layer model.load_state_dict(lowerCamelCase, strict=lowerCamelCase ) model.save_pretrained(lowerCamelCase ) transformers_config.save_pretrained(lowerCamelCase ) if __name__ == "__main__": _lowerCamelCase =argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument("""--enable_fusion""", action="""store_true""", help="""Whether to enable fusion or not""") _lowerCamelCase =parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
287
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """microsoft/conditional-detr-resnet-50""": ( """https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """conditional_detr""" _UpperCAmelCase : Optional[int] = ["""past_key_values"""] _UpperCAmelCase : Optional[Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , __magic_name__=True , __magic_name__=None , __magic_name__=3 , __magic_name__=3_0_0 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=True , __magic_name__="relu" , __magic_name__=2_5_6 , __magic_name__=0.1 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.02 , __magic_name__=1.0 , __magic_name__=False , __magic_name__="sine" , __magic_name__="resnet50" , __magic_name__=True , __magic_name__=False , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=1 , __magic_name__=1 , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=0.25 , **__magic_name__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) lowerCamelCase : Optional[int] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(__magic_name__ , __magic_name__ ): lowerCamelCase : List[Any] = backbone_config.get("""model_type""" ) lowerCamelCase : Dict = CONFIG_MAPPING[backbone_model_type] lowerCamelCase : str = config_class.from_dict(__magic_name__ ) lowerCamelCase : Dict = use_timm_backbone lowerCamelCase : str = backbone_config lowerCamelCase : Tuple = num_channels lowerCamelCase : Dict = num_queries lowerCamelCase : Any = d_model lowerCamelCase : Optional[Any] = encoder_ffn_dim lowerCamelCase : List[str] = encoder_layers lowerCamelCase : Union[str, Any] = encoder_attention_heads lowerCamelCase : Any = decoder_ffn_dim lowerCamelCase : Dict = decoder_layers lowerCamelCase : Union[str, Any] = decoder_attention_heads lowerCamelCase : Dict = dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Union[str, Any] = activation_dropout lowerCamelCase : Optional[int] = activation_function lowerCamelCase : int = init_std lowerCamelCase : str = init_xavier_std lowerCamelCase : Tuple = encoder_layerdrop lowerCamelCase : str = decoder_layerdrop lowerCamelCase : Tuple = encoder_layers lowerCamelCase : Optional[int] = auxiliary_loss lowerCamelCase : Optional[Any] = position_embedding_type lowerCamelCase : Optional[int] = backbone lowerCamelCase : Union[str, Any] = use_pretrained_backbone lowerCamelCase : str = dilation # Hungarian matcher lowerCamelCase : Optional[Any] = class_cost lowerCamelCase : Dict = bbox_cost lowerCamelCase : Tuple = giou_cost # Loss coefficients lowerCamelCase : Union[str, Any] = mask_loss_coefficient lowerCamelCase : Dict = dice_loss_coefficient lowerCamelCase : Optional[int] = cls_loss_coefficient lowerCamelCase : Optional[int] = bbox_loss_coefficient lowerCamelCase : Optional[int] = giou_loss_coefficient lowerCamelCase : Optional[int] = focal_alpha super().__init__(is_encoder_decoder=__magic_name__ , **__magic_name__ ) @property def UpperCamelCase__ ( self ): return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): return self.d_model def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: lowerCamelCase : Optional[int] = self.backbone_config.to_dict() lowerCamelCase : Optional[Any] = self.__class__.model_type return output class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-5 @property def UpperCamelCase__ ( self ): return 1_2
287
1
from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , __magic_name__ , __magic_name__ ): super().__init__() # make sure scheduler can always be converted to DDIM lowerCamelCase : Union[str, Any] = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=__magic_name__ , scheduler=__magic_name__ ) @torch.no_grad() def __call__( self , __magic_name__ = 1 , __magic_name__ = None , __magic_name__ = 0.0 , __magic_name__ = 5_0 , __magic_name__ = None , __magic_name__ = "pil" , __magic_name__ = True , ): # Sample gaussian noise to begin loop if isinstance(self.unet.config.sample_size , __magic_name__ ): lowerCamelCase : Tuple = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: lowerCamelCase : List[str] = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(__magic_name__ , __magic_name__ ) and len(__magic_name__ ) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(__magic_name__ )}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) lowerCamelCase : Dict = randn_tensor(__magic_name__ , generator=__magic_name__ , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(__magic_name__ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output lowerCamelCase : List[Any] = self.unet(__magic_name__ , __magic_name__ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 lowerCamelCase : str = self.scheduler.step( __magic_name__ , __magic_name__ , __magic_name__ , eta=__magic_name__ , use_clipped_model_output=__magic_name__ , generator=__magic_name__ ).prev_sample lowerCamelCase : Dict = (image / 2 + 0.5).clamp(0 , 1 ) lowerCamelCase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCamelCase : Union[str, Any] = self.numpy_to_pil(__magic_name__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=__magic_name__ )
287
import json import sys def _a ( lowerCamelCase, lowerCamelCase ): with open(lowerCamelCase, encoding="""utf-8""" ) as f: lowerCamelCase : List[Any] = json.load(lowerCamelCase ) lowerCamelCase : Optional[Any] = ["""<details>""", """<summary>Show updated benchmarks!</summary>""", """ """] for benchmark_name in sorted(lowerCamelCase ): lowerCamelCase : List[Any] = results[benchmark_name] lowerCamelCase : Union[str, Any] = benchmark_name.split("""/""" )[-1] output_md.append(F'''### Benchmark: {benchmark_file_name}''' ) lowerCamelCase : Any = """| metric |""" lowerCamelCase : str = """|--------|""" lowerCamelCase : List[Any] = """| new / old (diff) |""" for metric_name in sorted(lowerCamelCase ): lowerCamelCase : List[Any] = benchmark_res[metric_name] lowerCamelCase : Tuple = metric_vals["""new"""] lowerCamelCase : int = metric_vals.get("""old""", lowerCamelCase ) lowerCamelCase : Dict = metric_vals.get("""diff""", lowerCamelCase ) lowerCamelCase : Dict = F''' {new_val:f}''' if isinstance(lowerCamelCase, (int, float) ) else """None""" if old_val is not None: val_str += F''' / {old_val:f}''' if isinstance(lowerCamelCase, (int, float) ) else "None" if dif_val is not None: val_str += F''' ({dif_val:f})''' if isinstance(lowerCamelCase, (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append("""</details>""" ) with open(lowerCamelCase, """w""", encoding="""utf-8""" ) as f: f.writelines("""\n""".join(lowerCamelCase ) ) if __name__ == "__main__": _lowerCamelCase =sys.argv[1] _lowerCamelCase =sys.argv[2] format_json_to_md(input_json_file, output_md_file)
287
1
import math def _a ( lowerCamelCase ): assert isinstance(lowerCamelCase, lowerCamelCase ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False lowerCamelCase : int = range(3, int(math.sqrt(lowerCamelCase ) + 1 ), 2 ) return not any(not number % i for i in odd_numbers ) def _a ( lowerCamelCase, lowerCamelCase=1, **lowerCamelCase ): lowerCamelCase : int = factor * value lowerCamelCase : str = value while not is_prime(lowerCamelCase ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1, **lowerCamelCase ) return value
287
def _a ( lowerCamelCase ): return " ".join( """""".join(word[::-1] ) if len(lowerCamelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("""Hey wollef sroirraw"""))
287
1
import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def _a ( lowerCamelCase ): lowerCamelCase : List[str] = SwinConfig() lowerCamelCase : List[str] = swin_name.split("""_""" ) lowerCamelCase : Tuple = name_split[1] lowerCamelCase : Union[str, Any] = int(name_split[4] ) lowerCamelCase : Union[str, Any] = int(name_split[3][-1] ) if model_size == "tiny": lowerCamelCase : int = 96 lowerCamelCase : List[str] = (2, 2, 6, 2) lowerCamelCase : List[str] = (3, 6, 12, 24) elif model_size == "small": lowerCamelCase : Optional[Any] = 96 lowerCamelCase : List[str] = (2, 2, 18, 2) lowerCamelCase : Optional[Any] = (3, 6, 12, 24) elif model_size == "base": lowerCamelCase : Any = 128 lowerCamelCase : Optional[Any] = (2, 2, 18, 2) lowerCamelCase : Optional[Any] = (4, 8, 16, 32) else: lowerCamelCase : List[Any] = 192 lowerCamelCase : Any = (2, 2, 18, 2) lowerCamelCase : Optional[Any] = (6, 12, 24, 48) if "in22k" in swin_name: lowerCamelCase : int = 2_1841 else: lowerCamelCase : str = 1000 lowerCamelCase : int = """huggingface/label-files""" lowerCamelCase : Dict = """imagenet-1k-id2label.json""" lowerCamelCase : List[Any] = json.load(open(hf_hub_download(lowerCamelCase, lowerCamelCase, repo_type="""dataset""" ), """r""" ) ) lowerCamelCase : Optional[int] = {int(lowerCamelCase ): v for k, v in idalabel.items()} lowerCamelCase : Tuple = idalabel lowerCamelCase : List[Any] = {v: k for k, v in idalabel.items()} lowerCamelCase : List[str] = img_size lowerCamelCase : Optional[Any] = num_classes lowerCamelCase : Any = embed_dim lowerCamelCase : Tuple = depths lowerCamelCase : Optional[int] = num_heads lowerCamelCase : str = window_size return config def _a ( lowerCamelCase ): if "patch_embed.proj" in name: lowerCamelCase : List[str] = name.replace("""patch_embed.proj""", """embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: lowerCamelCase : Any = name.replace("""patch_embed.norm""", """embeddings.norm""" ) if "layers" in name: lowerCamelCase : List[Any] = """encoder.""" + name if "attn.proj" in name: lowerCamelCase : str = name.replace("""attn.proj""", """attention.output.dense""" ) if "attn" in name: lowerCamelCase : Optional[Any] = name.replace("""attn""", """attention.self""" ) if "norm1" in name: lowerCamelCase : Optional[Any] = name.replace("""norm1""", """layernorm_before""" ) if "norm2" in name: lowerCamelCase : int = name.replace("""norm2""", """layernorm_after""" ) if "mlp.fc1" in name: lowerCamelCase : Any = name.replace("""mlp.fc1""", """intermediate.dense""" ) if "mlp.fc2" in name: lowerCamelCase : List[Any] = name.replace("""mlp.fc2""", """output.dense""" ) if name == "norm.weight": lowerCamelCase : str = """layernorm.weight""" if name == "norm.bias": lowerCamelCase : List[Any] = """layernorm.bias""" if "head" in name: lowerCamelCase : Union[str, Any] = name.replace("""head""", """classifier""" ) else: lowerCamelCase : str = """swin.""" + name return name def _a ( lowerCamelCase, lowerCamelCase ): for key in orig_state_dict.copy().keys(): lowerCamelCase : List[Any] = orig_state_dict.pop(lowerCamelCase ) if "mask" in key: continue elif "qkv" in key: lowerCamelCase : Tuple = key.split(""".""" ) lowerCamelCase : Tuple = int(key_split[1] ) lowerCamelCase : Dict = int(key_split[3] ) lowerCamelCase : Dict = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: lowerCamelCase : List[Any] = val[:dim, :] lowerCamelCase : int = val[ dim : dim * 2, : ] lowerCamelCase : Optional[int] = val[-dim:, :] else: lowerCamelCase : Union[str, Any] = val[ :dim ] lowerCamelCase : Tuple = val[ dim : dim * 2 ] lowerCamelCase : List[Any] = val[ -dim: ] else: lowerCamelCase : Union[str, Any] = val return orig_state_dict def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : Optional[Any] = timm.create_model(lowerCamelCase, pretrained=lowerCamelCase ) timm_model.eval() lowerCamelCase : Union[str, Any] = get_swin_config(lowerCamelCase ) lowerCamelCase : int = SwinForImageClassification(lowerCamelCase ) model.eval() lowerCamelCase : Optional[Any] = convert_state_dict(timm_model.state_dict(), lowerCamelCase ) model.load_state_dict(lowerCamelCase ) lowerCamelCase : List[Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase : Tuple = AutoImageProcessor.from_pretrained("""microsoft/{}""".format(swin_name.replace("""_""", """-""" ) ) ) lowerCamelCase : List[str] = Image.open(requests.get(lowerCamelCase, stream=lowerCamelCase ).raw ) lowerCamelCase : str = image_processor(images=lowerCamelCase, return_tensors="""pt""" ) lowerCamelCase : Optional[int] = timm_model(inputs["""pixel_values"""] ) lowerCamelCase : Union[str, Any] = model(**lowerCamelCase ).logits assert torch.allclose(lowerCamelCase, lowerCamelCase, atol=1e-3 ) print(F'''Saving model {swin_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCamelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowerCamelCase ) if __name__ == "__main__": _lowerCamelCase =argparse.ArgumentParser() # Required parameters parser.add_argument( """--swin_name""", default="""swin_tiny_patch4_window7_224""", type=str, help="""Name of the Swin timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _lowerCamelCase =parser.parse_args() convert_swin_checkpoint(args.swin_name, args.pytorch_dump_folder_path)
287
# Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from packaging import version from .. import __version__ from .constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD from .doc import ( add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, copy_func, replace_return_docstrings, ) from .generic import ( ContextManagers, ExplicitEnum, ModelOutput, PaddingStrategy, TensorType, add_model_info_to_auto_map, cached_property, can_return_loss, expand_dims, find_labels, flatten_dict, infer_framework, is_jax_tensor, is_numpy_array, is_tensor, is_tf_symbolic_tensor, is_tf_tensor, is_torch_device, is_torch_dtype, is_torch_tensor, reshape, squeeze, strtobool, tensor_size, to_numpy, to_py_obj, transpose, working_or_temp_dir, ) from .hub import ( CLOUDFRONT_DISTRIB_PREFIX, DISABLE_TELEMETRY, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, EntryNotFoundError, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, cached_file, default_cache_path, define_sagemaker_information, download_url, extract_commit_hash, get_cached_models, get_file_from_repo, get_full_repo_name, has_file, http_user_agent, is_offline_mode, is_remote_url, move_cache, send_example_telemetry, try_to_load_from_cache, ) from .import_utils import ( ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, TORCH_FX_REQUIRED_VERSION, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, ccl_version, direct_transformers_import, get_torch_version, is_accelerate_available, is_apex_available, is_bitsandbytes_available, is_bsa_available, is_coloredlogs_available, is_cython_available, is_datasets_available, is_decord_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_jieba_available, is_jumanpp_available, is_kenlm_available, is_keras_nlp_available, is_librosa_available, is_natten_available, is_ninja_available, is_onnx_available, is_openai_available, is_optimum_available, is_pandas_available, is_peft_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytest_available, is_pytorch_quantization_available, is_rjieba_available, is_sacremoses_available, is_safetensors_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_sudachi_available, is_tensorflow_probability_available, is_tensorflow_text_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_bfaa_cpu_available, is_torch_bfaa_gpu_available, is_torch_compile_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_neuroncore_available, is_torch_tensorrt_fx_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_torchdistx_available, is_torchdynamo_available, is_torchvision_available, is_training_run_on_sagemaker, is_vision_available, requires_backends, torch_only_method, ) _lowerCamelCase ="""pytorch_model.bin""" _lowerCamelCase ="""pytorch_model.bin.index.json""" _lowerCamelCase ="""adapter_config.json""" _lowerCamelCase ="""adapter_model.bin""" _lowerCamelCase ="""adapter_model.safetensors""" _lowerCamelCase ="""tf_model.h5""" _lowerCamelCase ="""tf_model.h5.index.json""" _lowerCamelCase ="""model.ckpt""" _lowerCamelCase ="""flax_model.msgpack""" _lowerCamelCase ="""flax_model.msgpack.index.json""" _lowerCamelCase ="""model.safetensors""" _lowerCamelCase ="""model.safetensors.index.json""" _lowerCamelCase ="""config.json""" _lowerCamelCase ="""preprocessor_config.json""" _lowerCamelCase =FEATURE_EXTRACTOR_NAME _lowerCamelCase ="""generation_config.json""" _lowerCamelCase ="""modelcard.json""" _lowerCamelCase ="""▁""" _lowerCamelCase =SENTENCEPIECE_UNDERLINE # Kept for backward compatibility _lowerCamelCase =[ [[0, 1, 0, 1], [1, 0, 0, 1]] ] * 2 # Needs to have 0s and 1s only since XLM uses it for langs too. _lowerCamelCase =[[7, 6, 0, 0, 1], [1, 2, 3, 0, 0], [0, 0, 0, 4, 5]] _lowerCamelCase =[[1, 1, 1, 1, 1], [1, 1, 1, 0, 0], [0, 0, 0, 1, 1]] def _a ( lowerCamelCase ): if version.parse(lowerCamelCase ) < version.parse(lowerCamelCase ): if "dev" in min_version: lowerCamelCase : Optional[int] = ( """This example requires a source install from HuggingFace Transformers (see """ """`https://huggingface.co/docs/transformers/installation#install-from-source`),""" ) else: lowerCamelCase : int = F'''This example requires a minimum version of {min_version},''' error_message += F''' but the version found is {__version__}.\n''' raise ImportError( error_message + """Check out https://github.com/huggingface/transformers/tree/main/examples#important-note for the examples corresponding to other """ """versions of HuggingFace Transformers.""" )
287
1
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import evaluate import numpy as np from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/text-classification/requirements.txt""") _lowerCamelCase =logging.getLogger(__name__) @dataclass class A__ : _UpperCAmelCase : Optional[int] = field( default=128 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) _UpperCAmelCase : bool = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""}) _UpperCAmelCase : bool = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """Whether to pad all samples to `max_seq_length`. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch.""" ) } , ) _UpperCAmelCase : Optional[int] = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) _UpperCAmelCase : Optional[int] = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) _UpperCAmelCase : Optional[int] = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of prediction examples to this """ """value if set.""" ) } , ) @dataclass class A__ : _UpperCAmelCase : str = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""}) _UpperCAmelCase : str = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Evaluation language. Also train language if `train_language` is set to None."""}) _UpperCAmelCase : Optional[str] = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Train language if it is different from the evaluation language."""}) _UpperCAmelCase : Optional[str] = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""}) _UpperCAmelCase : Optional[str] = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""}) _UpperCAmelCase : Optional[str] = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) _UpperCAmelCase : Optional[bool] = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """arg to indicate if tokenizer should do lower case in AutoTokenizer.from_pretrained()"""} , ) _UpperCAmelCase : bool = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) _UpperCAmelCase : str = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) _UpperCAmelCase : bool = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) _UpperCAmelCase : bool = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Will enable to load a pretrained model whose head dimensions are different."""} , ) def _a ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCamelCase : str = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_xnli""", lowerCamelCase ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", handlers=[logging.StreamHandler(sys.stdout )], ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowerCamelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(lowerCamelCase ) datasets.utils.logging.set_verbosity(lowerCamelCase ) transformers.utils.logging.set_verbosity(lowerCamelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. lowerCamelCase : Optional[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCamelCase : List[Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Set seed before initializing model. set_seed(training_args.seed ) # In distributed training, the load_dataset function guarantees that only one local process can concurrently # download the dataset. # Downloading and loading xnli dataset from the hub. if training_args.do_train: if model_args.train_language is None: lowerCamelCase : Union[str, Any] = load_dataset( """xnli""", model_args.language, split="""train""", cache_dir=model_args.cache_dir, use_auth_token=True if model_args.use_auth_token else None, ) else: lowerCamelCase : List[str] = load_dataset( """xnli""", model_args.train_language, split="""train""", cache_dir=model_args.cache_dir, use_auth_token=True if model_args.use_auth_token else None, ) lowerCamelCase : Optional[Any] = train_dataset.features["""label"""].names if training_args.do_eval: lowerCamelCase : List[str] = load_dataset( """xnli""", model_args.language, split="""validation""", cache_dir=model_args.cache_dir, use_auth_token=True if model_args.use_auth_token else None, ) lowerCamelCase : int = eval_dataset.features["""label"""].names if training_args.do_predict: lowerCamelCase : List[str] = load_dataset( """xnli""", model_args.language, split="""test""", cache_dir=model_args.cache_dir, use_auth_token=True if model_args.use_auth_token else None, ) lowerCamelCase : Dict = predict_dataset.features["""label"""].names # Labels lowerCamelCase : Tuple = len(lowerCamelCase ) # Load pretrained model and tokenizer # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path, num_labels=lowerCamelCase, idalabel={str(lowerCamelCase ): label for i, label in enumerate(lowerCamelCase )}, labelaid={label: i for i, label in enumerate(lowerCamelCase )}, finetuning_task="""xnli""", cache_dir=model_args.cache_dir, revision=model_args.model_revision, use_auth_token=True if model_args.use_auth_token else None, ) lowerCamelCase : Dict = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path, do_lower_case=model_args.do_lower_case, cache_dir=model_args.cache_dir, use_fast=model_args.use_fast_tokenizer, revision=model_args.model_revision, use_auth_token=True if model_args.use_auth_token else None, ) lowerCamelCase : int = AutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path, from_tf=bool(""".ckpt""" in model_args.model_name_or_path ), config=lowerCamelCase, cache_dir=model_args.cache_dir, revision=model_args.model_revision, use_auth_token=True if model_args.use_auth_token else None, ignore_mismatched_sizes=model_args.ignore_mismatched_sizes, ) # Preprocessing the datasets # Padding strategy if data_args.pad_to_max_length: lowerCamelCase : Optional[Any] = """max_length""" else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch lowerCamelCase : Tuple = False def preprocess_function(lowerCamelCase ): # Tokenize the texts return tokenizer( examples["""premise"""], examples["""hypothesis"""], padding=lowerCamelCase, max_length=data_args.max_seq_length, truncation=lowerCamelCase, ) if training_args.do_train: if data_args.max_train_samples is not None: lowerCamelCase : int = min(len(lowerCamelCase ), data_args.max_train_samples ) lowerCamelCase : List[Any] = train_dataset.select(range(lowerCamelCase ) ) with training_args.main_process_first(desc="""train dataset map pre-processing""" ): lowerCamelCase : List[Any] = train_dataset.map( lowerCamelCase, batched=lowerCamelCase, load_from_cache_file=not data_args.overwrite_cache, desc="""Running tokenizer on train dataset""", ) # Log a few random samples from the training set: for index in random.sample(range(len(lowerCamelCase ) ), 3 ): logger.info(F'''Sample {index} of the training set: {train_dataset[index]}.''' ) if training_args.do_eval: if data_args.max_eval_samples is not None: lowerCamelCase : Union[str, Any] = min(len(lowerCamelCase ), data_args.max_eval_samples ) lowerCamelCase : Union[str, Any] = eval_dataset.select(range(lowerCamelCase ) ) with training_args.main_process_first(desc="""validation dataset map pre-processing""" ): lowerCamelCase : Union[str, Any] = eval_dataset.map( lowerCamelCase, batched=lowerCamelCase, load_from_cache_file=not data_args.overwrite_cache, desc="""Running tokenizer on validation dataset""", ) if training_args.do_predict: if data_args.max_predict_samples is not None: lowerCamelCase : int = min(len(lowerCamelCase ), data_args.max_predict_samples ) lowerCamelCase : Any = predict_dataset.select(range(lowerCamelCase ) ) with training_args.main_process_first(desc="""prediction dataset map pre-processing""" ): lowerCamelCase : Union[str, Any] = predict_dataset.map( lowerCamelCase, batched=lowerCamelCase, load_from_cache_file=not data_args.overwrite_cache, desc="""Running tokenizer on prediction dataset""", ) # Get the metric function lowerCamelCase : Union[str, Any] = evaluate.load("""xnli""" ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(lowerCamelCase ): lowerCamelCase : Optional[Any] = p.predictions[0] if isinstance(p.predictions, lowerCamelCase ) else p.predictions lowerCamelCase : Union[str, Any] = np.argmax(lowerCamelCase, axis=1 ) return metric.compute(predictions=lowerCamelCase, references=p.label_ids ) # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: lowerCamelCase : List[Any] = default_data_collator elif training_args.fpaa: lowerCamelCase : Dict = DataCollatorWithPadding(lowerCamelCase, pad_to_multiple_of=8 ) else: lowerCamelCase : List[str] = None # Initialize our Trainer lowerCamelCase : str = Trainer( model=lowerCamelCase, args=lowerCamelCase, train_dataset=train_dataset if training_args.do_train else None, eval_dataset=eval_dataset if training_args.do_eval else None, compute_metrics=lowerCamelCase, tokenizer=lowerCamelCase, data_collator=lowerCamelCase, ) # Training if training_args.do_train: lowerCamelCase : List[Any] = None if training_args.resume_from_checkpoint is not None: lowerCamelCase : List[Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowerCamelCase : int = last_checkpoint lowerCamelCase : Dict = trainer.train(resume_from_checkpoint=lowerCamelCase ) lowerCamelCase : Tuple = train_result.metrics lowerCamelCase : Union[str, Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCamelCase ) ) lowerCamelCase : Any = min(lowerCamelCase, len(lowerCamelCase ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics("""train""", lowerCamelCase ) trainer.save_metrics("""train""", lowerCamelCase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info("""*** Evaluate ***""" ) lowerCamelCase : str = trainer.evaluate(eval_dataset=lowerCamelCase ) lowerCamelCase : Optional[int] = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowerCamelCase ) lowerCamelCase : Any = min(lowerCamelCase, len(lowerCamelCase ) ) trainer.log_metrics("""eval""", lowerCamelCase ) trainer.save_metrics("""eval""", lowerCamelCase ) # Prediction if training_args.do_predict: logger.info("""*** Predict ***""" ) lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = trainer.predict(lowerCamelCase, metric_key_prefix="""predict""" ) lowerCamelCase : Optional[int] = ( data_args.max_predict_samples if data_args.max_predict_samples is not None else len(lowerCamelCase ) ) lowerCamelCase : Any = min(lowerCamelCase, len(lowerCamelCase ) ) trainer.log_metrics("""predict""", lowerCamelCase ) trainer.save_metrics("""predict""", lowerCamelCase ) lowerCamelCase : int = np.argmax(lowerCamelCase, axis=1 ) lowerCamelCase : List[Any] = os.path.join(training_args.output_dir, """predictions.txt""" ) if trainer.is_world_process_zero(): with open(lowerCamelCase, """w""" ) as writer: writer.write("""index\tprediction\n""" ) for index, item in enumerate(lowerCamelCase ): lowerCamelCase : Optional[int] = label_list[item] writer.write(F'''{index}\t{item}\n''' ) if __name__ == "__main__": main()
287
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = """camembert""" def __init__( self , __magic_name__=3_0_5_2_2 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , __magic_name__=None , **__magic_name__ , ): super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : int = num_hidden_layers lowerCamelCase : int = num_attention_heads lowerCamelCase : Optional[int] = hidden_act lowerCamelCase : List[Any] = intermediate_size lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Optional[int] = attention_probs_dropout_prob lowerCamelCase : Optional[int] = max_position_embeddings lowerCamelCase : str = type_vocab_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : int = layer_norm_eps lowerCamelCase : Any = position_embedding_type lowerCamelCase : Optional[int] = use_cache lowerCamelCase : Union[str, Any] = classifier_dropout class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): if self.task == "multiple-choice": lowerCamelCase : List[str] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase : List[str] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
287
1
import logging import os from .state import PartialState class A__ ( logging.LoggerAdapter): @staticmethod def UpperCamelCase__ ( __magic_name__ ): lowerCamelCase : str = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , *__magic_name__ , **__magic_name__ ): if PartialState._shared_state == {}: raise RuntimeError( """You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility.""" ) lowerCamelCase : Optional[Any] = kwargs.pop("""main_process_only""" , __magic_name__ ) lowerCamelCase : Optional[int] = kwargs.pop("""in_order""" , __magic_name__ ) if self.isEnabledFor(__magic_name__ ): if self._should_log(__magic_name__ ): lowerCamelCase , lowerCamelCase : Any = self.process(__magic_name__ , __magic_name__ ) self.logger.log(__magic_name__ , __magic_name__ , *__magic_name__ , **__magic_name__ ) elif in_order: lowerCamelCase : Tuple = PartialState() for i in range(state.num_processes ): if i == state.process_index: lowerCamelCase , lowerCamelCase : str = self.process(__magic_name__ , __magic_name__ ) self.logger.log(__magic_name__ , __magic_name__ , *__magic_name__ , **__magic_name__ ) state.wait_for_everyone() def _a ( lowerCamelCase, lowerCamelCase = None ): if log_level is None: lowerCamelCase : List[Any] = os.environ.get("""ACCELERATE_LOG_LEVEL""", lowerCamelCase ) lowerCamelCase : List[str] = logging.getLogger(lowerCamelCase ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(lowerCamelCase, {} )
287
_lowerCamelCase ={ "joule": 1.0, "kilojoule": 1_0_0_0, "megajoule": 1_0_0_0_0_0_0, "gigajoule": 1_0_0_0_0_0_0_0_0_0, "wattsecond": 1.0, "watthour": 3_6_0_0, "kilowatthour": 3_6_0_0_0_0_0, "newtonmeter": 1.0, "calorie_nutr": 4_1_8_6.8, "kilocalorie_nutr": 4_1_8_6_8_0_0.0_0, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1_0_5_5.0_5_5_8_5, "footpound": 1.355818, } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowerCamelCase : Dict = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {", ".join(lowerCamelCase )}''' ) raise ValueError(lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
287
1
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = (DEISMultistepScheduler,) _UpperCAmelCase : Optional[Any] = (("""num_inference_steps""", 25),) def UpperCamelCase__ ( self , **__magic_name__ ): lowerCamelCase : str = { """num_train_timesteps""": 1_0_0_0, """beta_start""": 0.0_001, """beta_end""": 0.02, """beta_schedule""": """linear""", """solver_order""": 2, } config.update(**__magic_name__ ) return config def UpperCamelCase__ ( self , __magic_name__=0 , **__magic_name__ ): lowerCamelCase : str = dict(self.forward_default_kwargs ) lowerCamelCase : Any = kwargs.pop("""num_inference_steps""" , __magic_name__ ) lowerCamelCase : List[str] = self.dummy_sample lowerCamelCase : Dict = 0.1 * sample lowerCamelCase : List[Any] = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowerCamelCase : Any = self.get_scheduler_config(**__magic_name__ ) lowerCamelCase : Optional[Any] = scheduler_class(**__magic_name__ ) scheduler.set_timesteps(__magic_name__ ) # copy over dummy past residuals lowerCamelCase : str = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__magic_name__ ) lowerCamelCase : Any = scheduler_class.from_pretrained(__magic_name__ ) new_scheduler.set_timesteps(__magic_name__ ) # copy over dummy past residuals lowerCamelCase : Optional[Any] = dummy_past_residuals[: new_scheduler.config.solver_order] lowerCamelCase , lowerCamelCase : Optional[int] = sample, sample for t in range(__magic_name__ , time_step + scheduler.config.solver_order + 1 ): lowerCamelCase : int = scheduler.step(__magic_name__ , __magic_name__ , __magic_name__ , **__magic_name__ ).prev_sample lowerCamelCase : str = new_scheduler.step(__magic_name__ , __magic_name__ , __magic_name__ , **__magic_name__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self ): pass def UpperCamelCase__ ( self , __magic_name__=0 , **__magic_name__ ): lowerCamelCase : Union[str, Any] = dict(self.forward_default_kwargs ) lowerCamelCase : Optional[Any] = kwargs.pop("""num_inference_steps""" , __magic_name__ ) lowerCamelCase : Union[str, Any] = self.dummy_sample lowerCamelCase : Tuple = 0.1 * sample lowerCamelCase : Optional[Any] = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowerCamelCase : List[str] = self.get_scheduler_config() lowerCamelCase : Dict = scheduler_class(**__magic_name__ ) scheduler.set_timesteps(__magic_name__ ) # copy over dummy past residuals (must be after setting timesteps) lowerCamelCase : Tuple = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__magic_name__ ) lowerCamelCase : int = scheduler_class.from_pretrained(__magic_name__ ) # copy over dummy past residuals new_scheduler.set_timesteps(__magic_name__ ) # copy over dummy past residual (must be after setting timesteps) lowerCamelCase : int = dummy_past_residuals[: new_scheduler.config.solver_order] lowerCamelCase : int = scheduler.step(__magic_name__ , __magic_name__ , __magic_name__ , **__magic_name__ ).prev_sample lowerCamelCase : Optional[Any] = new_scheduler.step(__magic_name__ , __magic_name__ , __magic_name__ , **__magic_name__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self , __magic_name__=None , **__magic_name__ ): if scheduler is None: lowerCamelCase : List[Any] = self.scheduler_classes[0] lowerCamelCase : str = self.get_scheduler_config(**__magic_name__ ) lowerCamelCase : List[Any] = scheduler_class(**__magic_name__ ) lowerCamelCase : Optional[Any] = self.scheduler_classes[0] lowerCamelCase : str = self.get_scheduler_config(**__magic_name__ ) lowerCamelCase : Tuple = scheduler_class(**__magic_name__ ) lowerCamelCase : Dict = 1_0 lowerCamelCase : Optional[int] = self.dummy_model() lowerCamelCase : Optional[Any] = self.dummy_sample_deter scheduler.set_timesteps(__magic_name__ ) for i, t in enumerate(scheduler.timesteps ): lowerCamelCase : str = model(__magic_name__ , __magic_name__ ) lowerCamelCase : List[str] = scheduler.step(__magic_name__ , __magic_name__ , __magic_name__ ).prev_sample return sample def UpperCamelCase__ ( self ): lowerCamelCase : int = dict(self.forward_default_kwargs ) lowerCamelCase : Optional[Any] = kwargs.pop("""num_inference_steps""" , __magic_name__ ) for scheduler_class in self.scheduler_classes: lowerCamelCase : List[str] = self.get_scheduler_config() lowerCamelCase : Union[str, Any] = scheduler_class(**__magic_name__ ) lowerCamelCase : List[Any] = self.dummy_sample lowerCamelCase : str = 0.1 * sample if num_inference_steps is not None and hasattr(__magic_name__ , """set_timesteps""" ): scheduler.set_timesteps(__magic_name__ ) elif num_inference_steps is not None and not hasattr(__magic_name__ , """set_timesteps""" ): lowerCamelCase : Union[str, Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowerCamelCase : List[str] = [residual + 0.2, residual + 0.15, residual + 0.10] lowerCamelCase : Any = dummy_past_residuals[: scheduler.config.solver_order] lowerCamelCase : Dict = scheduler.timesteps[5] lowerCamelCase : Union[str, Any] = scheduler.timesteps[6] lowerCamelCase : Union[str, Any] = scheduler.step(__magic_name__ , __magic_name__ , __magic_name__ , **__magic_name__ ).prev_sample lowerCamelCase : List[str] = scheduler.step(__magic_name__ , __magic_name__ , __magic_name__ , **__magic_name__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def UpperCamelCase__ ( self ): # make sure that iterating over schedulers with same config names gives same results # for defaults lowerCamelCase : Union[str, Any] = DEISMultistepScheduler(**self.get_scheduler_config() ) lowerCamelCase : Dict = self.full_loop(scheduler=__magic_name__ ) lowerCamelCase : Union[str, Any] = torch.mean(torch.abs(__magic_name__ ) ) assert abs(result_mean.item() - 0.23_916 ) < 1e-3 lowerCamelCase : str = DPMSolverSinglestepScheduler.from_config(scheduler.config ) lowerCamelCase : Dict = DPMSolverMultistepScheduler.from_config(scheduler.config ) lowerCamelCase : Union[str, Any] = UniPCMultistepScheduler.from_config(scheduler.config ) lowerCamelCase : Union[str, Any] = DEISMultistepScheduler.from_config(scheduler.config ) lowerCamelCase : List[Any] = self.full_loop(scheduler=__magic_name__ ) lowerCamelCase : Dict = torch.mean(torch.abs(__magic_name__ ) ) assert abs(result_mean.item() - 0.23_916 ) < 1e-3 def UpperCamelCase__ ( self ): for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=__magic_name__ ) def UpperCamelCase__ ( self ): self.check_over_configs(thresholding=__magic_name__ ) for order in [1, 2, 3]: for solver_type in ["logrho"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=__magic_name__ , prediction_type=__magic_name__ , sample_max_value=__magic_name__ , algorithm_type="""deis""" , solver_order=__magic_name__ , solver_type=__magic_name__ , ) def UpperCamelCase__ ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__magic_name__ ) def UpperCamelCase__ ( self ): for algorithm_type in ["deis"]: for solver_type in ["logrho"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=__magic_name__ , solver_type=__magic_name__ , prediction_type=__magic_name__ , algorithm_type=__magic_name__ , ) lowerCamelCase : Union[str, Any] = self.full_loop( solver_order=__magic_name__ , solver_type=__magic_name__ , prediction_type=__magic_name__ , algorithm_type=__magic_name__ , ) assert not torch.isnan(__magic_name__ ).any(), "Samples have nan numbers" def UpperCamelCase__ ( self ): self.check_over_configs(lower_order_final=__magic_name__ ) self.check_over_configs(lower_order_final=__magic_name__ ) def UpperCamelCase__ ( self ): for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=__magic_name__ , time_step=0 ) def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = self.full_loop() lowerCamelCase : Union[str, Any] = torch.mean(torch.abs(__magic_name__ ) ) assert abs(result_mean.item() - 0.23_916 ) < 1e-3 def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = self.full_loop(prediction_type="""v_prediction""" ) lowerCamelCase : Tuple = torch.mean(torch.abs(__magic_name__ ) ) assert abs(result_mean.item() - 0.091 ) < 1e-3 def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = self.scheduler_classes[0] lowerCamelCase : str = self.get_scheduler_config(thresholding=__magic_name__ , dynamic_thresholding_ratio=0 ) lowerCamelCase : List[Any] = scheduler_class(**__magic_name__ ) lowerCamelCase : Dict = 1_0 lowerCamelCase : Optional[int] = self.dummy_model() lowerCamelCase : List[Any] = self.dummy_sample_deter.half() scheduler.set_timesteps(__magic_name__ ) for i, t in enumerate(scheduler.timesteps ): lowerCamelCase : Optional[int] = model(__magic_name__ , __magic_name__ ) lowerCamelCase : Union[str, Any] = scheduler.step(__magic_name__ , __magic_name__ , __magic_name__ ).prev_sample assert sample.dtype == torch.floataa
287
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=__SCREAMING_SNAKE_CASE) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = field(default="""automatic-speech-recognition""" , metadata={"""include_in_asdict_even_if_is_default""": True}) _UpperCAmelCase : ClassVar[Features] = Features({"""audio""": Audio()}) _UpperCAmelCase : ClassVar[Features] = Features({"""transcription""": Value("""string""")}) _UpperCAmelCase : str = "audio" _UpperCAmelCase : str = "transcription" def UpperCamelCase__ ( self , __magic_name__ ): if self.audio_column not in features: raise ValueError(F'''Column {self.audio_column} is not present in features.''' ) if not isinstance(features[self.audio_column] , __magic_name__ ): raise ValueError(F'''Column {self.audio_column} is not an Audio type.''' ) lowerCamelCase : Optional[Any] = copy.deepcopy(self ) lowerCamelCase : List[Any] = self.input_schema.copy() lowerCamelCase : Tuple = features[self.audio_column] lowerCamelCase : int = input_schema return task_template @property def UpperCamelCase__ ( self ): return {self.audio_column: "audio", self.transcription_column: "transcription"}
287
1
def _a ( lowerCamelCase ): if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True lowerCamelCase : Any = 4 lowerCamelCase : List[str] = (1 << p) - 1 for _ in range(p - 2 ): lowerCamelCase : List[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(1_1))
287
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower_vision_model""" def __init__( self , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=3 , __magic_name__=1_6 , __magic_name__=2_8_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__=True , __magic_name__=False , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : str = num_hidden_layers lowerCamelCase : Optional[int] = num_channels lowerCamelCase : List[str] = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Any = initializer_factor lowerCamelCase : Tuple = layer_norm_eps lowerCamelCase : Tuple = stop_gradient lowerCamelCase : Optional[int] = share_layernorm lowerCamelCase : str = remove_last_layer @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : str = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """bridgetower_text_model""" def __init__( self , __magic_name__=5_0_2_6_5 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=1 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_4 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Any = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : Tuple = hidden_act lowerCamelCase : Optional[int] = initializer_factor lowerCamelCase : Any = intermediate_size lowerCamelCase : List[str] = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : str = max_position_embeddings lowerCamelCase : Union[str, Any] = type_vocab_size lowerCamelCase : Optional[int] = layer_norm_eps lowerCamelCase : Optional[int] = position_embedding_type lowerCamelCase : List[str] = use_cache lowerCamelCase : List[str] = pad_token_id lowerCamelCase : List[str] = bos_token_id lowerCamelCase : Optional[int] = eos_token_id @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): lowerCamelCase , lowerCamelCase : int = cls.get_config_dict(__magic_name__ , **__magic_name__ ) if config_dict.get("""model_type""" ) == "bridgetower": lowerCamelCase : Optional[int] = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__magic_name__ , **__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Dict = """bridgetower""" def __init__( self , __magic_name__=True , __magic_name__="gelu" , __magic_name__=7_6_8 , __magic_name__=1 , __magic_name__=1e-05 , __magic_name__=False , __magic_name__="add" , __magic_name__=1_2 , __magic_name__=6 , __magic_name__=False , __magic_name__=False , __magic_name__=None , __magic_name__=None , **__magic_name__ , ): # TODO: remove this once the Hub files are updated. lowerCamelCase : int = kwargs.pop("""text_config_dict""" , __magic_name__ ) lowerCamelCase : str = kwargs.pop("""vision_config_dict""" , __magic_name__ ) super().__init__(**__magic_name__ ) lowerCamelCase : str = share_cross_modal_transformer_layers lowerCamelCase : Union[str, Any] = hidden_act lowerCamelCase : str = hidden_size lowerCamelCase : Tuple = initializer_factor lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = share_link_tower_layers lowerCamelCase : List[Any] = link_tower_type lowerCamelCase : Tuple = num_attention_heads lowerCamelCase : int = num_hidden_layers lowerCamelCase : Union[str, Any] = tie_word_embeddings lowerCamelCase : Tuple = init_layernorm_from_vision_encoder if text_config is None: lowerCamelCase : Any = {} logger.info("""`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.""" ) if vision_config is None: lowerCamelCase : int = {} logger.info("""`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.""" ) lowerCamelCase : Any = BridgeTowerTextConfig(**__magic_name__ ) lowerCamelCase : Optional[Any] = BridgeTowerVisionConfig(**__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , **__magic_name__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : str = copy.deepcopy(self.__dict__ ) lowerCamelCase : int = self.text_config.to_dict() lowerCamelCase : Dict = self.vision_config.to_dict() lowerCamelCase : List[str] = self.__class__.model_type return output
287
1
import argparse import struct import unittest class A__ : def __init__( self , __magic_name__ ): lowerCamelCase : List[str] = data # Initialize hash values lowerCamelCase : List[str] = [ 0X6A09E667, 0XBB67AE85, 0X3C6EF372, 0XA54FF53A, 0X510E527F, 0X9B05688C, 0X1F83D9AB, 0X5BE0CD19, ] # Initialize round constants lowerCamelCase : Dict = [ 0X428A2F98, 0X71374491, 0XB5C0FBCF, 0XE9B5DBA5, 0X3956C25B, 0X59F111F1, 0X923F82A4, 0XAB1C5ED5, 0XD807AA98, 0X12835B01, 0X243185BE, 0X550C7DC3, 0X72BE5D74, 0X80DEB1FE, 0X9BDC06A7, 0XC19BF174, 0XE49B69C1, 0XEFBE4786, 0X0FC19DC6, 0X240CA1CC, 0X2DE92C6F, 0X4A7484AA, 0X5CB0A9DC, 0X76F988DA, 0X983E5152, 0XA831C66D, 0XB00327C8, 0XBF597FC7, 0XC6E00BF3, 0XD5A79147, 0X06CA6351, 0X14292967, 0X27B70A85, 0X2E1B2138, 0X4D2C6DFC, 0X53380D13, 0X650A7354, 0X766A0ABB, 0X81C2C92E, 0X92722C85, 0XA2BFE8A1, 0XA81A664B, 0XC24B8B70, 0XC76C51A3, 0XD192E819, 0XD6990624, 0XF40E3585, 0X106AA070, 0X19A4C116, 0X1E376C08, 0X2748774C, 0X34B0BCB5, 0X391C0CB3, 0X4ED8AA4A, 0X5B9CCA4F, 0X682E6FF3, 0X748F82EE, 0X78A5636F, 0X84C87814, 0X8CC70208, 0X90BEFFFA, 0XA4506CEB, 0XBEF9A3F7, 0XC67178F2, ] lowerCamelCase : Optional[int] = self.preprocessing(self.data ) self.final_hash() @staticmethod def UpperCamelCase__ ( __magic_name__ ): lowerCamelCase : Optional[int] = b"""\x80""" + (b"""\x00""" * (6_3 - (len(__magic_name__ ) + 8) % 6_4)) lowerCamelCase : str = struct.pack(""">Q""" , (len(__magic_name__ ) * 8) ) return data + padding + big_endian_integer def UpperCamelCase__ ( self ): # Convert into blocks of 64 bytes lowerCamelCase : Optional[Any] = [ self.preprocessed_data[x : x + 6_4] for x in range(0 , len(self.preprocessed_data ) , 6_4 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers lowerCamelCase : List[str] = list(struct.unpack(""">16L""" , __magic_name__ ) ) # add 48 0-ed integers words += [0] * 4_8 lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[Any] = self.hashes for index in range(0 , 6_4 ): if index > 1_5: # modify the zero-ed indexes at the end of the array lowerCamelCase : Optional[Any] = ( self.ror(words[index - 1_5] , 7 ) ^ self.ror(words[index - 1_5] , 1_8 ) ^ (words[index - 1_5] >> 3) ) lowerCamelCase : Dict = ( self.ror(words[index - 2] , 1_7 ) ^ self.ror(words[index - 2] , 1_9 ) ^ (words[index - 2] >> 1_0) ) lowerCamelCase : str = ( words[index - 1_6] + sa + words[index - 7] + sa ) % 0X100000000 # Compression lowerCamelCase : List[Any] = self.ror(__magic_name__ , 6 ) ^ self.ror(__magic_name__ , 1_1 ) ^ self.ror(__magic_name__ , 2_5 ) lowerCamelCase : List[str] = (e & f) ^ ((~e & 0XFFFFFFFF) & g) lowerCamelCase : Optional[Any] = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X100000000 lowerCamelCase : Optional[Any] = self.ror(__magic_name__ , 2 ) ^ self.ror(__magic_name__ , 1_3 ) ^ self.ror(__magic_name__ , 2_2 ) lowerCamelCase : Optional[Any] = (a & b) ^ (a & c) ^ (b & c) lowerCamelCase : Optional[int] = (sa + maj) % 0X100000000 lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[Any] = ( g, f, e, ((d + tempa) % 0X100000000), c, b, a, ((tempa + tempa) % 0X100000000), ) lowerCamelCase : List[str] = [a, b, c, d, e, f, g, h] # Modify final values lowerCamelCase : int = [ ((element + mutated_hash_values[index]) % 0X100000000) for index, element in enumerate(self.hashes ) ] lowerCamelCase : Union[str, Any] = """""".join([hex(__magic_name__ )[2:].zfill(8 ) for value in self.hashes] ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): return 0XFFFFFFFF & (value << (3_2 - rotations)) | (value >> rotations) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): import hashlib lowerCamelCase : int = bytes("""Test String""" , """utf-8""" ) self.assertEqual(SHAaaa(__magic_name__ ).hash , hashlib.shaaaa(__magic_name__ ).hexdigest() ) def _a ( ): import doctest doctest.testmod() lowerCamelCase : int = argparse.ArgumentParser() parser.add_argument( """-s""", """--string""", dest="""input_string""", default="""Hello World!! Welcome to Cryptography""", help="""Hash the string""", ) parser.add_argument( """-f""", """--file""", dest="""input_file""", help="""Hash contents of a file""" ) lowerCamelCase : Optional[Any] = parser.parse_args() lowerCamelCase : List[Any] = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file, """rb""" ) as f: lowerCamelCase : Dict = f.read() else: lowerCamelCase : List[str] = bytes(lowerCamelCase, """utf-8""" ) print(SHAaaa(lowerCamelCase ).hash ) if __name__ == "__main__": main()
287
def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = set(range(3, lowerCamelCase, 2 ) ) primes.add(2 ) for p in range(3, lowerCamelCase, 2 ): if p not in primes: continue primes.difference_update(set(range(p * p, lowerCamelCase, lowerCamelCase ) ) ) lowerCamelCase : Any = [float(lowerCamelCase ) for n in range(limit + 1 )] for p in primes: for n in range(lowerCamelCase, limit + 1, lowerCamelCase ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'''{solution() = }''')
287
1
from collections import namedtuple _lowerCamelCase =namedtuple("""from_to""", """from_ to""") _lowerCamelCase ={ """cubicmeter""": from_to(1, 1), """litre""": from_to(0.001, 1_0_0_0), """kilolitre""": from_to(1, 1), """gallon""": from_to(0.00454, 264.172), """cubicyard""": from_to(0.76455, 1.30795), """cubicfoot""": from_to(0.028, 35.3147), """cup""": from_to(0.000236588, 4226.75), } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if from_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'from_type\' value: {from_type!r} Supported values are:\n''' + """, """.join(lowerCamelCase ) ) if to_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'to_type\' value: {to_type!r}. Supported values are:\n''' + """, """.join(lowerCamelCase ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
287
from __future__ import annotations import collections import pprint from pathlib import Path def _a ( lowerCamelCase ): return "".join(sorted(lowerCamelCase ) ) def _a ( lowerCamelCase ): return word_by_signature[signature(lowerCamelCase )] _lowerCamelCase =Path(__file__).parent.joinpath("""words.txt""").read_text(encoding="""utf-8""") _lowerCamelCase =sorted({word.strip().lower() for word in data.splitlines()}) _lowerCamelCase =collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _lowerCamelCase ={word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open("""anagrams.txt""", """w""") as file: file.write("""all_anagrams = \n """) file.write(pprint.pformat(all_anagrams))
287
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : str = TFAutoModelForSeqaSeqLM.from_pretrained("""google/mt5-small""" ) lowerCamelCase : int = AutoTokenizer.from_pretrained("""google/mt5-small""" ) lowerCamelCase : Tuple = tokenizer("""Hello there""" , return_tensors="""tf""" ).input_ids lowerCamelCase : int = tokenizer("""Hi I am""" , return_tensors="""tf""" ).input_ids lowerCamelCase : Dict = model(__magic_name__ , labels=__magic_name__ ).loss lowerCamelCase : Any = -tf.math.reduce_mean(__magic_name__ ).numpy() lowerCamelCase : List[str] = -21.228_168 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2e-4 )
287
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = ["""pixel_values"""] def __init__( self , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , __magic_name__ = PILImageResampling.BILINEAR , __magic_name__ = True , __magic_name__ = 1 / 2_5_5 , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = size if size is not None else {"""shortest_edge""": 3_8_4} lowerCamelCase : Tuple = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Dict = do_resize lowerCamelCase : List[Any] = size # Default value set here for backwards compatibility where the value in config is None lowerCamelCase : Any = crop_pct if crop_pct is not None else 2_2_4 / 2_5_6 lowerCamelCase : Union[str, Any] = resample lowerCamelCase : str = do_rescale lowerCamelCase : Union[str, Any] = rescale_factor lowerCamelCase : Tuple = do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCamelCase : int = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = PILImageResampling.BICUBIC , __magic_name__ = None , **__magic_name__ , ): lowerCamelCase : Union[str, Any] = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) lowerCamelCase : str = size["""shortest_edge"""] if shortest_edge < 3_8_4: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct lowerCamelCase : List[str] = int(shortest_edge / crop_pct ) lowerCamelCase : Optional[Any] = get_resize_output_image_size(__magic_name__ , size=__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Optional[int] = resize(image=__magic_name__ , size=__magic_name__ , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=__magic_name__ , size=(shortest_edge, shortest_edge) , data_format=__magic_name__ , **__magic_name__ ) else: # warping (no cropping) when evaluated at 384 or larger return resize( __magic_name__ , size=(shortest_edge, shortest_edge) , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return rescale(__magic_name__ , scale=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return normalize(__magic_name__ , mean=__magic_name__ , std=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = ChannelDimension.FIRST , **__magic_name__ , ): lowerCamelCase : str = do_resize if do_resize is not None else self.do_resize lowerCamelCase : Optional[Any] = crop_pct if crop_pct is not None else self.crop_pct lowerCamelCase : Optional[int] = resample if resample is not None else self.resample lowerCamelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else self.image_mean lowerCamelCase : Tuple = image_std if image_std is not None else self.image_std lowerCamelCase : Dict = size if size is not None else self.size lowerCamelCase : str = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : List[str] = make_list_of_images(__magic_name__ ) if not valid_images(__magic_name__ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_resize and size["shortest_edge"] < 3_8_4 and crop_pct is None: raise ValueError("""crop_pct must be specified if size < 384.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. lowerCamelCase : Optional[Any] = [to_numpy_array(__magic_name__ ) for image in images] if do_resize: lowerCamelCase : List[Any] = [self.resize(image=__magic_name__ , size=__magic_name__ , crop_pct=__magic_name__ , resample=__magic_name__ ) for image in images] if do_rescale: lowerCamelCase : Union[str, Any] = [self.rescale(image=__magic_name__ , scale=__magic_name__ ) for image in images] if do_normalize: lowerCamelCase : List[Any] = [self.normalize(image=__magic_name__ , mean=__magic_name__ , std=__magic_name__ ) for image in images] lowerCamelCase : Optional[int] = [to_channel_dimension_format(__magic_name__ , __magic_name__ ) for image in images] lowerCamelCase : List[str] = {"""pixel_values""": images} return BatchFeature(data=__magic_name__ , tensor_type=__magic_name__ )
287
1
from __future__ import annotations _lowerCamelCase =[True] * 1_0_0_0_0_0_1 _lowerCamelCase =2 while i * i <= 1_0_0_0_0_0_0: if seive[i]: for j in range(i * i, 1_0_0_0_0_0_1, i): _lowerCamelCase =False i += 1 def _a ( lowerCamelCase ): return seive[n] def _a ( lowerCamelCase ): return any(digit in """02468""" for digit in str(lowerCamelCase ) ) def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : List[str] = [2] # result already includes the number 2. for num in range(3, limit + 1, 2 ): if is_prime(lowerCamelCase ) and not contains_an_even_digit(lowerCamelCase ): lowerCamelCase : List[Any] = str(lowerCamelCase ) lowerCamelCase : int = [int(str_num[j:] + str_num[:j] ) for j in range(len(lowerCamelCase ) )] if all(is_prime(lowerCamelCase ) for i in list_nums ): result.append(lowerCamelCase ) return result def _a ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(f'''{len(find_circular_primes()) = }''')
287
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 _lowerCamelCase ={ """return_dict""": False, """output_hidden_states""": True, """output_attentions""": True, """torchscript""": True, """torch_dtype""": """float16""", """use_bfloat16""": True, """tf_legacy_loss""": True, """pruned_heads""": {"""a""": 1}, """tie_word_embeddings""": False, """is_decoder""": True, """cross_attention_hidden_size""": 1_2_8, """add_cross_attention""": True, """tie_encoder_decoder""": True, """max_length""": 5_0, """min_length""": 3, """do_sample""": True, """early_stopping""": True, """num_beams""": 3, """num_beam_groups""": 3, """diversity_penalty""": 0.5, """temperature""": 2.0, """top_k""": 1_0, """top_p""": 0.7, """typical_p""": 0.2, """repetition_penalty""": 0.8, """length_penalty""": 0.8, """no_repeat_ngram_size""": 5, """encoder_no_repeat_ngram_size""": 5, """bad_words_ids""": [1, 2, 3], """num_return_sequences""": 3, """chunk_size_feed_forward""": 5, """output_scores""": True, """return_dict_in_generate""": True, """forced_bos_token_id""": 2, """forced_eos_token_id""": 3, """remove_invalid_values""": True, """architectures""": ["""BertModel"""], """finetuning_task""": """translation""", """id2label""": {0: """label"""}, """label2id""": {"""label""": """0"""}, """tokenizer_class""": """BertTokenizerFast""", """prefix""": """prefix""", """bos_token_id""": 6, """pad_token_id""": 7, """eos_token_id""": 8, """sep_token_id""": 9, """decoder_start_token_id""": 1_0, """exponential_decay_length_penalty""": (5, 1.01), """suppress_tokens""": [0, 1], """begin_suppress_tokens""": 2, """task_specific_params""": {"""translation""": """some_params"""}, """problem_type""": """regression""", } @is_staging_test class A__ ( unittest.TestCase): @classmethod def UpperCamelCase__ ( cls ): lowerCamelCase : int = TOKEN HfFolder.save_token(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls ): try: delete_repo(token=cls._token , repo_id="""test-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-config-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-config""" ) except HTTPError: pass def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""test-config""" , use_auth_token=self._token ) lowerCamelCase : Any = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__magic_name__ , repo_id="""test-config""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : Optional[Any] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""valid_org/test-config-org""" , use_auth_token=self._token ) lowerCamelCase : Optional[int] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( __magic_name__ , repo_id="""valid_org/test-config-org""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : List[str] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): CustomConfig.register_for_auto_class() lowerCamelCase : Optional[Any] = CustomConfig(attribute=4_2 ) config.push_to_hub("""test-dynamic-config""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"""AutoConfig""": """custom_configuration.CustomConfig"""} ) lowerCamelCase : List[str] = AutoConfig.from_pretrained(F'''{USER}/test-dynamic-config''' , trust_remote_code=__magic_name__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , """CustomConfig""" ) self.assertEqual(new_config.attribute , 4_2 ) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : str = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated lowerCamelCase : Optional[int] = c.n_embd + 1 # int lowerCamelCase : Optional[int] = c.resid_pdrop + 1.0 # float lowerCamelCase : Tuple = not c.scale_attn_weights # bool lowerCamelCase : Any = c.summary_type + """foo""" # str c.update_from_string( F'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(__magic_name__ , c.n_embd , """mismatch for key: n_embd""" ) self.assertEqual(__magic_name__ , c.resid_pdrop , """mismatch for key: resid_pdrop""" ) self.assertEqual(__magic_name__ , c.scale_attn_weights , """mismatch for key: scale_attn_weights""" ) self.assertEqual(__magic_name__ , c.summary_type , """mismatch for key: summary_type""" ) def UpperCamelCase__ ( self ): lowerCamelCase : str = PretrainedConfig() lowerCamelCase : int = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( __magic_name__ , ["""is_encoder_decoder""", """_name_or_path""", """_commit_hash""", """transformers_version"""] ) lowerCamelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(__magic_name__ , __magic_name__ )] if len(__magic_name__ ) > 0: raise ValueError( """The following keys are set with the default values in""" """ `test_configuration_common.config_common_kwargs` pick another value for them:""" F''' {", ".join(__magic_name__ )}.''' ) def UpperCamelCase__ ( self ): with self.assertRaises(__magic_name__ ): # config is in subfolder, the following should not work without specifying the subfolder lowerCamelCase : Dict = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" ) lowerCamelCase : str = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" , subfolder="""bert""" ) self.assertIsNotNone(__magic_name__ ) def UpperCamelCase__ ( self ): # A mock response for an HTTP head request to emulate server down lowerCamelCase : Dict = mock.Mock() lowerCamelCase : Optional[int] = 5_0_0 lowerCamelCase : List[Any] = {} lowerCamelCase : Tuple = HTTPError lowerCamelCase : Union[str, Any] = {} # Download this model to make sure it's in the cache. lowerCamelCase : List[str] = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=__magic_name__ ) as mock_head: lowerCamelCase : Any = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase__ ( self ): # This test is for deprecated behavior and can be removed in v5 lowerCamelCase : List[str] = BertConfig.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = AutoConfig.from_pretrained("""bert-base-cased""" ) lowerCamelCase : Optional[Any] = ["""config.4.0.0.json"""] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(__magic_name__ ) lowerCamelCase : str = 2 json.dump(configuration.to_dict() , open(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , """w""" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 lowerCamelCase : Any = ["""config.42.0.0.json"""] lowerCamelCase : Optional[Any] = 7_6_8 configuration.save_pretrained(__magic_name__ ) shutil.move(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , os.path.join(__magic_name__ , """config.42.0.0.json""" ) ) lowerCamelCase : int = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 7_6_8 ) def UpperCamelCase__ ( self ): # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. lowerCamelCase : str = """hf-internal-testing/test-two-configs""" import transformers as new_transformers lowerCamelCase : Tuple = """v4.0.0""" lowerCamelCase , lowerCamelCase : Optional[int] = new_transformers.models.auto.AutoConfig.from_pretrained( __magic_name__ , return_unused_kwargs=__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(__magic_name__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers lowerCamelCase : Tuple = """v3.0.0""" lowerCamelCase : Any = old_transformers.models.auto.AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(old_configuration.hidden_size , 7_6_8 )
287
1
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _lowerCamelCase =1_6 _lowerCamelCase =3_2 def _a ( lowerCamelCase, lowerCamelCase = 16 ): lowerCamelCase : Any = AutoTokenizer.from_pretrained("""bert-base-cased""" ) lowerCamelCase : Tuple = load_dataset("""glue""", """mrpc""" ) def tokenize_function(lowerCamelCase ): # max_length=None => use the model max length (it's actually the default) lowerCamelCase : Tuple = tokenizer(examples["""sentence1"""], examples["""sentence2"""], truncation=lowerCamelCase, max_length=lowerCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowerCamelCase : Optional[int] = datasets.map( lowerCamelCase, batched=lowerCamelCase, remove_columns=["""idx""", """sentence1""", """sentence2"""], ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowerCamelCase : Union[str, Any] = tokenized_datasets.rename_column("""label""", """labels""" ) def collate_fn(lowerCamelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. lowerCamelCase : List[Any] = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowerCamelCase : Union[str, Any] = 16 elif accelerator.mixed_precision != "no": lowerCamelCase : Optional[int] = 8 else: lowerCamelCase : Tuple = None return tokenizer.pad( lowerCamelCase, padding="""longest""", max_length=lowerCamelCase, pad_to_multiple_of=lowerCamelCase, return_tensors="""pt""", ) # Instantiate dataloaders. lowerCamelCase : Tuple = DataLoader( tokenized_datasets["""train"""], shuffle=lowerCamelCase, collate_fn=lowerCamelCase, batch_size=lowerCamelCase ) lowerCamelCase : Tuple = DataLoader( tokenized_datasets["""validation"""], shuffle=lowerCamelCase, collate_fn=lowerCamelCase, batch_size=lowerCamelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders _lowerCamelCase =mocked_dataloaders # noqa: F811 def _a ( lowerCamelCase, lowerCamelCase ): # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", lowerCamelCase ) == "1": lowerCamelCase : Optional[Any] = 2 # Initialize accelerator lowerCamelCase : int = Accelerator(cpu=args.cpu, mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowerCamelCase : Union[str, Any] = config["""lr"""] lowerCamelCase : Any = int(config["""num_epochs"""] ) lowerCamelCase : int = int(config["""seed"""] ) lowerCamelCase : Optional[int] = int(config["""batch_size"""] ) lowerCamelCase : Tuple = evaluate.load("""glue""", """mrpc""" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=lowerCamelCase ) def inner_training_loop(lowerCamelCase ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(lowerCamelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowerCamelCase : str = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""", return_dict=lowerCamelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowerCamelCase : str = model.to(accelerator.device ) # Instantiate optimizer lowerCamelCase : Dict = AdamW(params=model.parameters(), lr=lowerCamelCase ) lowerCamelCase , lowerCamelCase : Dict = get_dataloaders(lowerCamelCase, lowerCamelCase ) # Instantiate scheduler lowerCamelCase : List[str] = get_linear_schedule_with_warmup( optimizer=lowerCamelCase, num_warmup_steps=100, num_training_steps=(len(lowerCamelCase ) * num_epochs), ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Tuple = accelerator.prepare( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Now we train the model for epoch in range(lowerCamelCase ): model.train() for step, batch in enumerate(lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowerCamelCase : Any = model(**lowerCamelCase ) lowerCamelCase : List[Any] = outputs.loss accelerator.backward(lowerCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowerCamelCase : Tuple = model(**lowerCamelCase ) lowerCamelCase : Optional[int] = outputs.logits.argmax(dim=-1 ) lowerCamelCase , lowerCamelCase : int = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=lowerCamelCase, references=lowerCamelCase, ) lowerCamelCase : Tuple = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''', lowerCamelCase ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def _a ( ): lowerCamelCase : Any = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""", type=lowerCamelCase, default=lowerCamelCase, choices=["""no""", """fp16""", """bf16""", """fp8"""], help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""", ) parser.add_argument("""--cpu""", action="""store_true""", help="""If passed, will train on the CPU.""" ) lowerCamelCase : Optional[Any] = parser.parse_args() lowerCamelCase : str = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(lowerCamelCase, lowerCamelCase ) if __name__ == "__main__": main()
287
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase): @slow def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) lowerCamelCase : Any = tf.convert_to_tensor( [[5, 1_2_1, 1_1, 6_6_0, 1_6, 7_3_0, 2_5_5_4_3, 1_1_0, 8_3, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowerCamelCase : str = model(__magic_name__ )["""last_hidden_state"""] lowerCamelCase : Union[str, Any] = tf.TensorShape((1, 1_0, 7_6_8) ) self.assertEqual(output.shape , __magic_name__ ) # compare the actual values for a slice. lowerCamelCase : Dict = tf.convert_to_tensor( [[[-0.0_254, 0.0_235, 0.1_027], [0.0_606, -0.1_811, -0.0_418], [-0.1_561, -0.1_127, 0.2_687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
287
1
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """microsoft/conditional-detr-resnet-50""": ( """https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = """conditional_detr""" _UpperCAmelCase : Optional[int] = ["""past_key_values"""] _UpperCAmelCase : Optional[Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , __magic_name__=True , __magic_name__=None , __magic_name__=3 , __magic_name__=3_0_0 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=6 , __magic_name__=2_0_4_8 , __magic_name__=8 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=True , __magic_name__="relu" , __magic_name__=2_5_6 , __magic_name__=0.1 , __magic_name__=0.0 , __magic_name__=0.0 , __magic_name__=0.02 , __magic_name__=1.0 , __magic_name__=False , __magic_name__="sine" , __magic_name__="resnet50" , __magic_name__=True , __magic_name__=False , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=1 , __magic_name__=1 , __magic_name__=2 , __magic_name__=5 , __magic_name__=2 , __magic_name__=0.25 , **__magic_name__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) lowerCamelCase : Optional[int] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(__magic_name__ , __magic_name__ ): lowerCamelCase : List[Any] = backbone_config.get("""model_type""" ) lowerCamelCase : Dict = CONFIG_MAPPING[backbone_model_type] lowerCamelCase : str = config_class.from_dict(__magic_name__ ) lowerCamelCase : Dict = use_timm_backbone lowerCamelCase : str = backbone_config lowerCamelCase : Tuple = num_channels lowerCamelCase : Dict = num_queries lowerCamelCase : Any = d_model lowerCamelCase : Optional[Any] = encoder_ffn_dim lowerCamelCase : List[str] = encoder_layers lowerCamelCase : Union[str, Any] = encoder_attention_heads lowerCamelCase : Any = decoder_ffn_dim lowerCamelCase : Dict = decoder_layers lowerCamelCase : Union[str, Any] = decoder_attention_heads lowerCamelCase : Dict = dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Union[str, Any] = activation_dropout lowerCamelCase : Optional[int] = activation_function lowerCamelCase : int = init_std lowerCamelCase : str = init_xavier_std lowerCamelCase : Tuple = encoder_layerdrop lowerCamelCase : str = decoder_layerdrop lowerCamelCase : Tuple = encoder_layers lowerCamelCase : Optional[int] = auxiliary_loss lowerCamelCase : Optional[Any] = position_embedding_type lowerCamelCase : Optional[int] = backbone lowerCamelCase : Union[str, Any] = use_pretrained_backbone lowerCamelCase : str = dilation # Hungarian matcher lowerCamelCase : Optional[Any] = class_cost lowerCamelCase : Dict = bbox_cost lowerCamelCase : Tuple = giou_cost # Loss coefficients lowerCamelCase : Union[str, Any] = mask_loss_coefficient lowerCamelCase : Dict = dice_loss_coefficient lowerCamelCase : Optional[int] = cls_loss_coefficient lowerCamelCase : Optional[int] = bbox_loss_coefficient lowerCamelCase : Optional[int] = giou_loss_coefficient lowerCamelCase : Optional[int] = focal_alpha super().__init__(is_encoder_decoder=__magic_name__ , **__magic_name__ ) @property def UpperCamelCase__ ( self ): return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): return self.d_model def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: lowerCamelCase : Optional[int] = self.backbone_config.to_dict() lowerCamelCase : Optional[Any] = self.__class__.model_type return output class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = version.parse("""1.11""") @property def UpperCamelCase__ ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def UpperCamelCase__ ( self ): return 1e-5 @property def UpperCamelCase__ ( self ): return 1_2
287
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _lowerCamelCase =get_logger(__name__) class A__ : def __init__( self , __magic_name__ = None ): lowerCamelCase : Dict = ( os.path.join(__magic_name__ , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) lowerCamelCase : List[str] = Extractor def UpperCamelCase__ ( self , __magic_name__ ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" lowerCamelCase : int = os.path.abspath(__magic_name__ ) return os.path.join(self.extract_dir , hash_url_to_filename(__magic_name__ ) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ ): return force_extract or ( not os.path.isfile(__magic_name__ ) and not (os.path.isdir(__magic_name__ ) and os.listdir(__magic_name__ )) ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = False ): lowerCamelCase : Union[str, Any] = self.extractor.infer_extractor_format(__magic_name__ ) if not extractor_format: return input_path lowerCamelCase : int = self._get_output_path(__magic_name__ ) if self._do_extract(__magic_name__ , __magic_name__ ): self.extractor.extract(__magic_name__ , __magic_name__ , __magic_name__ ) return output_path class A__ ( __SCREAMING_SNAKE_CASE): @classmethod @abstractmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): ... @staticmethod @abstractmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): ... class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[bytes] = [] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with open(__magic_name__ , """rb""" ) as f: return f.read(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if not magic_number: lowerCamelCase : Optional[Any] = max(len(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) try: lowerCamelCase : Tuple = cls.read_magic_number(__magic_name__ , __magic_name__ ) except OSError: return False return any(magic_number.startswith(__magic_name__ ) for cls_magic_number in cls.magic_numbers ) class A__ ( __SCREAMING_SNAKE_CASE): @classmethod def UpperCamelCase__ ( cls , __magic_name__ , **__magic_name__ ): return tarfile.is_tarfile(__magic_name__ ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): def resolved(__magic_name__ ) -> str: return os.path.realpath(os.path.abspath(__magic_name__ ) ) def badpath(__magic_name__ , __magic_name__ ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__magic_name__ , __magic_name__ ) ).startswith(__magic_name__ ) def badlink(__magic_name__ , __magic_name__ ) -> bool: # Links are interpreted relative to the directory containing the link lowerCamelCase : List[str] = resolved(os.path.join(__magic_name__ , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=__magic_name__ ) lowerCamelCase : Optional[Any] = resolved(__magic_name__ ) for finfo in members: if badpath(finfo.name , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(__magic_name__ , __magic_name__ ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Dict = tarfile.open(__magic_name__ ) tar_file.extractall(__magic_name__ , members=TarExtractor.safemembers(__magic_name__ , __magic_name__ ) ) tar_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : str = [B"""\x1F\x8B"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with gzip.open(__magic_name__ , """rb""" ) as gzip_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Optional[int] = [ B"""PK\x03\x04""", B"""PK\x05\x06""", # empty archive B"""PK\x07\x08""", # spanned archive ] @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = b"" ): if super().is_extractable(__magic_name__ , magic_number=__magic_name__ ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__magic_name__ , """rb""" ) as fp: lowerCamelCase : List[str] = _EndRecData(__magic_name__ ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: lowerCamelCase : List[Any] = fp.read(__magic_name__ ) # CD is where we expect it to be if len(__magic_name__ ) == sizeCentralDir: lowerCamelCase : str = struct.unpack(__magic_name__ , __magic_name__ ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with zipfile.ZipFile(__magic_name__ , """r""" ) as zip_file: zip_file.extractall(__magic_name__ ) zip_file.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[str] = [B"""\xFD\x37\x7A\x58\x5A\x00"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with lzma.open(__magic_name__ ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""Rar!\x1a\x07\x00""", B"""Rar!\x1a\x07\x01\x00"""] # RAR_ID # RAR5_ID @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) lowerCamelCase : Union[str, Any] = rarfile.RarFile(__magic_name__ ) rf.extractall(__magic_name__ ) rf.close() class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = [B"""\x28\xb5\x2F\xFD"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd lowerCamelCase : int = zstd.ZstdDecompressor() with open(__magic_name__ , """rb""" ) as ifh, open(__magic_name__ , """wb""" ) as ofh: dctx.copy_stream(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Any = [B"""\x42\x5A\x68"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): with bza.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x37\x7A\xBC\xAF\x27\x1C"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(__magic_name__ , exist_ok=__magic_name__ ) with pyazr.SevenZipFile(__magic_name__ , """r""" ) as archive: archive.extractall(__magic_name__ ) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : List[Any] = [B"""\x04\x22\x4D\x18"""] @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(__magic_name__ , """rb""" ) as compressed_file: with open(__magic_name__ , """wb""" ) as extracted_file: shutil.copyfileobj(__magic_name__ , __magic_name__ ) class A__ : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) _UpperCAmelCase : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def UpperCamelCase__ ( cls ): return max( len(__magic_name__ ) for extractor in cls.extractors.values() if issubclass(__magic_name__ , __magic_name__ ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def UpperCamelCase__ ( __magic_name__ , __magic_name__ ): try: return MagicNumberBaseExtractor.read_magic_number(__magic_name__ , magic_number_length=__magic_name__ ) except OSError: return b"" @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ = False ): warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = cls.infer_extractor_format(__magic_name__ ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def UpperCamelCase__ ( cls , __magic_name__ ): # <Added version="2.4.0"/> lowerCamelCase : Dict = cls._get_magic_number_max_length() lowerCamelCase : Optional[Any] = cls._read_magic_number(__magic_name__ , __magic_name__ ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__magic_name__ , magic_number=__magic_name__ ): return extractor_format @classmethod def UpperCamelCase__ ( cls , __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = "deprecated" , ): os.makedirs(os.path.dirname(__magic_name__ ) , exist_ok=__magic_name__ ) # Prevent parallel extractions lowerCamelCase : Tuple = str(Path(__magic_name__ ).with_suffix(""".lock""" ) ) with FileLock(__magic_name__ ): shutil.rmtree(__magic_name__ , ignore_errors=__magic_name__ ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__magic_name__ , __magic_name__ ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=__magic_name__ , ) lowerCamelCase : int = extractor if extractor != """deprecated""" else extractor_format else: lowerCamelCase : Optional[int] = cls.extractors[extractor_format] return extractor.extract(__magic_name__ , __magic_name__ ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=__magic_name__ , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__magic_name__ ): return extractor.extract(__magic_name__ , __magic_name__ )
287
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import AlignProcessor, EfficientNetImageProcessor @require_vision class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = tempfile.mkdtemp() lowerCamelCase : Tuple = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] lowerCamelCase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) lowerCamelCase : Any = { """do_resize""": True, """size""": 2_0, """do_center_crop""": True, """crop_size""": 1_8, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } lowerCamelCase : int = os.path.join(self.tmpdirname , __magic_name__ ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self , **__magic_name__ ): return BertTokenizer.from_pretrained(self.tmpdirname , **__magic_name__ ) def UpperCamelCase__ ( self , **__magic_name__ ): return BertTokenizerFast.from_pretrained(self.tmpdirname , **__magic_name__ ) def UpperCamelCase__ ( self , **__magic_name__ ): return EfficientNetImageProcessor.from_pretrained(self.tmpdirname , **__magic_name__ ) def UpperCamelCase__ ( self ): shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] lowerCamelCase : Any = [Image.fromarray(np.moveaxis(__magic_name__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase__ ( self ): lowerCamelCase : Dict = self.get_tokenizer() lowerCamelCase : List[Any] = self.get_rust_tokenizer() lowerCamelCase : Tuple = self.get_image_processor() lowerCamelCase : List[Any] = AlignProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) processor_slow.save_pretrained(self.tmpdirname ) lowerCamelCase : List[str] = AlignProcessor.from_pretrained(self.tmpdirname , use_fast=__magic_name__ ) lowerCamelCase : List[str] = AlignProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) processor_fast.save_pretrained(self.tmpdirname ) lowerCamelCase : int = AlignProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __magic_name__ ) self.assertIsInstance(processor_fast.tokenizer , __magic_name__ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __magic_name__ ) self.assertIsInstance(processor_fast.image_processor , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Any = AlignProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase : Optional[Any] = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) lowerCamelCase : List[str] = self.get_image_processor(do_normalize=__magic_name__ , padding_value=1.0 ) lowerCamelCase : List[Any] = AlignProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__magic_name__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __magic_name__ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : int = self.get_image_processor() lowerCamelCase : Union[str, Any] = self.get_tokenizer() lowerCamelCase : Optional[Any] = AlignProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) lowerCamelCase : str = self.prepare_image_inputs() lowerCamelCase : Tuple = image_processor(__magic_name__ , return_tensors="""np""" ) lowerCamelCase : Dict = processor(images=__magic_name__ , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCamelCase__ ( self ): lowerCamelCase : int = self.get_image_processor() lowerCamelCase : Any = self.get_tokenizer() lowerCamelCase : Union[str, Any] = AlignProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) lowerCamelCase : List[str] = """lower newer""" lowerCamelCase : List[str] = processor(text=__magic_name__ ) lowerCamelCase : List[str] = tokenizer(__magic_name__ , padding="""max_length""" , max_length=6_4 ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = self.get_image_processor() lowerCamelCase : Dict = self.get_tokenizer() lowerCamelCase : str = AlignProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) lowerCamelCase : int = """lower newer""" lowerCamelCase : Optional[int] = self.prepare_image_inputs() lowerCamelCase : Dict = processor(text=__magic_name__ , images=__magic_name__ ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """token_type_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__magic_name__ ): processor() def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = self.get_image_processor() lowerCamelCase : Optional[int] = self.get_tokenizer() lowerCamelCase : Dict = AlignProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) lowerCamelCase : Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase : List[Any] = processor.batch_decode(__magic_name__ ) lowerCamelCase : List[str] = tokenizer.batch_decode(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) def UpperCamelCase__ ( self ): lowerCamelCase : Tuple = self.get_image_processor() lowerCamelCase : int = self.get_tokenizer() lowerCamelCase : Optional[int] = AlignProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) lowerCamelCase : Dict = """lower newer""" lowerCamelCase : int = self.prepare_image_inputs() lowerCamelCase : int = processor(text=__magic_name__ , images=__magic_name__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
287
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration _lowerCamelCase =5_0_0_0_0_0 _lowerCamelCase , _lowerCamelCase =os.path.split(__file__) _lowerCamelCase =os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.map(**lowerCamelCase ) @get_duration def _a ( lowerCamelCase, **lowerCamelCase ): lowerCamelCase : Optional[Any] = dataset.filter(**lowerCamelCase ) def _a ( ): lowerCamelCase : Optional[Any] = {"""num examples""": SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : Any = datasets.Features({"""text""": datasets.Value("""string""" ), """numbers""": datasets.Value("""float32""" )} ) lowerCamelCase : Tuple = generate_example_dataset( os.path.join(lowerCamelCase, """dataset.arrow""" ), lowerCamelCase, num_examples=lowerCamelCase ) lowerCamelCase : Tuple = transformers.AutoTokenizer.from_pretrained("""bert-base-cased""", use_fast=lowerCamelCase ) def tokenize(lowerCamelCase ): return tokenizer(examples["""text"""] ) lowerCamelCase : List[str] = map(lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""numpy""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""pandas""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""torch""", columns="""numbers""" ): lowerCamelCase : List[str] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) with dataset.formatted_as(type="""tensorflow""", columns="""numbers""" ): lowerCamelCase : Optional[int] = map(lowerCamelCase, function=lambda lowerCamelCase : None, batched=lowerCamelCase ) lowerCamelCase : int = map(lowerCamelCase, function=lowerCamelCase, batched=lowerCamelCase ) lowerCamelCase : Union[str, Any] = filter(lowerCamelCase ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(lowerCamelCase, """wb""" ) as f: f.write(json.dumps(lowerCamelCase ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
287
1