code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Optional[Any] ) -> List[Any]: """simple docstring""" super().tearDown() gc.collect() def snake_case__ ( self : List[Any] ) -> List[str]: """simple docstring""" __lowercase , __lowercase = FlaxControlNetModel.from_pretrained( """lllyasviel/sd-controlnet-canny""" , from_pt=lowercase , dtype=jnp.bfloataa ) __lowercase , __lowercase = FlaxStableDiffusionControlNetPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , controlnet=lowercase , from_pt=lowercase , dtype=jnp.bfloataa ) __lowercase = controlnet_params __lowercase = """bird""" __lowercase = jax.device_count() __lowercase = pipe.prepare_text_inputs([prompts] * num_samples ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png""" ) __lowercase = pipe.prepare_image_inputs([canny_image] * num_samples ) __lowercase = jax.random.PRNGKey(0 ) __lowercase = jax.random.split(lowercase , jax.device_count() ) __lowercase = replicate(lowercase ) __lowercase = shard(lowercase ) __lowercase = shard(lowercase ) __lowercase = pipe( prompt_ids=lowercase , image=lowercase , params=lowercase , prng_seed=lowercase , num_inference_steps=50 , jit=lowercase , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) __lowercase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) __lowercase = images[0, 253:256, 253:256, -1] __lowercase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) __lowercase = jnp.array( [0.16_7969, 0.11_6699, 0.08_1543, 0.15_4297, 0.13_2812, 0.10_8887, 0.16_9922, 0.16_9922, 0.20_5078] ) print(F"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2 def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase , __lowercase = FlaxControlNetModel.from_pretrained( """lllyasviel/sd-controlnet-openpose""" , from_pt=lowercase , dtype=jnp.bfloataa ) __lowercase , __lowercase = FlaxStableDiffusionControlNetPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , controlnet=lowercase , from_pt=lowercase , dtype=jnp.bfloataa ) __lowercase = controlnet_params __lowercase = """Chef in the kitchen""" __lowercase = jax.device_count() __lowercase = pipe.prepare_text_inputs([prompts] * num_samples ) __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png""" ) __lowercase = pipe.prepare_image_inputs([pose_image] * num_samples ) __lowercase = jax.random.PRNGKey(0 ) __lowercase = jax.random.split(lowercase , jax.device_count() ) __lowercase = replicate(lowercase ) __lowercase = shard(lowercase ) __lowercase = shard(lowercase ) __lowercase = pipe( prompt_ids=lowercase , image=lowercase , params=lowercase , prng_seed=lowercase , num_inference_steps=50 , jit=lowercase , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) __lowercase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) __lowercase = images[0, 253:256, 253:256, -1] __lowercase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) __lowercase = jnp.array( [[0.27_1484, 0.26_1719, 0.27_5391, 0.27_7344, 0.27_9297, 0.29_1016, 0.29_4922, 0.30_2734, 0.30_2734]] ) print(F"output_slice: {output_slice}" ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
634
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "asapp/sew-tiny-100k": "https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json", # See all SEW models at https://huggingface.co/models?filter=sew } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """sew""" def __init__( self : List[Any] , lowercase : int=32 , lowercase : List[str]=768 , lowercase : Dict=12 , lowercase : str=12 , lowercase : str=3_072 , lowercase : Optional[int]=2 , lowercase : List[str]="gelu" , lowercase : List[str]=0.1 , lowercase : Tuple=0.1 , lowercase : Dict=0.1 , lowercase : Any=0.0 , lowercase : Dict=0.1 , lowercase : Optional[int]=0.1 , lowercase : List[str]=0.02 , lowercase : Dict=1E-5 , lowercase : Tuple="group" , lowercase : int="gelu" , lowercase : Any=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , lowercase : Optional[int]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowercase : Any=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowercase : List[str]=False , lowercase : Tuple=128 , lowercase : int=16 , lowercase : Union[str, Any]=True , lowercase : List[str]=0.05 , lowercase : Optional[int]=10 , lowercase : Any=2 , lowercase : Optional[Any]=0.0 , lowercase : Optional[Any]=10 , lowercase : int=0 , lowercase : Optional[int]="mean" , lowercase : List[Any]=False , lowercase : str=False , lowercase : int=256 , lowercase : str=0 , lowercase : List[Any]=1 , lowercase : List[Any]=2 , **lowercase : List[Any] , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase , pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase ) __lowercase = hidden_size __lowercase = feat_extract_norm __lowercase = feat_extract_activation __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = conv_bias __lowercase = num_conv_pos_embeddings __lowercase = num_conv_pos_embedding_groups __lowercase = len(self.conv_dim ) __lowercase = num_hidden_layers __lowercase = intermediate_size __lowercase = squeeze_factor __lowercase = hidden_act __lowercase = num_attention_heads __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = feat_proj_dropout __lowercase = final_dropout __lowercase = layerdrop __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" F"but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)" F"= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowercase = apply_spec_augment __lowercase = mask_time_prob __lowercase = mask_time_length __lowercase = mask_time_min_masks __lowercase = mask_feature_prob __lowercase = mask_feature_length __lowercase = mask_feature_min_masks # ctc loss __lowercase = ctc_loss_reduction __lowercase = ctc_zero_infinity # sequence classification __lowercase = use_weighted_layer_sum __lowercase = classifier_proj_size @property def snake_case__ ( self : Dict ) -> str: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
634
1
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml UpperCamelCase__ = NewType("DataClass", Any) UpperCamelCase__ = NewType("DataClassType", Any) def UpperCAmelCase__ ( lowercase__ ) -> List[Any]: if isinstance(lowercase__ , lowercase__ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F"Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive)." ) def UpperCAmelCase__ ( lowercase__ ) -> Callable[[str], Any]: __lowercase = {str(lowercase__ ): choice for choice in choices} return lambda lowercase__ : str_to_choice.get(lowercase__ , lowercase__ ) def UpperCAmelCase__ ( *, lowercase__ = None , lowercase__ = None , lowercase__ = dataclasses.MISSING , lowercase__ = dataclasses.MISSING , lowercase__ = None , **lowercase__ , ) -> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls __lowercase = {} if aliases is not None: __lowercase = aliases if help is not None: __lowercase = help return dataclasses.field(metadata=lowercase__ , default=lowercase__ , default_factory=lowercase__ , **lowercase__ ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Iterable[DataClassType] def __init__( self : Union[str, Any] , lowercase : Union[DataClassType, Iterable[DataClassType]] , **lowercase : Dict ) -> Dict: """simple docstring""" if "formatter_class" not in kwargs: __lowercase = ArgumentDefaultsHelpFormatter super().__init__(**lowercase ) if dataclasses.is_dataclass(lowercase ): __lowercase = [dataclass_types] __lowercase = list(lowercase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(lowercase ) @staticmethod def snake_case__ ( lowercase : ArgumentParser , lowercase : dataclasses.Field ) -> Optional[int]: """simple docstring""" __lowercase = F"--{field.name}" __lowercase = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , lowercase ): raise RuntimeError( """Unresolved type detected, which should have been done with the help of """ """`typing.get_type_hints` method by default""" ) __lowercase = kwargs.pop("""aliases""" , [] ) if isinstance(lowercase , lowercase ): __lowercase = [aliases] __lowercase = getattr(field.type , """__origin__""" , field.type ) if origin_type is Union or (hasattr(lowercase , """UnionType""" ) and isinstance(lowercase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(lowercase ) not in field.type.__args__ ): raise ValueError( """Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because""" """ the argument parser only supports one type per argument.""" F" Problem encountered in field '{field.name}'." ) if type(lowercase ) not in field.type.__args__: # filter `str` in Union __lowercase = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] __lowercase = getattr(field.type , """__origin__""" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) __lowercase = ( field.type.__args__[0] if isinstance(lowercase , field.type.__args__[1] ) else field.type.__args__[1] ) __lowercase = getattr(field.type , """__origin__""" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) __lowercase = {} if origin_type is Literal or (isinstance(field.type , lowercase ) and issubclass(field.type , lowercase )): if origin_type is Literal: __lowercase = field.type.__args__ else: __lowercase = [x.value for x in field.type] __lowercase = make_choice_type_function(kwargs["""choices"""] ) if field.default is not dataclasses.MISSING: __lowercase = field.default else: __lowercase = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument __lowercase = copy(lowercase ) # Hack because type=bool in argparse does not behave as we want. __lowercase = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. __lowercase = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way __lowercase = default # This tells argparse we accept 0 or 1 value after --field_name __lowercase = """?""" # This is the value that will get picked if we do --field_name (without value) __lowercase = True elif isclass(lowercase ) and issubclass(lowercase , lowercase ): __lowercase = field.type.__args__[0] __lowercase = """+""" if field.default_factory is not dataclasses.MISSING: __lowercase = field.default_factory() elif field.default is dataclasses.MISSING: __lowercase = True else: __lowercase = field.type if field.default is not dataclasses.MISSING: __lowercase = field.default elif field.default_factory is not dataclasses.MISSING: __lowercase = field.default_factory() else: __lowercase = True parser.add_argument(lowercase , *lowercase , **lowercase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): __lowercase = False parser.add_argument(F"--no_{field.name}" , action="""store_false""" , dest=field.name , **lowercase ) def snake_case__ ( self : str , lowercase : DataClassType ) -> Dict: """simple docstring""" if hasattr(lowercase , """_argument_group_name""" ): __lowercase = self.add_argument_group(dtype._argument_group_name ) else: __lowercase = self try: __lowercase = get_type_hints(lowercase ) except NameError: raise RuntimeError( F"Type resolution failed for {dtype}. Try declaring the class in global scope or " """removing line of `from __future__ import annotations` which opts in Postponed """ """Evaluation of Annotations (PEP 563)""" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(lowercase ): __lowercase = """.""".join(map(lowercase , sys.version_info[:3] ) ) raise RuntimeError( F"Type resolution failed for {dtype} on Python {python_version}. Try removing " """line of `from __future__ import annotations` which opts in union types as """ """`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To """ """support Python versions that lower than 3.10, you need to use """ """`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of """ """`X | None`.""" ) from ex raise for field in dataclasses.fields(lowercase ): if not field.init: continue __lowercase = type_hints[field.name] self._parse_dataclass_field(lowercase , lowercase ) def snake_case__ ( self : Dict , lowercase : Any=None , lowercase : int=False , lowercase : int=True , lowercase : Any=None , lowercase : str=None , ) -> Tuple[DataClass, ...]: """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): __lowercase = [] if args_filename: args_files.append(Path(lowercase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(""".args""" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values __lowercase = ArgumentParser() args_file_parser.add_argument(lowercase , type=lowercase , action="""append""" ) # Use only remaining args for further parsing (remove the args_file_flag) __lowercase , __lowercase = args_file_parser.parse_known_args(args=lowercase ) __lowercase = vars(lowercase ).get(args_file_flag.lstrip("""-""" ) , lowercase ) if cmd_args_file_paths: args_files.extend([Path(lowercase ) for p in cmd_args_file_paths] ) __lowercase = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last __lowercase = file_args + args if args is not None else file_args + sys.argv[1:] __lowercase , __lowercase = self.parse_known_args(args=lowercase ) __lowercase = [] for dtype in self.dataclass_types: __lowercase = {f.name for f in dataclasses.fields(lowercase ) if f.init} __lowercase = {k: v for k, v in vars(lowercase ).items() if k in keys} for k in keys: delattr(lowercase , lowercase ) __lowercase = dtype(**lowercase ) outputs.append(lowercase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(lowercase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F"Some specified arguments are not used by the HfArgumentParser: {remaining_args}" ) return (*outputs,) def snake_case__ ( self : Any , lowercase : Dict[str, Any] , lowercase : bool = False ) -> Tuple[DataClass, ...]: """simple docstring""" __lowercase = set(args.keys() ) __lowercase = [] for dtype in self.dataclass_types: __lowercase = {f.name for f in dataclasses.fields(lowercase ) if f.init} __lowercase = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) __lowercase = dtype(**lowercase ) outputs.append(lowercase ) if not allow_extra_keys and unused_keys: raise ValueError(F"Some keys are not used by the HfArgumentParser: {sorted(lowercase )}" ) return tuple(lowercase ) def snake_case__ ( self : List[Any] , lowercase : str , lowercase : bool = False ) -> Tuple[DataClass, ...]: """simple docstring""" with open(Path(lowercase ) , encoding="""utf-8""" ) as open_json_file: __lowercase = json.loads(open_json_file.read() ) __lowercase = self.parse_dict(lowercase , allow_extra_keys=lowercase ) return tuple(lowercase ) def snake_case__ ( self : int , lowercase : str , lowercase : bool = False ) -> Tuple[DataClass, ...]: """simple docstring""" __lowercase = self.parse_dict(yaml.safe_load(Path(lowercase ).read_text() ) , allow_extra_keys=lowercase ) return tuple(lowercase )
634
import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizerOutput from transformers.testing_utils import require_phonemizer from ...test_tokenization_common import TokenizerTesterMixin @require_phonemizer class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = WavaVecaPhonemeCTCTokenizer lowercase__ : Optional[int] = False def snake_case__ ( self : str ) -> int: """simple docstring""" super().setUp() __lowercase = ( """<s> <pad> </s> <unk> n s t ə l a i k d m ɛ ɾ e ɪ p o ɐ z ð f j v b ɹ ʁ ʊ iː r w ʌ u ɡ æ aɪ ʃ h ɔ ɑː """ """ŋ ɚ eɪ β uː y ɑ̃ oʊ ᵻ eː θ aʊ ts oː ɔ̃ ɣ ɜ ɑ dʒ əl x ɜː ç ʒ tʃ ɔː ɑːɹ ɛ̃ ʎ ɔːɹ ʋ aː ɕ œ ø oːɹ ɲ yː """ """ʔ iə i5 s. tɕ ?? nʲ ɛː œ̃ ɭ ɔø ʑ tʲ ɨ ɛɹ ts. rʲ ɪɹ ɭʲ i.5 ɔɪ q sʲ u5 ʊɹ iɜ a5 iɛ5 øː ʕ ja əɜ th ɑ5 """ """oɪ dʲ ə5 tɕh ts.h mʲ ɯ dʑ vʲ e̞ tʃʲ ei5 o5 onɡ5 ɑu5 iɑ5 ai5 aɪɚ kh ə1 ʐ i2 ʉ ħ t[ aɪə ʲ ju ə2 u2 oɜ """ """pː iɛɜ ou5 y5 uɜ tː uo5 d[ uoɜ tsh ɑɜ ɵ i̪5 uei5 ɟ aɜ ɑɨ i.ɜ eʊ o2 ɐ̃ ä pʲ kʲ n̩ ɒ ph ɑu2 uɨ əɪ ɫ ɬ """ """yɜ bʲ ɑ2 s̪ aiɜ χ ɐ̃ʊ̃ 1 ə4 yæɜ a2 ɨː t̪ iouɜ ũ onɡɜ aɨ iɛ2 ɔɨ ɑuɜ o̞ ei2 iou2 c kː y2 ɖ oe dˤ yɛɜ """ """əʊ S ɡʲ onɡ2 u\" eiɜ ʈ ɯᵝ iou5 dZ r̝̊ i.2 tS s^ ʝ yə5 iɑɜ uə5 pf ɨu iɑ2 ou2 ər2 fʲ ai2 r̝ uəɜ ɳ əɨ """ """ua5 uɪ ɽ bː yu5 uo2 yɛ5 l̩ ɻ ərɜ ʂ i̪2 ouɜ uaɜ a. a.ː yæ5 dː r̩ ee ɪu ər5 i̪ ɜ æi u: i.ː t^ o1 ɪ^ """ """ai ueiɜ æː ɛɪ eə i. ɴ ie ua2 ɑ1 o4 tʃː o: ɑ: u1 N i̪1 au yæ2 u. qː yəɜ y: kʰ tʃʰ iʊ sx õ uo tʰ """ """uai5 bʰ u.ː uə2 ʊə d^ s̪ː yiɜ dʰ r. oe: i1 ɟː yu2 nʲʲ i̪4 uei2 tsʲ ɸ ĩ ɑ4 t̪ː eɑ u4 e: tsː ʈʰ ɡʰ """ """ɯɯ dʒʲ ʂʲ X ɵː uaiɜ tɕʲ ã t^ː ẽː yɛ2 cː i.1 ɛʊ dˤdˤ dʒː i4 ɡː yi ɕʲ ɟʰ pʰ dʑʲ yuɜ ua1 ua4 æiː ɐɐ """ """ui iou1 ʊː a1 iou4 cʰ iɛ1 yə2 ɖʰ ẽ ʒʲ ää ər4 iːː ɪː iɑ1 ər1 œː øi ɪuː cʰcʰ əː1 iː1 ũ kʰː o̞o̞ xʲ """ """ou1 iɛ4 e̞e̞ y1 dzː dʲʲ dʰː ɯᵝɯᵝ lː uo1 i.4 i: yɛ5ʲ a4""" ).split(""" """ ) __lowercase = dict(zip(lowercase , range(len(lowercase ) ) ) ) __lowercase = {"""pad_token""": """<pad>""", """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>"""} __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowercase ) + """\n""" ) def snake_case__ ( self : List[Any] , lowercase : Optional[Any] , lowercase : List[str]=False , lowercase : List[str]=20 , lowercase : str=5 ) -> Tuple[str, list]: """simple docstring""" __lowercase = [(i, tokenizer.decode([i] , clean_up_tokenization_spaces=lowercase )) for i in range(len(lowercase ) )] __lowercase = list(filter(lambda lowercase : [t[0]] == tokenizer.encode(t[1] , do_phonemize=lowercase ) , lowercase ) ) if max_length is not None and len(lowercase ) > max_length: __lowercase = toks[:max_length] if min_length is not None and len(lowercase ) < min_length and len(lowercase ) > 0: while len(lowercase ) < min_length: __lowercase = toks + toks # toks_str = [t[1] for t in toks] __lowercase = [t[0] for t in toks] # Ensure consistency __lowercase = tokenizer.decode(lowercase , clean_up_tokenization_spaces=lowercase ) if " " not in output_txt and len(lowercase ) > 1: __lowercase = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=lowercase ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=lowercase ) ) if with_prefix_space: __lowercase = """ """ + output_txt __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) return output_txt, output_ids def snake_case__ ( self : Tuple , **lowercase : int ) -> Union[str, Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return WavaVecaPhonemeCTCTokenizer.from_pretrained(self.tmpdirname , **lowercase ) def snake_case__ ( self : Any ) -> List[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) # check adding a single token tokenizer.add_tokens("""xxx""" ) __lowercase = tokenizer("""m xxx ɪ""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 392, 17] ) # xxx should be last token tokenizer.add_tokens(["""aaa""", """bbb""", """ccc"""] ) __lowercase = tokenizer("""m aaa ɪ ccc""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 393, 17, 395] ) # aaa and ccc should be after xxx and 2 after aaa __lowercase = tokenizer("""maɪ c""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [3, 200] ) # mai should be <unk> (=3) def snake_case__ ( self : int ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) def snake_case__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98], [24, 22, 5, 24, 22, 5, 77], ] __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> Optional[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ | h aʊ | ɑːɹ | j uː |""" ) def snake_case__ ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : List[Any] ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, tokenizer.word_delimiter_token_id, 15, 8, tokenizer.word_delimiter_token_id, 98], [tokenizer.word_delimiter_token_id, 24, 22, tokenizer.word_delimiter_token_id, 5, 24, 22, 5, 77], ] # fmt: on # decode with word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) # decode with no word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] , filter_word_delimiter_token=lowercase ) __lowercase = tokenizer.batch_decode(lowercase , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ | ɾ l | ɭʲ""", """| j ð | s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : Optional[int] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(""" """.join([p.strip() for p in phonemes.split(""" |""" )] ).strip() , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token=lowercase ) __lowercase = """Hello how are you""" __lowercase = tokenizer(lowercase , phonemizer_lang="""en-us""" ).input_ids __lowercase = tokenizer(lowercase , phonemizer_lang="""fr-fr""" ).input_ids self.assertNotEqual(lowercase , lowercase ) __lowercase = tokenizer.decode(lowercase ) __lowercase = tokenizer.decode(lowercase ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) self.assertEqual(lowercase , """ɛ l o h aʊ a ʁ j u""" ) def snake_case__ ( self : int ) -> int: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how Are you""" __lowercase = """hello how are you""" __lowercase = tokenizer(lowercase ).input_ids __lowercase = tokenizer(lowercase ).input_ids self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : int ) -> Union[str, Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) tokenizer.add_tokens(["""!""", """?"""] ) tokenizer.add_special_tokens({"""cls_token""": """$$$"""} ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98, 392, 392, 393, 392, 392, 393, 394, 394], [24, 22, 5, 24, 22, 5, 77, tokenizer.pad_token_id, 394, 394], ] # fmt: on __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ!?!? $$$""", """j ð s j ð s oːɹ $$$"""] ) @staticmethod def snake_case__ ( lowercase : List[str] , lowercase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = [d[key] for d in offsets] return retrieved_list def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off # ksssɾɾ|ɾɾ<pad>ɾɾ|<pad>ɾlll|ɭʲ -> k s ɾ ɾ | ɾ l | ɭʲ" __lowercase = [11, 5, 5, 5, 15, 15, tokenizer.pad_token_id, 15, 15, tokenizer.word_delimiter_token_id, tokenizer.pad_token_id, 15, 8, 8, 8, tokenizer.word_delimiter_token_id, 98] # fmt: on __lowercase = tokenizer.decode(lowercase , output_char_offsets=lowercase , filter_word_delimiter_token=lowercase ) # check Wav2Vec2CTCTokenizerOutput keys for char self.assertEqual(len(outputs.keys() ) , 2 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""char_offsets""" in outputs ) self.assertTrue(isinstance(lowercase , lowercase ) ) # check that order of chars is correct and identical for both outputs self.assertEqual(""" """.join(self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) ) , outputs.text ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) , ["""k""", """s""", """ɾ""", """ɾ""", """|""", """ɾ""", """l""", """|""", """ɭʲ"""] ) # check that offsets are actually correct for char # 0-1 is 11, 1-4 is 5, 4-6 is first 15, 6-7 is <pad> (thus not shown), 7-9 is second 15, 9-10 is word_delimiter_token, # 10-11 is <pad> (thus not shown), 11-12 is third 15, 12-15 is 8, 15-16 is word_delimiter_token, 16-17 is 98 self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """start_offset""" ) , [0, 1, 4, 7, 9, 11, 12, 15, 16] ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """end_offset""" ) , [1, 4, 6, 9, 10, 12, 15, 16, 17] ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) def check_list_tuples_equal(lowercase : List[str] , lowercase : Dict ): self.assertTrue(isinstance(lowercase , lowercase ) ) self.assertTrue(isinstance(outputs_list[0] , lowercase ) ) # transform list to ModelOutput __lowercase = WavaVecaPhonemeCTCTokenizerOutput( {k: [d[k] for d in outputs_list] for k in outputs_list[0]} ) self.assertListEqual(outputs_batch["""text"""] , outputs_batch_a["""text"""] ) def recursive_check(lowercase : List[Any] , lowercase : Optional[int] ): if isinstance(lowercase , lowercase ): [recursive_check(lowercase , lowercase ) for la, la in zip(lowercase , lowercase )] self.assertEqual(lowercase , lowercase ) if "char_offsets" in outputs_batch: recursive_check(outputs_batch["""char_offsets"""] , outputs_batch_a["""char_offsets"""] ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 4, 8, 98, 32, 32, 32, 32, 4, 33, tokenizer.word_delimiter_token_id, 32, 32, 33, 34, 34], [24, 22, 5, tokenizer.word_delimiter_token_id, tokenizer.word_delimiter_token_id, 24, 22, 22, 22, 4, 5, 77, tokenizer.pad_token_id, 22, 22, 4, 34, 34, 34, 34], ] # fmt: on # We assume that `decode` works as expected. All we will check now is # the output type is correct and the output is identical to `decode` # char __lowercase = tokenizer.batch_decode(lowercase , output_char_offsets=lowercase ) __lowercase = [tokenizer.decode(lowercase , output_char_offsets=lowercase ) for ids in sample_ids] check_list_tuples_equal(lowercase , lowercase ) @unittest.skip("""Wav2Vec2PhonemeTokenizer always lower cases letters to correctly map to phonemes""" ) def snake_case__ ( self : List[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeTokenizer always puts spaces between phonemes""" ) def snake_case__ ( self : List[str] ) -> List[str]: """simple docstring""" pass @unittest.skip("""encodes to text to ids, but decodes ids to phonemes -> not possible to have internal consistency""" ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeModel has no max model length => no testing""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass def snake_case__ ( self : Any ) -> int: """simple docstring""" __lowercase = self.get_tokenizers(do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) __lowercase = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] __lowercase = tokenizer.add_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size + len(lowercase ) ) __lowercase = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) __lowercase = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} __lowercase = tokenizer.add_special_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size_a + len(lowercase ) ) __lowercase = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" pass def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizers(fast=lowercase , do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = ["""ð""", """ɪ""", """s""", """ɪ""", """z""", """ɐ""", """t""", """ɛ""", """k""", """s""", """t"""] __lowercase = tokenizer.convert_tokens_to_string(lowercase ) self.assertIsInstance(output["""text"""] , lowercase )
634
1
import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets UpperCamelCase__ = "\n@inproceedings{xu-etal-2016-optimizing,\n title = {Optimizing Statistical Machine Translation for Text Simplification},\n authors={Xu, Wei and Napoles, Courtney and Pavlick, Ellie and Chen, Quanze and Callison-Burch, Chris},\n journal = {Transactions of the Association for Computational Linguistics},\n volume = {4},\n year={2016},\n url = {https://www.aclweb.org/anthology/Q16-1029},\n pages = {401--415\n},\n@inproceedings{post-2018-call,\n title = \"A Call for Clarity in Reporting {BLEU} Scores\",\n author = \"Post, Matt\",\n booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\",\n month = oct,\n year = \"2018\",\n address = \"Belgium, Brussels\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W18-6319\",\n pages = \"186--191\",\n}\n" UpperCamelCase__ = "\\nWIKI_SPLIT is the combination of three metrics SARI, EXACT and SACREBLEU\nIt can be used to evaluate the quality of machine-generated texts.\n" UpperCamelCase__ = "\nCalculates sari score (between 0 and 100) given a list of source and predicted\nsentences, and a list of lists of reference sentences. It also computes the BLEU score as well as the exact match score.\nArgs:\n sources: list of source sentences where each sentence should be a string.\n predictions: list of predicted sentences where each sentence should be a string.\n references: list of lists of reference sentences where each sentence should be a string.\nReturns:\n sari: sari score\n sacrebleu: sacrebleu score\n exact: exact score\n\nExamples:\n >>> sources=[\"About 95 species are currently accepted .\"]\n >>> predictions=[\"About 95 you now get in .\"]\n >>> references=[[\"About 95 species are currently known .\"]]\n >>> wiki_split = datasets.load_metric(\"wiki_split\")\n >>> results = wiki_split.compute(sources=sources, predictions=predictions, references=references)\n >>> print(results)\n {'sari': 21.805555555555557, 'sacrebleu': 14.535768424205482, 'exact': 0.0}\n" def UpperCAmelCase__ ( lowercase__ ) -> Dict: def remove_articles(lowercase__ ): __lowercase = re.compile(r"""\b(a|an|the)\b""" , re.UNICODE ) return re.sub(lowercase__ , """ """ , lowercase__ ) def white_space_fix(lowercase__ ): return " ".join(text.split() ) def remove_punc(lowercase__ ): __lowercase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase__ ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase__ ) ) ) ) def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> List[Any]: return int(normalize_answer(lowercase__ ) == normalize_answer(lowercase__ ) ) def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> int: __lowercase = [any(compute_exact(lowercase__ , lowercase__ ) for ref in refs ) for pred, refs in zip(lowercase__ , lowercase__ )] return (sum(lowercase__ ) / len(lowercase__ )) * 100 def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ) -> Tuple: __lowercase = [rgram for rgrams in rgramslist for rgram in rgrams] __lowercase = Counter(lowercase__ ) __lowercase = Counter(lowercase__ ) __lowercase = Counter() for sgram, scount in sgramcounter.items(): __lowercase = scount * numref __lowercase = Counter(lowercase__ ) __lowercase = Counter() for cgram, ccount in cgramcounter.items(): __lowercase = ccount * numref # KEEP __lowercase = sgramcounter_rep & cgramcounter_rep __lowercase = keepgramcounter_rep & rgramcounter __lowercase = sgramcounter_rep & rgramcounter __lowercase = 0 __lowercase = 0 for keepgram in keepgramcountergood_rep: keeptmpscorea += keepgramcountergood_rep[keepgram] / keepgramcounter_rep[keepgram] # Fix an alleged bug [2] in the keep score computation. # keeptmpscore2 += keepgramcountergood_rep[keepgram] / keepgramcounterall_rep[keepgram] keeptmpscorea += keepgramcountergood_rep[keepgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. __lowercase = 1 __lowercase = 1 if len(lowercase__ ) > 0: __lowercase = keeptmpscorea / len(lowercase__ ) if len(lowercase__ ) > 0: # Fix an alleged bug [2] in the keep score computation. # keepscore_recall = keeptmpscore2 / len(keepgramcounterall_rep) __lowercase = keeptmpscorea / sum(keepgramcounterall_rep.values() ) __lowercase = 0 if keepscore_precision > 0 or keepscore_recall > 0: __lowercase = 2 * keepscore_precision * keepscore_recall / (keepscore_precision + keepscore_recall) # DELETION __lowercase = sgramcounter_rep - cgramcounter_rep __lowercase = delgramcounter_rep - rgramcounter __lowercase = sgramcounter_rep - rgramcounter __lowercase = 0 __lowercase = 0 for delgram in delgramcountergood_rep: deltmpscorea += delgramcountergood_rep[delgram] / delgramcounter_rep[delgram] deltmpscorea += delgramcountergood_rep[delgram] / delgramcounterall_rep[delgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. __lowercase = 1 if len(lowercase__ ) > 0: __lowercase = deltmpscorea / len(lowercase__ ) # ADDITION __lowercase = set(lowercase__ ) - set(lowercase__ ) __lowercase = set(lowercase__ ) & set(lowercase__ ) __lowercase = set(lowercase__ ) - set(lowercase__ ) __lowercase = 0 for addgram in addgramcountergood: addtmpscore += 1 # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. __lowercase = 1 __lowercase = 1 if len(lowercase__ ) > 0: __lowercase = addtmpscore / len(lowercase__ ) if len(lowercase__ ) > 0: __lowercase = addtmpscore / len(lowercase__ ) __lowercase = 0 if addscore_precision > 0 or addscore_recall > 0: __lowercase = 2 * addscore_precision * addscore_recall / (addscore_precision + addscore_recall) return (keepscore, delscore_precision, addscore) def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> Tuple: __lowercase = len(lowercase__ ) __lowercase = ssent.split(""" """ ) __lowercase = csent.split(""" """ ) __lowercase = [] __lowercase = [] __lowercase = [] __lowercase = [] __lowercase = [] __lowercase = [] __lowercase = [] __lowercase = [] __lowercase = [] __lowercase = [] for rsent in rsents: __lowercase = rsent.split(""" """ ) __lowercase = [] __lowercase = [] __lowercase = [] ragramslist.append(lowercase__ ) for i in range(0 , len(lowercase__ ) - 1 ): if i < len(lowercase__ ) - 1: __lowercase = ragrams[i] + """ """ + ragrams[i + 1] ragrams.append(lowercase__ ) if i < len(lowercase__ ) - 2: __lowercase = ragrams[i] + """ """ + ragrams[i + 1] + """ """ + ragrams[i + 2] ragrams.append(lowercase__ ) if i < len(lowercase__ ) - 3: __lowercase = ragrams[i] + """ """ + ragrams[i + 1] + """ """ + ragrams[i + 2] + """ """ + ragrams[i + 3] ragrams.append(lowercase__ ) ragramslist.append(lowercase__ ) ragramslist.append(lowercase__ ) ragramslist.append(lowercase__ ) for i in range(0 , len(lowercase__ ) - 1 ): if i < len(lowercase__ ) - 1: __lowercase = sagrams[i] + """ """ + sagrams[i + 1] sagrams.append(lowercase__ ) if i < len(lowercase__ ) - 2: __lowercase = sagrams[i] + """ """ + sagrams[i + 1] + """ """ + sagrams[i + 2] sagrams.append(lowercase__ ) if i < len(lowercase__ ) - 3: __lowercase = sagrams[i] + """ """ + sagrams[i + 1] + """ """ + sagrams[i + 2] + """ """ + sagrams[i + 3] sagrams.append(lowercase__ ) for i in range(0 , len(lowercase__ ) - 1 ): if i < len(lowercase__ ) - 1: __lowercase = cagrams[i] + """ """ + cagrams[i + 1] cagrams.append(lowercase__ ) if i < len(lowercase__ ) - 2: __lowercase = cagrams[i] + """ """ + cagrams[i + 1] + """ """ + cagrams[i + 2] cagrams.append(lowercase__ ) if i < len(lowercase__ ) - 3: __lowercase = cagrams[i] + """ """ + cagrams[i + 1] + """ """ + cagrams[i + 2] + """ """ + cagrams[i + 3] cagrams.append(lowercase__ ) ((__lowercase) , (__lowercase) , (__lowercase)) = SARIngram(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) ((__lowercase) , (__lowercase) , (__lowercase)) = SARIngram(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) ((__lowercase) , (__lowercase) , (__lowercase)) = SARIngram(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) ((__lowercase) , (__lowercase) , (__lowercase)) = SARIngram(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) __lowercase = sum([keepascore, keepascore, keepascore, keepascore] ) / 4 __lowercase = sum([delascore, delascore, delascore, delascore] ) / 4 __lowercase = sum([addascore, addascore, addascore, addascore] ) / 4 __lowercase = (avgkeepscore + avgdelscore + avgaddscore) / 3 return finalscore def UpperCAmelCase__ ( lowercase__ , lowercase__ = True , lowercase__ = "13a" , lowercase__ = True ) -> List[str]: # Normalization is requried for the ASSET dataset (one of the primary # datasets in sentence simplification) to allow using space # to split the sentence. Even though Wiki-Auto and TURK datasets, # do not require normalization, we do it for consistency. # Code adapted from the EASSE library [1] written by the authors of the ASSET dataset. # [1] https://github.com/feralvam/easse/blob/580bba7e1378fc8289c663f864e0487188fe8067/easse/utils/preprocessing.py#L7 if lowercase: __lowercase = sentence.lower() if tokenizer in ["13a", "intl"]: if version.parse(sacrebleu.__version__ ).major >= 2: __lowercase = sacrebleu.metrics.bleu._get_tokenizer(lowercase__ )()(lowercase__ ) else: __lowercase = sacrebleu.TOKENIZERS[tokenizer]()(lowercase__ ) elif tokenizer == "moses": __lowercase = sacremoses.MosesTokenizer().tokenize(lowercase__ , return_str=lowercase__ , escape=lowercase__ ) elif tokenizer == "penn": __lowercase = sacremoses.MosesTokenizer().penn_tokenize(lowercase__ , return_str=lowercase__ ) else: __lowercase = sentence if not return_str: __lowercase = normalized_sent.split() return normalized_sent def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> int: if not (len(lowercase__ ) == len(lowercase__ ) == len(lowercase__ )): raise ValueError("""Sources length must match predictions and references lengths.""" ) __lowercase = 0 for src, pred, refs in zip(lowercase__ , lowercase__ , lowercase__ ): sari_score += SARIsent(normalize(lowercase__ ) , normalize(lowercase__ ) , [normalize(lowercase__ ) for sent in refs] ) __lowercase = sari_score / len(lowercase__ ) return 100 * sari_score def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__="exp" , lowercase__=None , lowercase__=False , lowercase__=False , lowercase__=False , ) -> List[Any]: __lowercase = len(references[0] ) if any(len(lowercase__ ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) __lowercase = [[refs[i] for refs in references] for i in range(lowercase__ )] __lowercase = sacrebleu.corpus_bleu( lowercase__ , lowercase__ , smooth_method=lowercase__ , smooth_value=lowercase__ , force=lowercase__ , lowercase=lowercase__ , use_effective_order=lowercase__ , ) return output.score @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def snake_case__ ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=[ """https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py""", """https://github.com/cocoxu/simplification/blob/master/SARI.py""", """https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/sari_hook.py""", """https://github.com/mjpost/sacreBLEU""", ] , reference_urls=[ """https://www.aclweb.org/anthology/Q16-1029.pdf""", """https://github.com/mjpost/sacreBLEU""", """https://en.wikipedia.org/wiki/BLEU""", """https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213""", ] , ) def snake_case__ ( self : str , lowercase : Dict , lowercase : Optional[int] , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = {} result.update({"""sari""": compute_sari(sources=lowercase , predictions=lowercase , references=lowercase )} ) result.update({"""sacrebleu""": compute_sacrebleu(predictions=lowercase , references=lowercase )} ) result.update({"""exact""": compute_em(predictions=lowercase , references=lowercase )} ) return result
634
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) # TODO Update this UpperCamelCase__ = { "facebook/esm-1b": "https://huggingface.co/facebook/esm-1b/resolve/main/config.json", # See all ESM models at https://huggingface.co/models?filter=esm } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """esm""" def __init__( self : Any , lowercase : Optional[Any]=None , lowercase : Optional[int]=None , lowercase : List[Any]=None , lowercase : Optional[int]=768 , lowercase : str=12 , lowercase : Union[str, Any]=12 , lowercase : Dict=3_072 , lowercase : Optional[int]=0.1 , lowercase : str=0.1 , lowercase : Dict=1_026 , lowercase : Tuple=0.02 , lowercase : str=1E-1_2 , lowercase : Dict="absolute" , lowercase : Optional[Any]=True , lowercase : int=None , lowercase : int=False , lowercase : List[str]=False , lowercase : Tuple=None , lowercase : Tuple=None , **lowercase : Union[str, Any] , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase , mask_token_id=lowercase , **lowercase ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = emb_layer_norm_before __lowercase = token_dropout __lowercase = is_folding_model if is_folding_model: if esmfold_config is None: logger.info("""No esmfold_config supplied for folding model, using default values.""" ) __lowercase = EsmFoldConfig() elif isinstance(lowercase , lowercase ): __lowercase = EsmFoldConfig(**lowercase ) __lowercase = esmfold_config if vocab_list is None: logger.warning("""No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!""" ) __lowercase = get_default_vocab_list() else: __lowercase = vocab_list else: __lowercase = None __lowercase = None if self.esmfold_config is not None and getattr(self.esmfold_config , """use_esm_attn_map""" , lowercase ): raise ValueError("""The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!""" ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = super().to_dict() if isinstance(self.esmfold_config , lowercase ): __lowercase = self.esmfold_config.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : str = None lowercase__ : bool = True lowercase__ : bool = False lowercase__ : bool = False lowercase__ : bool = False lowercase__ : float = 0 lowercase__ : bool = True lowercase__ : bool = False lowercase__ : int = 128 lowercase__ : "TrunkConfig" = None def snake_case__ ( self : List[str] ) -> Any: """simple docstring""" if self.trunk is None: __lowercase = TrunkConfig() elif isinstance(self.trunk , lowercase ): __lowercase = TrunkConfig(**self.trunk ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.trunk.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 48 lowercase__ : int = 1_024 lowercase__ : int = 128 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : float = 0 lowercase__ : float = 0 lowercase__ : bool = False lowercase__ : int = 4 lowercase__ : Optional[int] = 128 lowercase__ : "StructureModuleConfig" = None def snake_case__ ( self : Tuple ) -> str: """simple docstring""" if self.structure_module is None: __lowercase = StructureModuleConfig() elif isinstance(self.structure_module , lowercase ): __lowercase = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(F"`max_recycles` should be positive, got {self.max_recycles}." ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( """`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got""" F" {self.sequence_state_dim} and {self.sequence_state_dim}." ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( """`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got""" F" {self.pairwise_state_dim} and {self.pairwise_state_dim}." ) __lowercase = self.sequence_state_dim // self.sequence_head_width __lowercase = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( """`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got""" F" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}." ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( """`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got""" F" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}." ) if self.pairwise_state_dim % 2 != 0: raise ValueError(F"`pairwise_state_dim` should be even, got {self.pairwise_state_dim}." ) if self.dropout >= 0.4: raise ValueError(F"`dropout` should not be greater than 0.4, got {self.dropout}." ) def snake_case__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.structure_module.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 384 lowercase__ : int = 128 lowercase__ : int = 16 lowercase__ : int = 128 lowercase__ : int = 12 lowercase__ : int = 4 lowercase__ : int = 8 lowercase__ : float = 0.1 lowercase__ : int = 8 lowercase__ : int = 1 lowercase__ : int = 2 lowercase__ : int = 7 lowercase__ : int = 10 lowercase__ : float = 1E-8 lowercase__ : float = 1E5 def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" return asdict(self ) def UpperCAmelCase__ ( ) -> List[Any]: return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
634
1
import math def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> float: if initial_intensity < 0: raise ValueError("""The value of intensity cannot be negative""" ) # handling of negative values of initial intensity if angle < 0 or angle > 360: raise ValueError("""In Malus Law, the angle is in the range 0-360 degrees""" ) # handling of values out of allowed range return initial_intensity * (math.cos(math.radians(lowercase__ ) ) ** 2) if __name__ == "__main__": import doctest doctest.testmod(name="malus_law")
634
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Tuple = LxmertTokenizer lowercase__ : List[str] = LxmertTokenizerFast lowercase__ : Optional[Any] = True lowercase__ : List[Any] = True def snake_case__ ( self : Tuple ) -> Tuple: """simple docstring""" super().setUp() __lowercase = [ """[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def snake_case__ ( self : Optional[int] , lowercase : int ) -> List[Any]: """simple docstring""" __lowercase = """UNwant\u00E9d,running""" __lowercase = """unwanted, running""" return input_text, output_text def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class(self.vocab_file ) __lowercase = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(lowercase , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , [7, 4, 5, 10, 8, 9] ) def snake_case__ ( self : Union[str, Any] ) -> int: """simple docstring""" if not self.test_rust_tokenizer: return __lowercase = self.get_tokenizer() __lowercase = self.get_rust_tokenizer() __lowercase = """I was born in 92000, and this is falsé.""" __lowercase = tokenizer.tokenize(lowercase ) __lowercase = rust_tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) __lowercase = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = self.get_rust_tokenizer() __lowercase = tokenizer.encode(lowercase ) __lowercase = rust_tokenizer.encode(lowercase ) self.assertListEqual(lowercase , lowercase )
634
1
from __future__ import annotations def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> dict[str, float]: if (voltage, current, resistance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if resistance < 0: raise ValueError("""Resistance cannot be negative""" ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
634
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> bool: __lowercase = len(lowercase__ ) __lowercase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): __lowercase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): __lowercase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: __lowercase = subset[i - 1][j] if arr[i - 1] <= j: __lowercase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
634
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__ = { "configuration_funnel": ["FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP", "FunnelConfig"], "convert_funnel_original_tf_checkpoint_to_pytorch": [], "tokenization_funnel": ["FunnelTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ["FunnelTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST", "FunnelBaseModel", "FunnelForMaskedLM", "FunnelForMultipleChoice", "FunnelForPreTraining", "FunnelForQuestionAnswering", "FunnelForSequenceClassification", "FunnelForTokenClassification", "FunnelModel", "FunnelPreTrainedModel", "load_tf_weights_in_funnel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST", "TFFunnelBaseModel", "TFFunnelForMaskedLM", "TFFunnelForMultipleChoice", "TFFunnelForPreTraining", "TFFunnelForQuestionAnswering", "TFFunnelForSequenceClassification", "TFFunnelForTokenClassification", "TFFunnelModel", "TFFunnelPreTrainedModel", ] if TYPE_CHECKING: from .configuration_funnel import FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP, FunnelConfig from .tokenization_funnel import FunnelTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_funnel_fast import FunnelTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_funnel import ( FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, FunnelBaseModel, FunnelForMaskedLM, FunnelForMultipleChoice, FunnelForPreTraining, FunnelForQuestionAnswering, FunnelForSequenceClassification, FunnelForTokenClassification, FunnelModel, FunnelPreTrainedModel, load_tf_weights_in_funnel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_funnel import ( TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, TFFunnelPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "hustvl/yolos-small": "https://huggingface.co/hustvl/yolos-small/resolve/main/config.json", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """yolos""" def __init__( self : Optional[int] , lowercase : Any=768 , lowercase : Tuple=12 , lowercase : Tuple=12 , lowercase : str=3_072 , lowercase : Optional[Any]="gelu" , lowercase : Union[str, Any]=0.0 , lowercase : Dict=0.0 , lowercase : Optional[int]=0.02 , lowercase : Optional[Any]=1E-1_2 , lowercase : Tuple=[512, 864] , lowercase : Optional[int]=16 , lowercase : Dict=3 , lowercase : Optional[Any]=True , lowercase : Optional[int]=100 , lowercase : Optional[int]=True , lowercase : Any=False , lowercase : Any=1 , lowercase : Any=5 , lowercase : List[str]=2 , lowercase : Union[str, Any]=5 , lowercase : str=2 , lowercase : Tuple=0.1 , **lowercase : str , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase ) __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = qkv_bias __lowercase = num_detection_tokens __lowercase = use_mid_position_embeddings __lowercase = auxiliary_loss # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = eos_coefficient class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Dict = version.parse("""1.11""" ) @property def snake_case__ ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def snake_case__ ( self : int ) -> float: """simple docstring""" return 1E-4 @property def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" return 12
634
1
from timeit import timeit def UpperCAmelCase__ ( lowercase__ ) -> int: if number < 0: raise ValueError("""the value of input must not be negative""" ) __lowercase = 0 while number: number &= number - 1 result += 1 return result def UpperCAmelCase__ ( lowercase__ ) -> int: if number < 0: raise ValueError("""the value of input must not be negative""" ) __lowercase = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def UpperCAmelCase__ ( ) -> None: def do_benchmark(lowercase__ ) -> None: __lowercase = """import __main__ as z""" print(F"Benchmark when {number = }:" ) print(F"{get_set_bits_count_using_modulo_operator(lowercase__ ) = }" ) __lowercase = timeit("""z.get_set_bits_count_using_modulo_operator(25)""" , setup=lowercase__ ) print(F"timeit() runs in {timing} seconds" ) print(F"{get_set_bits_count_using_brian_kernighans_algorithm(lowercase__ ) = }" ) __lowercase = timeit( """z.get_set_bits_count_using_brian_kernighans_algorithm(25)""" , setup=lowercase__ , ) print(F"timeit() runs in {timing} seconds" ) for number in (25, 37, 58, 0): do_benchmark(lowercase__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
634
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = IFImgaImgSuperResolutionPipeline lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) lowercase__ : Tuple = PipelineTesterMixin.required_optional_params - {"""latents"""} def snake_case__ ( self : Tuple ) -> Any: """simple docstring""" return self._get_superresolution_dummy_components() def snake_case__ ( self : List[str] , lowercase : Optional[int] , lowercase : Optional[Any]=0 ) -> Union[str, Any]: """simple docstring""" if str(lowercase ).startswith("""mps""" ): __lowercase = torch.manual_seed(lowercase ) else: __lowercase = torch.Generator(device=lowercase ).manual_seed(lowercase ) __lowercase = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """original_image""": original_image, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def snake_case__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1E-1 ) def snake_case__ ( self : Dict ) -> int: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" self._test_save_load_local() def snake_case__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
634
1
import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def UpperCAmelCase__ ( lowercase__ ) -> str: __lowercase = FileLock(str(tmpdir / """foo.lock""" ) ) __lowercase = FileLock(str(tmpdir / """foo.lock""" ) ) __lowercase = 0.01 with locka.acquire(): with pytest.raises(lowercase__ ): __lowercase = time.time() locka.acquire(lowercase__ ) assert time.time() - _start > timeout def UpperCAmelCase__ ( lowercase__ ) -> List[str]: __lowercase = """a""" * 1_000 + """.lock""" __lowercase = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith(""".lock""" ) assert not locka._lock_file.endswith(lowercase__ ) assert len(os.path.basename(locka._lock_file ) ) <= 255 __lowercase = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(lowercase__ ): locka.acquire(0 )
634
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ = { "configuration_luke": ["LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP", "LukeConfig"], "tokenization_luke": ["LukeTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "LUKE_PRETRAINED_MODEL_ARCHIVE_LIST", "LukeForEntityClassification", "LukeForEntityPairClassification", "LukeForEntitySpanClassification", "LukeForMultipleChoice", "LukeForQuestionAnswering", "LukeForSequenceClassification", "LukeForTokenClassification", "LukeForMaskedLM", "LukeModel", "LukePreTrainedModel", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
1
def UpperCAmelCase__ ( lowercase__ = 1_000 ) -> int: return sum(2 * a * ((a - 1) // 2) for a in range(3 , n + 1 ) ) if __name__ == "__main__": print(solution())
634
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> int: if len(lowercase__ ) != len(lowercase__ ): raise ValueError("""String lengths must match!""" ) __lowercase = 0 for chara, chara in zip(lowercase__ , lowercase__ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
634
1
class _lowerCAmelCase : """simple docstring""" def __init__( self : Tuple ) -> None: """simple docstring""" __lowercase = {} # Mapping from char to TrieNode __lowercase = False def snake_case__ ( self : Tuple , lowercase : list[str] ) -> None: """simple docstring""" for word in words: self.insert(lowercase ) def snake_case__ ( self : List[str] , lowercase : str ) -> None: """simple docstring""" __lowercase = self for char in word: if char not in curr.nodes: __lowercase = TrieNode() __lowercase = curr.nodes[char] __lowercase = True def snake_case__ ( self : Optional[int] , lowercase : str ) -> bool: """simple docstring""" __lowercase = self for char in word: if char not in curr.nodes: return False __lowercase = curr.nodes[char] return curr.is_leaf def snake_case__ ( self : Any , lowercase : str ) -> None: """simple docstring""" def _delete(lowercase : TrieNode , lowercase : str , lowercase : int ) -> bool: if index == len(lowercase ): # If word does not exist if not curr.is_leaf: return False __lowercase = False return len(curr.nodes ) == 0 __lowercase = word[index] __lowercase = curr.nodes.get(lowercase ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted __lowercase = _delete(lowercase , lowercase , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , lowercase , 0 ) def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> None: if node.is_leaf: print(lowercase__ , end=""" """ ) for key, value in node.nodes.items(): print_words(lowercase__ , word + key ) def UpperCAmelCase__ ( ) -> bool: __lowercase = """banana bananas bandana band apple all beast""".split() __lowercase = TrieNode() root.insert_many(lowercase__ ) # print_words(root, "") assert all(root.find(lowercase__ ) for word in words ) assert root.find("""banana""" ) assert not root.find("""bandanas""" ) assert not root.find("""apps""" ) assert root.find("""apple""" ) assert root.find("""all""" ) root.delete("""all""" ) assert not root.find("""all""" ) root.delete("""banana""" ) assert not root.find("""banana""" ) assert root.find("""bananas""" ) return True def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> None: print(str(lowercase__ ) , """works!""" if passes else """doesn't work :(""" ) def UpperCAmelCase__ ( ) -> None: assert test_trie() def UpperCAmelCase__ ( ) -> None: print_results("""Testing trie functionality""" , test_trie() ) if __name__ == "__main__": main()
634
from __future__ import annotations from collections.abc import Callable UpperCamelCase__ = list[list[float | int]] def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Matrix: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(size + 1 )] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for row in range(lowercase__ ): for col in range(lowercase__ ): __lowercase = matrix[row][col] __lowercase = vector[row][0] __lowercase = 0 __lowercase = 0 while row < size and col < size: # pivoting __lowercase = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase__ , lowercase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: __lowercase , __lowercase = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase__ ): __lowercase = augmented[rowa][col] / augmented[row][col] __lowercase = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase__ ): for row in range(lowercase__ ): __lowercase = augmented[row][col] / augmented[col][col] for cola in range(lowercase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase__ ) ] def UpperCAmelCase__ ( lowercase__ ) -> Callable[[int], int]: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(lowercase__ )] for _ in range(lowercase__ )] __lowercase = [[0] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for x_val, y_val in enumerate(lowercase__ ): for col in range(lowercase__ ): __lowercase = (x_val + 1) ** (size - col - 1) __lowercase = y_val __lowercase = solve(lowercase__ , lowercase__ ) def interpolated_func(lowercase__ ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase__ ) ) return interpolated_func def UpperCAmelCase__ ( lowercase__ ) -> int: return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCAmelCase__ ( lowercase__ = question_function , lowercase__ = 10 ) -> int: __lowercase = [func(lowercase__ ) for x_val in range(1 , order + 1 )] __lowercase = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] __lowercase = 0 __lowercase = 42 __lowercase = 42 for poly in polynomials: __lowercase = 1 while func(lowercase__ ) == poly(lowercase__ ): x_val += 1 ret += poly(lowercase__ ) return ret if __name__ == "__main__": print(F"""{solution() = }""")
634
1
from math import pi, sqrt def UpperCAmelCase__ ( lowercase__ ) -> float: if num <= 0: raise ValueError("""math domain error""" ) if num > 171.5: raise OverflowError("""math range error""" ) elif num - int(lowercase__ ) not in (0, 0.5): raise NotImplementedError("""num must be an integer or a half-integer""" ) elif num == 0.5: return sqrt(lowercase__ ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def UpperCAmelCase__ ( ) -> None: assert gamma(0.5 ) == sqrt(lowercase__ ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() UpperCamelCase__ = 1.0 while num: UpperCamelCase__ = float(input("Gamma of: ")) print(F"""gamma({num}) = {gamma(num)}""") print("\nEnter 0 to exit...")
634
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ = { "configuration_ctrl": ["CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP", "CTRLConfig"], "tokenization_ctrl": ["CTRLTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "CTRLForSequenceClassification", "CTRLLMHeadModel", "CTRLModel", "CTRLPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCTRLForSequenceClassification", "TFCTRLLMHeadModel", "TFCTRLModel", "TFCTRLPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
1
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def UpperCAmelCase__ ( *lowercase__ , lowercase__ = None , lowercase__=True , lowercase__=2 ) -> List[Any]: from .. import __version__ __lowercase = take_from __lowercase = () if not isinstance(args[0] , lowercase__ ): __lowercase = (args,) for attribute, version_name, message in args: if version.parse(version.parse(lowercase__ ).base_version ) >= version.parse(lowercase__ ): raise ValueError( F"The deprecation tuple {(attribute, version_name, message)} should be removed since diffusers'" F" version {__version__} is >= {version_name}" ) __lowercase = None if isinstance(lowercase__ , lowercase__ ) and attribute in deprecated_kwargs: values += (deprecated_kwargs.pop(lowercase__ ),) __lowercase = F"The `{attribute}` argument is deprecated and will be removed in version {version_name}." elif hasattr(lowercase__ , lowercase__ ): values += (getattr(lowercase__ , lowercase__ ),) __lowercase = F"The `{attribute}` attribute is deprecated and will be removed in version {version_name}." elif deprecated_kwargs is None: __lowercase = F"`{attribute}` is deprecated and will be removed in version {version_name}." if warning is not None: __lowercase = warning + """ """ if standard_warn else """""" warnings.warn(warning + message , lowercase__ , stacklevel=lowercase__ ) if isinstance(lowercase__ , lowercase__ ) and len(lowercase__ ) > 0: __lowercase = inspect.getouterframes(inspect.currentframe() )[1] __lowercase = call_frame.filename __lowercase = call_frame.lineno __lowercase = call_frame.function __lowercase , __lowercase = next(iter(deprecated_kwargs.items() ) ) raise TypeError(F"{function} in {filename} line {line_number-1} got an unexpected keyword argument `{key}`" ) if len(lowercase__ ) == 0: return elif len(lowercase__ ) == 1: return values[0] return values
634
import unittest import numpy as np def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = None , ) -> np.ndarray: __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) if shape_a[0] != shape_b[0]: __lowercase = ( """Expected the same number of rows for A and B. """ F"Instead found A of size {shape_a} and B of size {shape_b}" ) raise ValueError(lowercase__ ) if shape_b[1] != shape_c[1]: __lowercase = ( """Expected the same number of columns for B and C. """ F"Instead found B of size {shape_b} and C of size {shape_c}" ) raise ValueError(lowercase__ ) __lowercase = pseudo_inv if a_inv is None: try: __lowercase = np.linalg.inv(lowercase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Dict ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) __lowercase = schur_complement(lowercase , lowercase , lowercase ) __lowercase = np.block([[a, b], [b.T, c]] ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) self.assertAlmostEqual(lowercase , det_a * det_s ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
634
1
UpperCamelCase__ = "Alexander Joslin" import operator as op from .stack import Stack def UpperCAmelCase__ ( lowercase__ ) -> int: __lowercase = {"""*""": op.mul, """/""": op.truediv, """+""": op.add, """-""": op.sub} __lowercase = Stack() __lowercase = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(lowercase__ ) ) elif i in operators: # RULE 2 operator_stack.push(lowercase__ ) elif i == ")": # RULE 4 __lowercase = operator_stack.peek() operator_stack.pop() __lowercase = operand_stack.peek() operand_stack.pop() __lowercase = operand_stack.peek() operand_stack.pop() __lowercase = operators[opr](lowercase__ , lowercase__ ) operand_stack.push(lowercase__ ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": UpperCamelCase__ = "(5 + ((4 * 2) * (2 + 3)))" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
634
import random def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ = False ) -> dict: __lowercase = {i: [] for i in range(lowercase__ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(lowercase__ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(lowercase__ ): for j in range(i + 1 , lowercase__ ): if random.random() < probability: graph[i].append(lowercase__ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(lowercase__ ) return graph def UpperCAmelCase__ ( lowercase__ ) -> dict: return { i: [j for j in range(lowercase__ ) if i != j] for i in range(lowercase__ ) } if __name__ == "__main__": import doctest doctest.testmod()
634
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING import torch from ..models.auto import AutoModelForVisualQuestionAnswering, AutoProcessor from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Tuple = """dandelin/vilt-b32-finetuned-vqa""" lowercase__ : Tuple = ( """This is a tool that answers a question about an image. It takes an input named `image` which should be the """ """image containing the information, as well as a `question` which should be the question in English. It """ """returns a text that is the answer to the question.""" ) lowercase__ : int = """image_qa""" lowercase__ : Tuple = AutoProcessor lowercase__ : List[Any] = AutoModelForVisualQuestionAnswering lowercase__ : Optional[Any] = ["""image""", """text"""] lowercase__ : Optional[Any] = ["""text"""] def __init__( self : int , *lowercase : Optional[Any] , **lowercase : List[str] ) -> List[str]: """simple docstring""" requires_backends(self , ["""vision"""] ) super().__init__(*lowercase , **lowercase ) def snake_case__ ( self : int , lowercase : "Image" , lowercase : str ) -> Optional[Any]: """simple docstring""" return self.pre_processor(lowercase , lowercase , return_tensors="""pt""" ) def snake_case__ ( self : Optional[int] , lowercase : Union[str, Any] ) -> int: """simple docstring""" with torch.no_grad(): return self.model(**lowercase ).logits def snake_case__ ( self : int , lowercase : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = outputs.argmax(-1 ).item() return self.model.config.idalabel[idx]
634
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor UpperCamelCase__ = random.Random() def UpperCAmelCase__ ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ) -> str: if rng is None: __lowercase = global_rng __lowercase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : List[Any] , lowercase : Tuple , lowercase : Union[str, Any]=7 , lowercase : List[Any]=400 , lowercase : Any=2_000 , lowercase : Optional[int]=24 , lowercase : Any=24 , lowercase : List[str]=0.0 , lowercase : Dict=16_000 , lowercase : Union[str, Any]=True , lowercase : Dict=True , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = min_seq_length __lowercase = max_seq_length __lowercase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowercase = feature_size __lowercase = num_mel_bins __lowercase = padding_value __lowercase = sampling_rate __lowercase = return_attention_mask __lowercase = do_normalize def snake_case__ ( self : Optional[int] ) -> Any: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self : List[str] , lowercase : Tuple=False , lowercase : int=False ) -> Optional[Any]: """simple docstring""" def _flatten(lowercase : Optional[Any] ): return list(itertools.chain(*lowercase ) ) if equal_length: __lowercase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowercase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowercase = [np.asarray(lowercase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = SpeechaTextFeatureExtractor if is_speech_available() else None def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = SpeechaTextFeatureExtractionTester(self ) def snake_case__ ( self : Tuple , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" self.assertTrue(np.all(np.mean(lowercase , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowercase , axis=0 ) - 1 ) < 1E-3 ) ) def snake_case__ ( self : List[Any] ) -> str: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = [np.asarray(lowercase ) for speech_input in speech_inputs] # Test feature size __lowercase = feature_extractor(lowercase , padding=lowercase , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input __lowercase = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features __lowercase = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) # Test batched __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(lowercase , lowercase ): self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __lowercase = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowercase = np.asarray(lowercase ) __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(lowercase , lowercase ): self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) def snake_case__ ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(lowercase , lowercase ): __lowercase = feature_extractor( lowercase , padding=lowercase , max_length=lowercase , return_attention_mask=lowercase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(lowercase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def snake_case__ ( self : Any ) -> Dict: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(lowercase , lowercase ): __lowercase = feature_extractor( lowercase , max_length=lowercase , padding=lowercase , return_tensors="""np""" , return_attention_mask=lowercase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(lowercase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def snake_case__ ( self : str ) -> List[Any]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""max_length""" , max_length=4 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""longest""" , max_length=4 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""longest""" , max_length=16 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" import torch __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = np.random.rand(100 , 32 ).astype(np.floataa ) __lowercase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self : Optional[int] , lowercase : Union[str, Any] ) -> int: """simple docstring""" from datasets import load_dataset __lowercase = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __lowercase = ds.sort("""id""" ).select(range(lowercase ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on __lowercase = self._load_datasamples(1 ) __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = feature_extractor(lowercase , return_tensors="""pt""" ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , lowercase , atol=1E-4 ) )
634
1
from collections import Counter import numpy as np from sklearn import datasets from sklearn.model_selection import train_test_split UpperCamelCase__ = datasets.load_iris() UpperCamelCase__ = np.array(data["data"]) UpperCamelCase__ = np.array(data["target"]) UpperCamelCase__ = data["target_names"] UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = train_test_split(X, y) def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> str: return np.linalg.norm(np.array(lowercase__ ) - np.array(lowercase__ ) ) def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__=5 ) -> List[Any]: __lowercase = zip(lowercase__ , lowercase__ ) # List of distances of all points from the point to be classified __lowercase = [] for data_point in data: __lowercase = euclidean_distance(data_point[0] , lowercase__ ) distances.append((distance, data_point[1]) ) # Choosing 'k' points with the least distances. __lowercase = [i[1] for i in sorted(lowercase__ )[:k]] # Most commonly occurring class among them # is the class into which the point is classified __lowercase = Counter(lowercase__ ).most_common(1 )[0][0] return classes[result] if __name__ == "__main__": print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
634
def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: __lowercase = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def UpperCAmelCase__ ( ) -> List[str]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
634
1
import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self : List[Any] , lowercase : Union[str, Any] , lowercase : Tuple=13 , lowercase : Union[str, Any]=30 , lowercase : List[str]=2 , lowercase : Optional[Any]=3 , lowercase : List[Any]=True , lowercase : Dict=True , lowercase : List[Any]=32 , lowercase : int=5 , lowercase : Tuple=4 , lowercase : List[str]=37 , lowercase : Any="gelu" , lowercase : Any=0.1 , lowercase : Dict=0.1 , lowercase : Optional[Any]=10 , lowercase : Tuple=0.02 , lowercase : str=None , lowercase : int=2 , ) -> Tuple: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = scope __lowercase = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __lowercase = (image_size // patch_size) ** 2 __lowercase = num_patches + 1 def snake_case__ ( self : str ) -> str: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def snake_case__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def snake_case__ ( self : Optional[int] , lowercase : Tuple , lowercase : Tuple , lowercase : Optional[int] ) -> int: """simple docstring""" __lowercase = ViTModel(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : Tuple , lowercase : str , lowercase : Optional[int] , lowercase : Dict ) -> Any: """simple docstring""" __lowercase = ViTForMaskedImageModeling(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __lowercase = 1 __lowercase = ViTForMaskedImageModeling(lowercase ) model.to(lowercase ) model.eval() __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def snake_case__ ( self : Optional[Any] , lowercase : str , lowercase : int , lowercase : List[str] ) -> Any: """simple docstring""" __lowercase = self.type_sequence_label_size __lowercase = ViTForImageClassification(lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowercase = 1 __lowercase = ViTForImageClassification(lowercase ) model.to(lowercase ) model.eval() __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def snake_case__ ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : str = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) lowercase__ : Any = ( {"""feature-extraction""": ViTModel, """image-classification""": ViTForImageClassification} if is_torch_available() else {} ) lowercase__ : Tuple = True lowercase__ : int = False lowercase__ : Any = False lowercase__ : List[Any] = False def snake_case__ ( self : List[str] ) -> int: """simple docstring""" __lowercase = ViTModelTester(self ) __lowercase = ConfigTester(self , config_class=lowercase , has_text_modality=lowercase , hidden_size=37 ) def snake_case__ ( self : Union[str, Any] ) -> int: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def snake_case__ ( self : Dict ) -> List[Any]: """simple docstring""" pass def snake_case__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def snake_case__ ( self : int ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase ) def snake_case__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def snake_case__ ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowercase ) def snake_case__ ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase ) @slow def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ViTModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def UpperCAmelCase__ ( ) -> str: __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case__ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def snake_case__ ( self : int ) -> List[str]: """simple docstring""" __lowercase = ViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ).to(lowercase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase , return_tensors="""pt""" ).to(lowercase ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase ) # verify the logits __lowercase = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowercase ) __lowercase = torch.tensor([-0.2744, 0.8215, -0.0836] ).to(lowercase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase , atol=1E-4 ) ) @slow def snake_case__ ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = ViTModel.from_pretrained("""facebook/dino-vits8""" ).to(lowercase ) __lowercase = ViTImageProcessor.from_pretrained("""facebook/dino-vits8""" , size=480 ) __lowercase = prepare_img() __lowercase = image_processor(images=lowercase , return_tensors="""pt""" ) __lowercase = inputs.pixel_values.to(lowercase ) # forward pass with torch.no_grad(): __lowercase = model(lowercase , interpolate_pos_encoding=lowercase ) # verify the logits __lowercase = torch.Size((1, 3_601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , lowercase ) __lowercase = torch.tensor( [[4.2340, 4.3906, -6.6692], [4.5463, 1.8928, -6.7257], [4.4429, 0.8496, -5.8585]] ).to(lowercase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowercase , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = ViTModel.from_pretrained("""facebook/dino-vits8""" , torch_dtype=torch.floataa , device_map="""auto""" ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase , return_tensors="""pt""" ) __lowercase = inputs.pixel_values.to(lowercase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): __lowercase = model(lowercase )
634
def UpperCAmelCase__ ( lowercase__ = 100 ) -> int: __lowercase = n * (n + 1) * (2 * n + 1) / 6 __lowercase = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F"""{solution() = }""")
634
1
import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def UpperCAmelCase__ ( ) -> List[Any]: with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(lowercase__ ): requests.request("""GET""" , """https://huggingface.co""" ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request("""GET""" , """https://huggingface.co""" , timeout=1.0 ) @pytest.mark.integration def UpperCAmelCase__ ( ) -> Optional[Any]: with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request("""GET""" , """https://huggingface.co""" ) def UpperCAmelCase__ ( ) -> str: with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(lowercase__ ): http_head("""https://huggingface.co""" )
634
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets UpperCamelCase__ = datasets.logging.get_logger(__name__) UpperCamelCase__ = "\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = \"Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric\",\n author = \"Moosavi, Nafise Sadat and\n Strube, Michael\",\n booktitle = \"Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)\",\n month = aug,\n year = \"2016\",\n address = \"Berlin, Germany\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/P16-1060\",\n doi = \"10.18653/v1/P16-1060\",\n pages = \"632--642\",\n}\n\n" UpperCamelCase__ = "\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a \"-\"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section \"*_conll File Format\"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n" UpperCamelCase__ = "\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting 'keep_singletons=False', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the 'NP_only' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting 'min_span', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n 'mentions': mentions\n 'muc': MUC metric [Vilain et al, 1995]\n 'bcub': B-cubed [Bagga and Baldwin, 1998]\n 'ceafe': CEAFe [Luo et al., 2005]\n 'lea': LEA [Moosavi and Strube, 2016]\n 'conll_score': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric('coval')\n >>> words = ['bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -',\n ... 'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)',\n ... 'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)',\n ... 'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -',\n ... 'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -',\n ... 'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {'mentions/recall': 1.0,[...] 'conll_score': 100.0}\n" def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__=False , lowercase__=False , lowercase__=True , lowercase__=False , lowercase__="dummy_doc" ) -> str: __lowercase = {doc: key_lines} __lowercase = {doc: sys_lines} __lowercase = {} __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase , __lowercase = reader.get_doc_mentions(lowercase__ , key_doc_lines[doc] , lowercase__ ) key_singletons_num += singletons_num if NP_only or min_span: __lowercase = reader.set_annotated_parse_trees(lowercase__ , key_doc_lines[doc] , lowercase__ , lowercase__ ) __lowercase , __lowercase = reader.get_doc_mentions(lowercase__ , sys_doc_lines[doc] , lowercase__ ) sys_singletons_num += singletons_num if NP_only or min_span: __lowercase = reader.set_annotated_parse_trees(lowercase__ , key_doc_lines[doc] , lowercase__ , lowercase__ ) if remove_nested: __lowercase , __lowercase = reader.remove_nested_coref_mentions(lowercase__ , lowercase__ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters __lowercase , __lowercase = reader.remove_nested_coref_mentions(lowercase__ , lowercase__ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters __lowercase = reader.get_mention_assignments(lowercase__ , lowercase__ ) __lowercase = reader.get_mention_assignments(lowercase__ , lowercase__ ) __lowercase = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F"annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}" ) logger.info( """Number of resulting singleton clusters in the key """ F"annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}" ) if not keep_singletons: logger.info( F"{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system " """files, respectively""" ) return doc_coref_infos def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) -> Optional[Any]: __lowercase = get_coref_infos(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) __lowercase = {} __lowercase = 0 __lowercase = 0 for name, metric in metrics: __lowercase , __lowercase , __lowercase = evaluator.evaluate_documents(lowercase__ , lowercase__ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F"{name}/recall": recall, F"{name}/precision": precision, F"{name}/f1": fa} ) logger.info( name.ljust(10 ) , F"Recall: {recall * 100:.2f}" , F" Precision: {precision * 100:.2f}" , F" F1: {fa * 100:.2f}" , ) if conll_subparts_num == 3: __lowercase = (conll / 3) * 100 logger.info(F"CoNLL score: {conll:.2f}" ) output_scores.update({"""conll_score""": conll} ) return output_scores def UpperCAmelCase__ ( lowercase__ ) -> List[Any]: __lowercase = False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: __lowercase = line.split()[5] if not parse_col == "-": __lowercase = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def snake_case__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def snake_case__ ( self : Tuple , lowercase : Dict , lowercase : Optional[int] , lowercase : Dict=True , lowercase : List[str]=False , lowercase : int=False , lowercase : Dict=False ) -> str: """simple docstring""" __lowercase = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: __lowercase = util.check_gold_parse_annotation(lowercase ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" __lowercase = evaluate( key_lines=lowercase , sys_lines=lowercase , metrics=lowercase , NP_only=lowercase , remove_nested=lowercase , keep_singletons=lowercase , min_span=lowercase , ) return score
634
1
import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) UpperCamelCase__ = logging.getLogger(__name__) UpperCamelCase__ = "Hello world! cécé herlolip" UpperCamelCase__ = namedtuple( "BertAbsConfig", [ "temp_dir", "large", "use_bert_emb", "finetune_bert", "encoder", "share_emb", "max_pos", "enc_layers", "enc_hidden_size", "enc_heads", "enc_ff_size", "enc_dropout", "dec_layers", "dec_hidden_size", "dec_heads", "dec_ff_size", "dec_dropout", ], ) def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Any: __lowercase = BertAbsConfig( temp_dir=""".""" , finetune_bert=lowercase__ , large=lowercase__ , share_emb=lowercase__ , use_bert_emb=lowercase__ , encoder="""bert""" , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2_048 , dec_dropout=0.2 , ) __lowercase = torch.load(lowercase__ , lambda lowercase__ , lowercase__ : storage ) __lowercase = AbsSummarizer(lowercase__ , torch.device("""cpu""" ) , lowercase__ ) original.eval() __lowercase = BertAbsSummarizer(lowercase__ , torch.device("""cpu""" ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info("""convert the model""" ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info("""Make sure that the models' outputs are identical""" ) __lowercase = BertTokenizer.from_pretrained("""bert-base-uncased""" ) # prepare the model inputs __lowercase = tokenizer.encode("""This is sample éàalj'-.""" ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(lowercase__ )) ) __lowercase = torch.tensor(lowercase__ ).unsqueeze(0 ) __lowercase = tokenizer.encode("""This is sample 3 éàalj'-.""" ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(lowercase__ )) ) __lowercase = torch.tensor(lowercase__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ )[0] __lowercase = original.generator(lowercase__ ) __lowercase = new_model( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ )[0] __lowercase = new_model.generator(lowercase__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print("""Maximum absolute difference beween weights: {:.2f}""".format(lowercase__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print("""Maximum absolute difference beween weights: {:.2f}""".format(lowercase__ ) ) __lowercase = torch.allclose(lowercase__ , lowercase__ , atol=1E-3 ) if are_identical: logging.info("""all weights are equal up to 1e-3""" ) else: raise ValueError("""the weights are different. The new model is likely different from the original one.""" ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info("""saving the model's state dictionary""" ) torch.save( new_model.state_dict() , """./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin""" ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument( "--bertabs_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model.", ) UpperCamelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
634
UpperCamelCase__ = { "joule": 1.0, "kilojoule": 10_00, "megajoule": 1_00_00_00, "gigajoule": 10_00_00_00_00, "wattsecond": 1.0, "watthour": 36_00, "kilowatthour": 3_60_00_00, "newtonmeter": 1.0, "calorie_nutr": 41_86.8, "kilocalorie_nutr": 4_18_68_00.00, "electronvolt": 1.6_0217_6634e-19, "britishthermalunit_it": 10_55.0_55_85, "footpound": 1.355_818, } def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: __lowercase = ( F"Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n" F"Valid values are: {', '.join(lowercase__ )}" ) raise ValueError(lowercase__ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
634
1
import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : int ) -> List[str]: """simple docstring""" super().__init__() __lowercase = nn.Linear(3 , 4 ) __lowercase = nn.BatchNormad(4 ) __lowercase = nn.Linear(4 , 5 ) def snake_case__ ( self : int , lowercase : Dict ) -> Optional[int]: """simple docstring""" return self.lineara(self.batchnorm(self.lineara(lowercase ) ) ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" def snake_case__ ( self : List[Any] , lowercase : List[Any] , *lowercase : Tuple , **lowercase : List[str] ) -> Optional[Any]: """simple docstring""" return (args[0] + 1,) + args[1:], kwargs class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" def snake_case__ ( self : Optional[Any] , lowercase : Dict , lowercase : Dict ) -> Dict: """simple docstring""" return output + 1 class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = ModelForTest() __lowercase = ModelHook() add_hook_to_module(lowercase , lowercase ) self.assertEqual(test_model._hf_hook , lowercase ) self.assertTrue(hasattr(lowercase , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(lowercase ) self.assertFalse(hasattr(lowercase , """_hf_hook""" ) ) self.assertFalse(hasattr(lowercase , """_old_forward""" ) ) def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = ModelForTest() __lowercase = ModelHook() add_hook_to_module(lowercase , lowercase ) add_hook_to_module(lowercase , lowercase , append=lowercase ) self.assertEqual(isinstance(test_model._hf_hook , lowercase ) , lowercase ) self.assertEqual(len(test_model._hf_hook.hooks ) , 2 ) self.assertTrue(hasattr(lowercase , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(lowercase ) self.assertFalse(hasattr(lowercase , """_hf_hook""" ) ) self.assertFalse(hasattr(lowercase , """_old_forward""" ) ) def snake_case__ ( self : Union[str, Any] ) -> str: """simple docstring""" __lowercase = ModelForTest() __lowercase = torch.randn(2 , 3 ) __lowercase = test_model(x + 1 ) __lowercase = test_model(x + 2 ) __lowercase = PreForwardHook() add_hook_to_module(lowercase , lowercase ) __lowercase = test_model(lowercase ) self.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain __lowercase = PreForwardHook() add_hook_to_module(lowercase , lowercase ) __lowercase = test_model(lowercase ) self.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks __lowercase = SequentialHook(PreForwardHook() , PreForwardHook() ) add_hook_to_module(lowercase , lowercase ) __lowercase = test_model(lowercase ) assert torch.allclose(lowercase , lowercase , atol=1E-5 ) def snake_case__ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = ModelForTest() __lowercase = torch.randn(2 , 3 ) __lowercase = test_model(lowercase ) __lowercase = PostForwardHook() add_hook_to_module(lowercase , lowercase ) __lowercase = test_model(lowercase ) self.assertTrue(torch.allclose(lowercase , output + 1 , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain __lowercase = PostForwardHook() add_hook_to_module(lowercase , lowercase ) __lowercase = test_model(lowercase ) self.assertTrue(torch.allclose(lowercase , output + 1 , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks __lowercase = SequentialHook(PostForwardHook() , PostForwardHook() ) add_hook_to_module(lowercase , lowercase ) __lowercase = test_model(lowercase ) assert torch.allclose(lowercase , output + 2 , atol=1E-5 ) def snake_case__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = ModelForTest() __lowercase = torch.randn(2 , 3 ) __lowercase = test_model(lowercase ) __lowercase = PostForwardHook() add_hook_to_module(lowercase , lowercase ) __lowercase = test_model(lowercase ) self.assertTrue(torch.allclose(lowercase , output + 1 ) ) self.assertTrue(outputa.requires_grad ) __lowercase = True __lowercase = test_model(lowercase ) self.assertFalse(outputa.requires_grad ) @require_multi_gpu def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1 ) ) self.assertEqual(model.lineara.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0 ) ) self.assertEqual(model.lineara.weight.device , torch.device(1 ) ) # We can still make a forward pass. The input does not need to be on any particular device __lowercase = torch.randn(2 , 3 ) __lowercase = model(lowercase ) self.assertEqual(output.device , torch.device(1 ) ) # We can add a general hook to put back output on same device as input. add_hook_to_module(lowercase , AlignDevicesHook(io_same_device=lowercase ) ) __lowercase = torch.randn(2 , 3 ).to(0 ) __lowercase = model(lowercase ) self.assertEqual(output.device , torch.device(0 ) ) def snake_case__ ( self : List[Any] ) -> List[str]: """simple docstring""" __lowercase = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices __lowercase = {"""execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True} add_hook_to_module(model.lineara , AlignDevicesHook(**lowercase ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**lowercase ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**lowercase ) ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device __lowercase = torch.device(hook_kwargs["""execution_device"""] ) self.assertEqual(model.batchnorm.running_mean.device , lowercase ) __lowercase = torch.randn(2 , 3 ) __lowercase = model(lowercase ) self.assertEqual(output.device , lowercase ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload __lowercase = { """execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True, """offload_buffers""": True, } add_hook_to_module(model.lineara , AlignDevicesHook(**lowercase ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**lowercase ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**lowercase ) ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) __lowercase = torch.randn(2 , 3 ) __lowercase = model(lowercase ) self.assertEqual(output.device , lowercase ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def snake_case__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" __lowercase = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices __lowercase = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook(lowercase , execution_device=lowercase , offload=lowercase ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device __lowercase = torch.device(lowercase ) self.assertEqual(model.batchnorm.running_mean.device , lowercase ) __lowercase = torch.randn(2 , 3 ) __lowercase = model(lowercase ) self.assertEqual(output.device , lowercase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(lowercase ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook(lowercase , execution_device=lowercase , offload=lowercase , offload_buffers=lowercase ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) __lowercase = torch.randn(2 , 3 ) __lowercase = model(lowercase ) self.assertEqual(output.device , lowercase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(lowercase ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def snake_case__ ( self : List[Any] ) -> List[str]: """simple docstring""" __lowercase = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices __lowercase = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook( lowercase , execution_device=lowercase , offload=lowercase , weights_map=model.state_dict() ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device __lowercase = torch.device(lowercase ) self.assertEqual(model.batchnorm.running_mean.device , lowercase ) __lowercase = torch.randn(2 , 3 ) __lowercase = model(lowercase ) self.assertEqual(output.device , lowercase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(lowercase ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook( lowercase , execution_device=lowercase , offload=lowercase , weights_map=model.state_dict() , offload_buffers=lowercase , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) __lowercase = torch.randn(2 , 3 ) __lowercase = model(lowercase ) self.assertEqual(output.device , lowercase ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(lowercase ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) )
634
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_UpperCAmelCase ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : str = field(default="""summarization""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowercase__ : ClassVar[Features] = Features({"""text""": Value("""string""" )} ) lowercase__ : ClassVar[Features] = Features({"""summary""": Value("""string""" )} ) lowercase__ : str = "text" lowercase__ : str = "summary" @property def snake_case__ ( self : List[Any] ) -> Dict[str, str]: """simple docstring""" return {self.text_column: "text", self.summary_column: "summary"}
634
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Any = CycleDiffusionPipeline lowercase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { """negative_prompt""", """height""", """width""", """negative_prompt_embeds""", } lowercase__ : Union[str, Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} lowercase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""source_prompt"""} ) lowercase__ : str = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase__ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def snake_case__ ( self : Any ) -> Union[str, Any]: """simple docstring""" torch.manual_seed(0 ) __lowercase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) __lowercase = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , num_train_timesteps=1_000 , clip_sample=lowercase , set_alpha_to_one=lowercase , ) torch.manual_seed(0 ) __lowercase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) __lowercase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) __lowercase = CLIPTextModel(lowercase ) __lowercase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) __lowercase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def snake_case__ ( self : Tuple , lowercase : Dict , lowercase : Any=0 ) -> Any: """simple docstring""" __lowercase = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = image / 2 + 0.5 if str(lowercase ).startswith("""mps""" ): __lowercase = torch.manual_seed(lowercase ) else: __lowercase = torch.Generator(device=lowercase ).manual_seed(lowercase ) __lowercase = { """prompt""": """An astronaut riding an elephant""", """source_prompt""": """An astronaut riding a horse""", """image""": image, """generator""": generator, """num_inference_steps""": 2, """eta""": 0.1, """strength""": 0.8, """guidance_scale""": 3, """source_guidance_scale""": 1, """output_type""": """numpy""", } return inputs def snake_case__ ( self : Tuple ) -> Any: """simple docstring""" __lowercase = """cpu""" # ensure determinism for the device-dependent torch.Generator __lowercase = self.get_dummy_components() __lowercase = CycleDiffusionPipeline(**lowercase ) __lowercase = pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) __lowercase = self.get_dummy_inputs(lowercase ) __lowercase = pipe(**lowercase ) __lowercase = output.images __lowercase = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) __lowercase = np.array([0.4459, 0.4943, 0.4544, 0.6643, 0.5474, 0.4327, 0.5701, 0.5959, 0.5179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def snake_case__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_dummy_components() for name, module in components.items(): if hasattr(lowercase , """half""" ): __lowercase = module.half() __lowercase = CycleDiffusionPipeline(**lowercase ) __lowercase = pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) __lowercase = self.get_dummy_inputs(lowercase ) __lowercase = pipe(**lowercase ) __lowercase = output.images __lowercase = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) __lowercase = np.array([0.3506, 0.4543, 0.446, 0.4575, 0.5195, 0.4155, 0.5273, 0.518, 0.4116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def snake_case__ ( self : Optional[Any] ) -> str: """simple docstring""" return super().test_save_load_local() @unittest.skip("""non-deterministic pipeline""" ) def snake_case__ ( self : str ) -> Dict: """simple docstring""" return super().test_inference_batch_single_identical() @skip_mps def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" return super().test_dict_tuple_outputs_equivalent() @skip_mps def snake_case__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" return super().test_save_load_optional_components() @skip_mps def snake_case__ ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/cycle-diffusion/black_colored_car.png""" ) __lowercase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy""" ) __lowercase = init_image.resize((512, 512) ) __lowercase = """CompVis/stable-diffusion-v1-4""" __lowercase = DDIMScheduler.from_pretrained(lowercase , subfolder="""scheduler""" ) __lowercase = CycleDiffusionPipeline.from_pretrained( lowercase , scheduler=lowercase , safety_checker=lowercase , torch_dtype=torch.floataa , revision="""fp16""" ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() __lowercase = """A black colored car""" __lowercase = """A blue colored car""" __lowercase = torch.manual_seed(0 ) __lowercase = pipe( prompt=lowercase , source_prompt=lowercase , image=lowercase , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowercase , output_type="""np""" , ) __lowercase = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5E-1 def snake_case__ ( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/cycle-diffusion/black_colored_car.png""" ) __lowercase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy""" ) __lowercase = init_image.resize((512, 512) ) __lowercase = """CompVis/stable-diffusion-v1-4""" __lowercase = DDIMScheduler.from_pretrained(lowercase , subfolder="""scheduler""" ) __lowercase = CycleDiffusionPipeline.from_pretrained(lowercase , scheduler=lowercase , safety_checker=lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() __lowercase = """A black colored car""" __lowercase = """A blue colored car""" __lowercase = torch.manual_seed(0 ) __lowercase = pipe( prompt=lowercase , source_prompt=lowercase , image=lowercase , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowercase , output_type="""np""" , ) __lowercase = output.images assert np.abs(image - expected_image ).max() < 2E-2
634
def UpperCAmelCase__ ( lowercase__ ) -> Optional[int]: __lowercase = len(lowercase__ ) __lowercase = sum(lowercase__ ) __lowercase = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): __lowercase = True for i in range(1 , s + 1 ): __lowercase = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): __lowercase = dp[i][j - 1] if arr[i - 1] <= j: __lowercase = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: __lowercase = s - 2 * j break return diff
634
1
from __future__ import annotations def UpperCAmelCase__ ( lowercase__ ) -> list: if len(lowercase__ ) == 0: return [] __lowercase , __lowercase = min(lowercase__ ), max(lowercase__ ) __lowercase = int(max_value - min_value ) + 1 __lowercase = [[] for _ in range(lowercase__ )] for i in my_list: buckets[int(i - min_value )].append(lowercase__ ) return [v for bucket in buckets for v in sorted(lowercase__ )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
634
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class _lowerCAmelCase : """simple docstring""" def __init__( self : List[Any] , lowercase : str , lowercase : Union[str, Any]=13 , lowercase : Tuple=32 , lowercase : Optional[Any]=2 , lowercase : Tuple=3 , lowercase : Tuple=16 , lowercase : Tuple=[1, 2, 1] , lowercase : Optional[Any]=[2, 2, 4] , lowercase : Dict=2 , lowercase : Optional[int]=2.0 , lowercase : List[Any]=True , lowercase : str=0.0 , lowercase : Any=0.0 , lowercase : Optional[int]=0.1 , lowercase : int="gelu" , lowercase : Tuple=False , lowercase : Optional[Any]=True , lowercase : int=0.02 , lowercase : Union[str, Any]=1E-5 , lowercase : Dict=True , lowercase : Any=None , lowercase : str=True , lowercase : str=10 , lowercase : Dict=8 , lowercase : int=["stage1", "stage2", "stage3"] , lowercase : Optional[int]=[1, 2, 3] , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def snake_case__ ( self : List[str] ) -> int: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def snake_case__ ( self : Any , lowercase : List[Any] , lowercase : Optional[int] , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = MaskFormerSwinModel(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def snake_case__ ( self : Any , lowercase : Tuple , lowercase : Any , lowercase : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(lowercase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=lowercase ) def snake_case__ ( self : int ) -> Any: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Optional[int] = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) lowercase__ : List[str] = {"""feature-extraction""": MaskFormerSwinModel} if is_torch_available() else {} lowercase__ : List[str] = False lowercase__ : int = False lowercase__ : int = False lowercase__ : Tuple = False lowercase__ : Optional[Any] = False def snake_case__ ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=lowercase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" pass def snake_case__ ( self : Tuple ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" return def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def snake_case__ ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowercase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def snake_case__ ( self : int ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def snake_case__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" pass def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def snake_case__ ( self : Any ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def snake_case__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" pass def snake_case__ ( self : Tuple , lowercase : Tuple , lowercase : Optional[int] , lowercase : Optional[int] , lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowercase ) , lowercase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def snake_case__ ( self : int ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , lowercase ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def snake_case__ ( self : Any ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def snake_case__ ( self : List[str] ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" pass def snake_case__ ( self : Tuple ) -> List[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(lowercase : Optional[int] ): __lowercase = 0 return t def check_equivalence(lowercase : Optional[int] , lowercase : str , lowercase : str , lowercase : Tuple={} ): with torch.no_grad(): __lowercase = model(**lowercase , return_dict=lowercase , **lowercase ) __lowercase = model(**lowercase , return_dict=lowercase , **lowercase ).to_tuple() def recursive_check(lowercase : int , lowercase : Optional[Any] ): if isinstance(lowercase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(lowercase , lowercase ): recursive_check(lowercase , lowercase ) elif isinstance(lowercase , lowercase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(lowercase , lowercase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(lowercase ) , set_nan_tensor_to_zero(lowercase ) , atol=1E-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:" F" {torch.isnan(lowercase ).any()} and `inf`: {torch.isinf(lowercase )}. Dict has" F" `nan`: {torch.isnan(lowercase ).any()} and `inf`: {torch.isinf(lowercase )}." ) , ) recursive_check(lowercase , lowercase ) for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() __lowercase = self._prepare_for_class(lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) check_equivalence(lowercase , lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) check_equivalence(lowercase , lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) check_equivalence(lowercase , lowercase , lowercase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) check_equivalence(lowercase , lowercase , lowercase , {"""output_hidden_states""": True} ) @require_torch class _lowerCAmelCase ( unittest.TestCase , _UpperCAmelCase ): """simple docstring""" lowercase__ : List[str] = (MaskFormerSwinBackbone,) if is_torch_available() else () lowercase__ : Any = MaskFormerSwinConfig def snake_case__ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def snake_case__ ( self : Any ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(lowercase ) backbone.to(lowercase ) backbone.eval() __lowercase = backbone(**lowercase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , lowercase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**lowercase , output_hidden_states=lowercase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**lowercase , output_attentions=lowercase ) self.assertIsNotNone(outputs.attentions )
634
1
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed UpperCamelCase__ = logging.getLogger(__name__) def UpperCAmelCase__ ( lowercase__=2 , lowercase__=3 , lowercase__=16 , lowercase__ = 10 , lowercase__ = 2 ) -> List[Any]: def get_dataset(lowercase__ ): __lowercase = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(lowercase__ , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) __lowercase = get_dataset(lowercase__ ) __lowercase = get_dataset(lowercase__ ) __lowercase = DataLoader(lowercase__ , shuffle=lowercase__ , batch_size=lowercase__ , num_workers=4 ) __lowercase = DataLoader(lowercase__ , shuffle=lowercase__ , batch_size=lowercase__ , num_workers=4 ) return (train_dataloader, valid_dataloader) def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__=None ) -> Optional[int]: __lowercase = [] for epoch in range(lowercase__ ): # Train quickly model.train() for batch in dataloader: __lowercase , __lowercase = batch __lowercase = model(lowercase__ ) __lowercase = torch.nn.functional.mse_loss(lowercase__ , lowercase__ ) accelerator.backward(lowercase__ ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Dict ) -> int: """simple docstring""" super().__init__() __lowercase = nn.Parameter(torch.randn(1 ) ) __lowercase = nn.Parameter(torch.randn(1 ) ) def snake_case__ ( self : Union[str, Any] , lowercase : Dict ) -> Union[str, Any]: """simple docstring""" return x * self.a + self.b class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Optional[Any] ) -> List[str]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __lowercase , __lowercase = dummy_dataloaders() __lowercase = ProjectConfiguration(total_limit=1 , project_dir=lowercase , automatic_checkpoint_naming=lowercase ) # Train baseline __lowercase = Accelerator(project_config=lowercase ) __lowercase , __lowercase , __lowercase , __lowercase = accelerator.prepare( lowercase , lowercase , lowercase , lowercase ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def snake_case__ ( self : Dict ) -> str: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __lowercase , __lowercase = dummy_dataloaders() # Train baseline __lowercase = Accelerator() __lowercase , __lowercase , __lowercase , __lowercase = accelerator.prepare( lowercase , lowercase , lowercase , lowercase ) # Save initial __lowercase = os.path.join(lowercase , """initial""" ) accelerator.save_state(lowercase ) ((__lowercase) , (__lowercase)) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() __lowercase = train(3 , lowercase , lowercase , lowercase , lowercase ) ((__lowercase) , (__lowercase)) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() # Train partially set_seed(42 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __lowercase , __lowercase = dummy_dataloaders() __lowercase = Accelerator() __lowercase , __lowercase , __lowercase , __lowercase = accelerator.prepare( lowercase , lowercase , lowercase , lowercase ) accelerator.load_state(lowercase ) ((__lowercase) , (__lowercase)) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) __lowercase = train(2 , lowercase , lowercase , lowercase , lowercase ) # Save everything __lowercase = os.path.join(lowercase , """checkpoint""" ) accelerator.save_state(lowercase ) # Load everything back in and make sure all states work accelerator.load_state(lowercase ) test_rands += train(1 , lowercase , lowercase , lowercase , lowercase ) ((__lowercase) , (__lowercase)) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __lowercase , __lowercase = dummy_dataloaders() __lowercase = ProjectConfiguration(automatic_checkpoint_naming=lowercase ) # Train baseline __lowercase = Accelerator(project_dir=lowercase , project_config=lowercase ) __lowercase , __lowercase , __lowercase , __lowercase = accelerator.prepare( lowercase , lowercase , lowercase , lowercase ) # Save initial accelerator.save_state() ((__lowercase) , (__lowercase)) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() __lowercase = train(3 , lowercase , lowercase , lowercase , lowercase ) ((__lowercase) , (__lowercase)) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() # Train partially set_seed(42 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __lowercase , __lowercase = dummy_dataloaders() __lowercase = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=lowercase ) __lowercase = Accelerator(project_dir=lowercase , project_config=lowercase ) __lowercase , __lowercase , __lowercase , __lowercase = accelerator.prepare( lowercase , lowercase , lowercase , lowercase ) accelerator.load_state(os.path.join(lowercase , """checkpoints""" , """checkpoint_0""" ) ) ((__lowercase) , (__lowercase)) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) __lowercase = train(2 , lowercase , lowercase , lowercase , lowercase ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(lowercase , """checkpoints""" , """checkpoint_1""" ) ) test_rands += train(1 , lowercase , lowercase , lowercase , lowercase ) ((__lowercase) , (__lowercase)) = model.a.item(), model.b.item() __lowercase = optimizer.state_dict() self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = torch.tensor([1, 2, 3] ) __lowercase = torch.tensor([2, 3, 4] ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(net.parameters() ) __lowercase = Accelerator() with self.assertRaises(lowercase ) as ve: accelerator.register_for_checkpointing(lowercase , lowercase , lowercase , lowercase ) __lowercase = str(ve.exception ) self.assertTrue("""Item at index 0""" in message ) self.assertTrue("""Item at index 1""" in message ) self.assertFalse("""Item at index 2""" in message ) self.assertFalse("""Item at index 3""" in message ) def snake_case__ ( self : Optional[int] ) -> int: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __lowercase = DummyModel() __lowercase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __lowercase = torch.optim.lr_scheduler.StepLR(lowercase , step_size=1 , gamma=0.99 ) __lowercase , __lowercase = dummy_dataloaders() __lowercase = ProjectConfiguration(automatic_checkpoint_naming=lowercase ) # Train baseline __lowercase = Accelerator(project_dir=lowercase , project_config=lowercase ) __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = accelerator.prepare( lowercase , lowercase , lowercase , lowercase , lowercase ) # Save initial accelerator.save_state() __lowercase = scheduler.state_dict() train(3 , lowercase , lowercase , lowercase , lowercase , lowercase ) self.assertNotEqual(lowercase , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(lowercase , """checkpoints""" , """checkpoint_0""" ) ) self.assertEqual(lowercase , scheduler.state_dict() ) def snake_case__ ( self : Optional[int] ) -> List[Any]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __lowercase = DummyModel() __lowercase = ProjectConfiguration(automatic_checkpoint_naming=lowercase , total_limit=2 ) # Train baseline __lowercase = Accelerator(project_dir=lowercase , project_config=lowercase ) __lowercase = accelerator.prepare(lowercase ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(lowercase , """checkpoints""" , """checkpoint_0""" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase , """checkpoints""" , """checkpoint_9""" ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase , """checkpoints""" , """checkpoint_10""" ) ) ) @require_cuda def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = ["""torchrun""", F"--nproc_per_node={torch.cuda.device_count()}", inspect.getfile(self.__class__ )] execute_subprocess_async(lowercase , env=os.environ.copy() ) if __name__ == "__main__": UpperCamelCase__ = "/tmp/accelerate/state_checkpointing" UpperCamelCase__ = DummyModel() UpperCamelCase__ = torch.optim.Adam(params=model.parameters(), lr=1e-3) UpperCamelCase__ = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) UpperCamelCase__ , UpperCamelCase__ = dummy_dataloaders() UpperCamelCase__ = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline UpperCamelCase__ = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision="no") if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) UpperCamelCase__ , UpperCamelCase__ = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: UpperCamelCase__ = group["params"][0].device break assert param_device.type == accelerator.device.type UpperCamelCase__ = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, "checkpoints", "checkpoint_0"), map_location="cpu") for group in optimizer.param_groups: UpperCamelCase__ = group["params"][0].device break assert ( param_device.type == torch.device("cpu").type ), F"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, "checkpoints", "checkpoint_0"), map_location="on_device") for group in optimizer.param_groups: UpperCamelCase__ = group["params"][0].device break assert ( param_device.type == accelerator.device.type ), F"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match="Unsupported optimizer map location passed"): accelerator.load_state(os.path.join(savedir, "checkpoints", "checkpoint_0"), map_location="invalid") accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
634
import importlib import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Union import torch from ..utils import BaseOutput UpperCamelCase__ = "scheduler_config.json" class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : List[Any] = 1 lowercase__ : Tuple = 2 lowercase__ : Union[str, Any] = 3 lowercase__ : Union[str, Any] = 4 lowercase__ : str = 5 lowercase__ : Any = 6 lowercase__ : Any = 7 lowercase__ : List[str] = 8 lowercase__ : Union[str, Any] = 9 lowercase__ : int = 10 lowercase__ : List[str] = 11 lowercase__ : List[Any] = 12 lowercase__ : str = 13 lowercase__ : Optional[int] = 14 @dataclass class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : torch.FloatTensor class _lowerCAmelCase : """simple docstring""" lowercase__ : Optional[int] = SCHEDULER_CONFIG_NAME lowercase__ : int = [] lowercase__ : Dict = True @classmethod def snake_case__ ( cls : str , lowercase : Dict[str, Any] = None , lowercase : Optional[str] = None , lowercase : Any=False , **lowercase : List[str] , ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase , __lowercase = cls.load_config( pretrained_model_name_or_path=lowercase , subfolder=lowercase , return_unused_kwargs=lowercase , return_commit_hash=lowercase , **lowercase , ) return cls.from_config(lowercase , return_unused_kwargs=lowercase , **lowercase ) def snake_case__ ( self : Dict , lowercase : Union[str, os.PathLike] , lowercase : bool = False , **lowercase : List[str] ) -> Optional[Any]: """simple docstring""" self.save_config(save_directory=lowercase , push_to_hub=lowercase , **lowercase ) @property def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" return self._get_compatibles() @classmethod def snake_case__ ( cls : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = list(set([cls.__name__] + cls._compatibles ) ) __lowercase = importlib.import_module(__name__.split(""".""" )[0] ) __lowercase = [ getattr(lowercase , lowercase ) for c in compatible_classes_str if hasattr(lowercase , lowercase ) ] return compatible_classes
634
1
import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self : Union[str, Any] , lowercase : Any , lowercase : Optional[Any]=13 , lowercase : Optional[Any]=30 , lowercase : Optional[Any]=2 , lowercase : Optional[Any]=3 , lowercase : str=True , lowercase : Dict=True , lowercase : int=32 , lowercase : Optional[Any]=5 , lowercase : Dict=4 , lowercase : Any=37 , lowercase : Optional[Any]="gelu" , lowercase : List[Any]=0.1 , lowercase : Union[str, Any]=0.1 , lowercase : List[str]=10 , lowercase : str=0.02 , lowercase : int=3 , lowercase : Dict=0.6 , lowercase : List[Any]=None , ) -> str: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = mask_ratio __lowercase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) __lowercase = (image_size // patch_size) ** 2 __lowercase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def snake_case__ ( self : str ) -> List[Any]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def snake_case__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def snake_case__ ( self : List[Any] , lowercase : Optional[Any] , lowercase : Union[str, Any] , lowercase : List[str] ) -> List[str]: """simple docstring""" __lowercase = ViTMAEModel(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : str , lowercase : List[Any] , lowercase : Union[str, Any] , lowercase : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = ViTMAEForPreTraining(lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) __lowercase = (self.image_size // self.patch_size) ** 2 __lowercase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images __lowercase = 1 __lowercase = ViTMAEForPreTraining(lowercase ) model.to(lowercase ) model.eval() __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase ) __lowercase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def snake_case__ ( self : int ) -> Optional[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Dict = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () lowercase__ : str = {"""feature-extraction""": ViTMAEModel} if is_torch_available() else {} lowercase__ : Optional[int] = False lowercase__ : Tuple = False lowercase__ : List[Any] = False lowercase__ : Dict = False def snake_case__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = ViTMAEModelTester(self ) __lowercase = ConfigTester(self , config_class=lowercase , has_text_modality=lowercase , hidden_size=37 ) def snake_case__ ( self : int ) -> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def snake_case__ ( self : Tuple ) -> List[Any]: """simple docstring""" pass def snake_case__ ( self : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def snake_case__ ( self : int ) -> Optional[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def snake_case__ ( self : Any ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase ) def snake_case__ ( self : int , lowercase : Dict , lowercase : int , lowercase : List[Any] ) -> Tuple: """simple docstring""" np.random.seed(2 ) __lowercase = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) __lowercase = torch.from_numpy(lowercase ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument __lowercase = pt_noise super().check_pt_tf_models(lowercase , lowercase , lowercase ) def snake_case__ ( self : int ) -> Any: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) __lowercase = outputs[0].cpu().numpy() __lowercase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase ) __lowercase = model_class.from_pretrained(lowercase ) model.to(lowercase ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) # Make sure we don't have nans __lowercase = after_outputs[0].cpu().numpy() __lowercase = 0 __lowercase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowercase , 1E-5 ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def snake_case__ ( self : List[Any] ) -> List[str]: """simple docstring""" pass @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def snake_case__ ( self : List[str] ) -> Dict: """simple docstring""" pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def snake_case__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def snake_case__ ( self : Dict ) -> int: """simple docstring""" pass @slow def snake_case__ ( self : str ) -> Any: """simple docstring""" for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ViTMAEModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def UpperCAmelCase__ ( ) -> Tuple: __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case__ ( self : int ) -> Any: """simple docstring""" return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def snake_case__ ( self : Dict ) -> List[str]: """simple docstring""" np.random.seed(2 ) __lowercase = ViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ).to(lowercase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase , return_tensors="""pt""" ).to(lowercase ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) __lowercase = ViTMAEConfig() __lowercase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase , noise=torch.from_numpy(lowercase ).to(device=lowercase ) ) # verify the logits __lowercase = torch.Size((1, 196, 768) ) self.assertEqual(outputs.logits.shape , lowercase ) __lowercase = torch.tensor( [[-0.0548, -1.7023, -0.9325], [0.3721, -0.5670, -0.2233], [0.8235, -1.3878, -0.3524]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(lowercase ) , atol=1E-4 ) )
634
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "asapp/sew-tiny-100k": "https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json", # See all SEW models at https://huggingface.co/models?filter=sew } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """sew""" def __init__( self : List[Any] , lowercase : int=32 , lowercase : List[str]=768 , lowercase : Dict=12 , lowercase : str=12 , lowercase : str=3_072 , lowercase : Optional[int]=2 , lowercase : List[str]="gelu" , lowercase : List[str]=0.1 , lowercase : Tuple=0.1 , lowercase : Dict=0.1 , lowercase : Any=0.0 , lowercase : Dict=0.1 , lowercase : Optional[int]=0.1 , lowercase : List[str]=0.02 , lowercase : Dict=1E-5 , lowercase : Tuple="group" , lowercase : int="gelu" , lowercase : Any=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , lowercase : Optional[int]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowercase : Any=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowercase : List[str]=False , lowercase : Tuple=128 , lowercase : int=16 , lowercase : Union[str, Any]=True , lowercase : List[str]=0.05 , lowercase : Optional[int]=10 , lowercase : Any=2 , lowercase : Optional[Any]=0.0 , lowercase : Optional[Any]=10 , lowercase : int=0 , lowercase : Optional[int]="mean" , lowercase : List[Any]=False , lowercase : str=False , lowercase : int=256 , lowercase : str=0 , lowercase : List[Any]=1 , lowercase : List[Any]=2 , **lowercase : List[Any] , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase , pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase ) __lowercase = hidden_size __lowercase = feat_extract_norm __lowercase = feat_extract_activation __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = conv_bias __lowercase = num_conv_pos_embeddings __lowercase = num_conv_pos_embedding_groups __lowercase = len(self.conv_dim ) __lowercase = num_hidden_layers __lowercase = intermediate_size __lowercase = squeeze_factor __lowercase = hidden_act __lowercase = num_attention_heads __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = feat_proj_dropout __lowercase = final_dropout __lowercase = layerdrop __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" F"but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)" F"= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowercase = apply_spec_augment __lowercase = mask_time_prob __lowercase = mask_time_length __lowercase = mask_time_min_masks __lowercase = mask_feature_prob __lowercase = mask_feature_length __lowercase = mask_feature_min_masks # ctc loss __lowercase = ctc_loss_reduction __lowercase = ctc_zero_infinity # sequence classification __lowercase = use_weighted_layer_sum __lowercase = classifier_proj_size @property def snake_case__ ( self : Dict ) -> str: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
634
1
from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class _lowerCAmelCase : """simple docstring""" def snake_case__ ( self : Optional[int] , lowercase : List[Any] ) -> List[Any]: """simple docstring""" raise NotImplementedError() def snake_case__ ( self : Tuple ) -> str: """simple docstring""" raise NotImplementedError() class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , lowercase : "AutoTokenizer" , lowercase : bool = False , **lowercase : str ) -> Tuple: """simple docstring""" __lowercase = tokenizer __lowercase = skip_prompt __lowercase = decode_kwargs # variables used in the streaming process __lowercase = [] __lowercase = 0 __lowercase = True def snake_case__ ( self : Dict , lowercase : Optional[Any] ) -> List[Any]: """simple docstring""" if len(value.shape ) > 1 and value.shape[0] > 1: raise ValueError("""TextStreamer only supports batch size 1""" ) elif len(value.shape ) > 1: __lowercase = value[0] if self.skip_prompt and self.next_tokens_are_prompt: __lowercase = False return # Add the new token to the cache and decodes the entire thing. self.token_cache.extend(value.tolist() ) __lowercase = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) # After the symbol for a new line, we flush the cache. if text.endswith("""\n""" ): __lowercase = text[self.print_len :] __lowercase = [] __lowercase = 0 # If the last token is a CJK character, we print the characters. elif len(lowercase ) > 0 and self._is_chinese_char(ord(text[-1] ) ): __lowercase = text[self.print_len :] self.print_len += len(lowercase ) # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: __lowercase = text[self.print_len : text.rfind(""" """ ) + 1] self.print_len += len(lowercase ) self.on_finalized_text(lowercase ) def snake_case__ ( self : Any ) -> Optional[Any]: """simple docstring""" if len(self.token_cache ) > 0: __lowercase = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) __lowercase = text[self.print_len :] __lowercase = [] __lowercase = 0 else: __lowercase = """""" __lowercase = True self.on_finalized_text(lowercase , stream_end=lowercase ) def snake_case__ ( self : str , lowercase : str , lowercase : bool = False ) -> Dict: """simple docstring""" print(lowercase , flush=lowercase , end="""""" if not stream_end else None ) def snake_case__ ( self : List[str] , lowercase : Optional[int] ) -> Optional[int]: """simple docstring""" if ( (cp >= 0X4e00 and cp <= 0X9fff) or (cp >= 0X3400 and cp <= 0X4dbf) # or (cp >= 0X2_0000 and cp <= 0X2_a6df) # or (cp >= 0X2_a700 and cp <= 0X2_b73f) # or (cp >= 0X2_b740 and cp <= 0X2_b81f) # or (cp >= 0X2_b820 and cp <= 0X2_ceaf) # or (cp >= 0Xf900 and cp <= 0Xfaff) or (cp >= 0X2_f800 and cp <= 0X2_fa1f) # ): # return True return False class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowercase : "AutoTokenizer" , lowercase : bool = False , lowercase : Optional[float] = None , **lowercase : Tuple ) -> Dict: """simple docstring""" super().__init__(lowercase , lowercase , **lowercase ) __lowercase = Queue() __lowercase = None __lowercase = timeout def snake_case__ ( self : Optional[Any] , lowercase : str , lowercase : bool = False ) -> Union[str, Any]: """simple docstring""" self.text_queue.put(lowercase , timeout=self.timeout ) if stream_end: self.text_queue.put(self.stop_signal , timeout=self.timeout ) def __iter__( self : str ) -> List[str]: """simple docstring""" return self def snake_case__ ( self : str ) -> List[Any]: """simple docstring""" __lowercase = self.text_queue.get(timeout=self.timeout ) if value == self.stop_signal: raise StopIteration() else: return value
634
import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizerOutput from transformers.testing_utils import require_phonemizer from ...test_tokenization_common import TokenizerTesterMixin @require_phonemizer class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = WavaVecaPhonemeCTCTokenizer lowercase__ : Optional[int] = False def snake_case__ ( self : str ) -> int: """simple docstring""" super().setUp() __lowercase = ( """<s> <pad> </s> <unk> n s t ə l a i k d m ɛ ɾ e ɪ p o ɐ z ð f j v b ɹ ʁ ʊ iː r w ʌ u ɡ æ aɪ ʃ h ɔ ɑː """ """ŋ ɚ eɪ β uː y ɑ̃ oʊ ᵻ eː θ aʊ ts oː ɔ̃ ɣ ɜ ɑ dʒ əl x ɜː ç ʒ tʃ ɔː ɑːɹ ɛ̃ ʎ ɔːɹ ʋ aː ɕ œ ø oːɹ ɲ yː """ """ʔ iə i5 s. tɕ ?? nʲ ɛː œ̃ ɭ ɔø ʑ tʲ ɨ ɛɹ ts. rʲ ɪɹ ɭʲ i.5 ɔɪ q sʲ u5 ʊɹ iɜ a5 iɛ5 øː ʕ ja əɜ th ɑ5 """ """oɪ dʲ ə5 tɕh ts.h mʲ ɯ dʑ vʲ e̞ tʃʲ ei5 o5 onɡ5 ɑu5 iɑ5 ai5 aɪɚ kh ə1 ʐ i2 ʉ ħ t[ aɪə ʲ ju ə2 u2 oɜ """ """pː iɛɜ ou5 y5 uɜ tː uo5 d[ uoɜ tsh ɑɜ ɵ i̪5 uei5 ɟ aɜ ɑɨ i.ɜ eʊ o2 ɐ̃ ä pʲ kʲ n̩ ɒ ph ɑu2 uɨ əɪ ɫ ɬ """ """yɜ bʲ ɑ2 s̪ aiɜ χ ɐ̃ʊ̃ 1 ə4 yæɜ a2 ɨː t̪ iouɜ ũ onɡɜ aɨ iɛ2 ɔɨ ɑuɜ o̞ ei2 iou2 c kː y2 ɖ oe dˤ yɛɜ """ """əʊ S ɡʲ onɡ2 u\" eiɜ ʈ ɯᵝ iou5 dZ r̝̊ i.2 tS s^ ʝ yə5 iɑɜ uə5 pf ɨu iɑ2 ou2 ər2 fʲ ai2 r̝ uəɜ ɳ əɨ """ """ua5 uɪ ɽ bː yu5 uo2 yɛ5 l̩ ɻ ərɜ ʂ i̪2 ouɜ uaɜ a. a.ː yæ5 dː r̩ ee ɪu ər5 i̪ ɜ æi u: i.ː t^ o1 ɪ^ """ """ai ueiɜ æː ɛɪ eə i. ɴ ie ua2 ɑ1 o4 tʃː o: ɑ: u1 N i̪1 au yæ2 u. qː yəɜ y: kʰ tʃʰ iʊ sx õ uo tʰ """ """uai5 bʰ u.ː uə2 ʊə d^ s̪ː yiɜ dʰ r. oe: i1 ɟː yu2 nʲʲ i̪4 uei2 tsʲ ɸ ĩ ɑ4 t̪ː eɑ u4 e: tsː ʈʰ ɡʰ """ """ɯɯ dʒʲ ʂʲ X ɵː uaiɜ tɕʲ ã t^ː ẽː yɛ2 cː i.1 ɛʊ dˤdˤ dʒː i4 ɡː yi ɕʲ ɟʰ pʰ dʑʲ yuɜ ua1 ua4 æiː ɐɐ """ """ui iou1 ʊː a1 iou4 cʰ iɛ1 yə2 ɖʰ ẽ ʒʲ ää ər4 iːː ɪː iɑ1 ər1 œː øi ɪuː cʰcʰ əː1 iː1 ũ kʰː o̞o̞ xʲ """ """ou1 iɛ4 e̞e̞ y1 dzː dʲʲ dʰː ɯᵝɯᵝ lː uo1 i.4 i: yɛ5ʲ a4""" ).split(""" """ ) __lowercase = dict(zip(lowercase , range(len(lowercase ) ) ) ) __lowercase = {"""pad_token""": """<pad>""", """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>"""} __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowercase ) + """\n""" ) def snake_case__ ( self : List[Any] , lowercase : Optional[Any] , lowercase : List[str]=False , lowercase : List[str]=20 , lowercase : str=5 ) -> Tuple[str, list]: """simple docstring""" __lowercase = [(i, tokenizer.decode([i] , clean_up_tokenization_spaces=lowercase )) for i in range(len(lowercase ) )] __lowercase = list(filter(lambda lowercase : [t[0]] == tokenizer.encode(t[1] , do_phonemize=lowercase ) , lowercase ) ) if max_length is not None and len(lowercase ) > max_length: __lowercase = toks[:max_length] if min_length is not None and len(lowercase ) < min_length and len(lowercase ) > 0: while len(lowercase ) < min_length: __lowercase = toks + toks # toks_str = [t[1] for t in toks] __lowercase = [t[0] for t in toks] # Ensure consistency __lowercase = tokenizer.decode(lowercase , clean_up_tokenization_spaces=lowercase ) if " " not in output_txt and len(lowercase ) > 1: __lowercase = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=lowercase ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=lowercase ) ) if with_prefix_space: __lowercase = """ """ + output_txt __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) return output_txt, output_ids def snake_case__ ( self : Tuple , **lowercase : int ) -> Union[str, Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return WavaVecaPhonemeCTCTokenizer.from_pretrained(self.tmpdirname , **lowercase ) def snake_case__ ( self : Any ) -> List[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) # check adding a single token tokenizer.add_tokens("""xxx""" ) __lowercase = tokenizer("""m xxx ɪ""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 392, 17] ) # xxx should be last token tokenizer.add_tokens(["""aaa""", """bbb""", """ccc"""] ) __lowercase = tokenizer("""m aaa ɪ ccc""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 393, 17, 395] ) # aaa and ccc should be after xxx and 2 after aaa __lowercase = tokenizer("""maɪ c""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [3, 200] ) # mai should be <unk> (=3) def snake_case__ ( self : int ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) def snake_case__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98], [24, 22, 5, 24, 22, 5, 77], ] __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> Optional[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ | h aʊ | ɑːɹ | j uː |""" ) def snake_case__ ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : List[Any] ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, tokenizer.word_delimiter_token_id, 15, 8, tokenizer.word_delimiter_token_id, 98], [tokenizer.word_delimiter_token_id, 24, 22, tokenizer.word_delimiter_token_id, 5, 24, 22, 5, 77], ] # fmt: on # decode with word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) # decode with no word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] , filter_word_delimiter_token=lowercase ) __lowercase = tokenizer.batch_decode(lowercase , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ | ɾ l | ɭʲ""", """| j ð | s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : Optional[int] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(""" """.join([p.strip() for p in phonemes.split(""" |""" )] ).strip() , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token=lowercase ) __lowercase = """Hello how are you""" __lowercase = tokenizer(lowercase , phonemizer_lang="""en-us""" ).input_ids __lowercase = tokenizer(lowercase , phonemizer_lang="""fr-fr""" ).input_ids self.assertNotEqual(lowercase , lowercase ) __lowercase = tokenizer.decode(lowercase ) __lowercase = tokenizer.decode(lowercase ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) self.assertEqual(lowercase , """ɛ l o h aʊ a ʁ j u""" ) def snake_case__ ( self : int ) -> int: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how Are you""" __lowercase = """hello how are you""" __lowercase = tokenizer(lowercase ).input_ids __lowercase = tokenizer(lowercase ).input_ids self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : int ) -> Union[str, Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) tokenizer.add_tokens(["""!""", """?"""] ) tokenizer.add_special_tokens({"""cls_token""": """$$$"""} ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98, 392, 392, 393, 392, 392, 393, 394, 394], [24, 22, 5, 24, 22, 5, 77, tokenizer.pad_token_id, 394, 394], ] # fmt: on __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ!?!? $$$""", """j ð s j ð s oːɹ $$$"""] ) @staticmethod def snake_case__ ( lowercase : List[str] , lowercase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = [d[key] for d in offsets] return retrieved_list def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off # ksssɾɾ|ɾɾ<pad>ɾɾ|<pad>ɾlll|ɭʲ -> k s ɾ ɾ | ɾ l | ɭʲ" __lowercase = [11, 5, 5, 5, 15, 15, tokenizer.pad_token_id, 15, 15, tokenizer.word_delimiter_token_id, tokenizer.pad_token_id, 15, 8, 8, 8, tokenizer.word_delimiter_token_id, 98] # fmt: on __lowercase = tokenizer.decode(lowercase , output_char_offsets=lowercase , filter_word_delimiter_token=lowercase ) # check Wav2Vec2CTCTokenizerOutput keys for char self.assertEqual(len(outputs.keys() ) , 2 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""char_offsets""" in outputs ) self.assertTrue(isinstance(lowercase , lowercase ) ) # check that order of chars is correct and identical for both outputs self.assertEqual(""" """.join(self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) ) , outputs.text ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) , ["""k""", """s""", """ɾ""", """ɾ""", """|""", """ɾ""", """l""", """|""", """ɭʲ"""] ) # check that offsets are actually correct for char # 0-1 is 11, 1-4 is 5, 4-6 is first 15, 6-7 is <pad> (thus not shown), 7-9 is second 15, 9-10 is word_delimiter_token, # 10-11 is <pad> (thus not shown), 11-12 is third 15, 12-15 is 8, 15-16 is word_delimiter_token, 16-17 is 98 self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """start_offset""" ) , [0, 1, 4, 7, 9, 11, 12, 15, 16] ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """end_offset""" ) , [1, 4, 6, 9, 10, 12, 15, 16, 17] ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) def check_list_tuples_equal(lowercase : List[str] , lowercase : Dict ): self.assertTrue(isinstance(lowercase , lowercase ) ) self.assertTrue(isinstance(outputs_list[0] , lowercase ) ) # transform list to ModelOutput __lowercase = WavaVecaPhonemeCTCTokenizerOutput( {k: [d[k] for d in outputs_list] for k in outputs_list[0]} ) self.assertListEqual(outputs_batch["""text"""] , outputs_batch_a["""text"""] ) def recursive_check(lowercase : List[Any] , lowercase : Optional[int] ): if isinstance(lowercase , lowercase ): [recursive_check(lowercase , lowercase ) for la, la in zip(lowercase , lowercase )] self.assertEqual(lowercase , lowercase ) if "char_offsets" in outputs_batch: recursive_check(outputs_batch["""char_offsets"""] , outputs_batch_a["""char_offsets"""] ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 4, 8, 98, 32, 32, 32, 32, 4, 33, tokenizer.word_delimiter_token_id, 32, 32, 33, 34, 34], [24, 22, 5, tokenizer.word_delimiter_token_id, tokenizer.word_delimiter_token_id, 24, 22, 22, 22, 4, 5, 77, tokenizer.pad_token_id, 22, 22, 4, 34, 34, 34, 34], ] # fmt: on # We assume that `decode` works as expected. All we will check now is # the output type is correct and the output is identical to `decode` # char __lowercase = tokenizer.batch_decode(lowercase , output_char_offsets=lowercase ) __lowercase = [tokenizer.decode(lowercase , output_char_offsets=lowercase ) for ids in sample_ids] check_list_tuples_equal(lowercase , lowercase ) @unittest.skip("""Wav2Vec2PhonemeTokenizer always lower cases letters to correctly map to phonemes""" ) def snake_case__ ( self : List[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeTokenizer always puts spaces between phonemes""" ) def snake_case__ ( self : List[str] ) -> List[str]: """simple docstring""" pass @unittest.skip("""encodes to text to ids, but decodes ids to phonemes -> not possible to have internal consistency""" ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeModel has no max model length => no testing""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass def snake_case__ ( self : Any ) -> int: """simple docstring""" __lowercase = self.get_tokenizers(do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) __lowercase = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] __lowercase = tokenizer.add_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size + len(lowercase ) ) __lowercase = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) __lowercase = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} __lowercase = tokenizer.add_special_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size_a + len(lowercase ) ) __lowercase = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" pass def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizers(fast=lowercase , do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = ["""ð""", """ɪ""", """s""", """ɪ""", """z""", """ɐ""", """t""", """ɛ""", """k""", """s""", """t"""] __lowercase = tokenizer.convert_tokens_to_string(lowercase ) self.assertIsInstance(output["""text"""] , lowercase )
634
1
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def UpperCAmelCase__ ( lowercase__ ) -> Tuple: return EnvironmentCommand() def UpperCAmelCase__ ( lowercase__ ) -> Union[str, Any]: return EnvironmentCommand(args.accelerate_config_file ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" @staticmethod def snake_case__ ( lowercase : ArgumentParser ) -> Any: """simple docstring""" __lowercase = parser.add_parser("""env""" ) download_parser.set_defaults(func=lowercase ) download_parser.add_argument( """--accelerate-config_file""" , default=lowercase , help="""The accelerate config file to use for the default values in the launching script.""" , ) download_parser.set_defaults(func=lowercase ) def __init__( self : Dict , lowercase : Tuple , *lowercase : Optional[int] ) -> None: """simple docstring""" __lowercase = accelerate_config_file def snake_case__ ( self : List[Any] ) -> Any: """simple docstring""" __lowercase = """not installed""" if is_safetensors_available(): import safetensors __lowercase = safetensors.__version__ elif importlib.util.find_spec("""safetensors""" ) is not None: import safetensors __lowercase = F"{safetensors.__version__} but is ignored because of PyTorch version too old." __lowercase = """not installed""" __lowercase = __lowercase = """not found""" if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file __lowercase = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(lowercase ): __lowercase = load_config_from_file(self._accelerate_config_file ).to_dict() __lowercase = ( """\n""".join([F"\t- {prop}: {val}" for prop, val in accelerate_config.items()] ) if isinstance(lowercase , lowercase ) else F"\t{accelerate_config}" ) __lowercase = """not installed""" __lowercase = """NA""" if is_torch_available(): import torch __lowercase = torch.__version__ __lowercase = torch.cuda.is_available() __lowercase = """not installed""" __lowercase = """NA""" if is_tf_available(): import tensorflow as tf __lowercase = tf.__version__ try: # deprecated in v2.1 __lowercase = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool __lowercase = bool(tf.config.list_physical_devices("""GPU""" ) ) __lowercase = """not installed""" __lowercase = """not installed""" __lowercase = """not installed""" __lowercase = """NA""" if is_flax_available(): import flax import jax import jaxlib __lowercase = flax.__version__ __lowercase = jax.__version__ __lowercase = jaxlib.__version__ __lowercase = jax.lib.xla_bridge.get_backend().platform __lowercase = { """`transformers` version""": version, """Platform""": platform.platform(), """Python version""": platform.python_version(), """Huggingface_hub version""": huggingface_hub.__version__, """Safetensors version""": F"{safetensors_version}", """Accelerate version""": F"{accelerate_version}", """Accelerate config""": F"{accelerate_config_str}", """PyTorch version (GPU?)""": F"{pt_version} ({pt_cuda_available})", """Tensorflow version (GPU?)""": F"{tf_version} ({tf_cuda_available})", """Flax version (CPU?/GPU?/TPU?)""": F"{flax_version} ({jax_backend})", """Jax version""": F"{jax_version}", """JaxLib version""": F"{jaxlib_version}", """Using GPU in script?""": """<fill in>""", """Using distributed or parallel set-up in script?""": """<fill in>""", } print("""\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n""" ) print(self.format_dict(lowercase ) ) return info @staticmethod def snake_case__ ( lowercase : str ) -> str: """simple docstring""" return "\n".join([F"- {prop}: {val}" for prop, val in d.items()] ) + "\n"
634
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) # TODO Update this UpperCamelCase__ = { "facebook/esm-1b": "https://huggingface.co/facebook/esm-1b/resolve/main/config.json", # See all ESM models at https://huggingface.co/models?filter=esm } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """esm""" def __init__( self : Any , lowercase : Optional[Any]=None , lowercase : Optional[int]=None , lowercase : List[Any]=None , lowercase : Optional[int]=768 , lowercase : str=12 , lowercase : Union[str, Any]=12 , lowercase : Dict=3_072 , lowercase : Optional[int]=0.1 , lowercase : str=0.1 , lowercase : Dict=1_026 , lowercase : Tuple=0.02 , lowercase : str=1E-1_2 , lowercase : Dict="absolute" , lowercase : Optional[Any]=True , lowercase : int=None , lowercase : int=False , lowercase : List[str]=False , lowercase : Tuple=None , lowercase : Tuple=None , **lowercase : Union[str, Any] , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase , mask_token_id=lowercase , **lowercase ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = emb_layer_norm_before __lowercase = token_dropout __lowercase = is_folding_model if is_folding_model: if esmfold_config is None: logger.info("""No esmfold_config supplied for folding model, using default values.""" ) __lowercase = EsmFoldConfig() elif isinstance(lowercase , lowercase ): __lowercase = EsmFoldConfig(**lowercase ) __lowercase = esmfold_config if vocab_list is None: logger.warning("""No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!""" ) __lowercase = get_default_vocab_list() else: __lowercase = vocab_list else: __lowercase = None __lowercase = None if self.esmfold_config is not None and getattr(self.esmfold_config , """use_esm_attn_map""" , lowercase ): raise ValueError("""The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!""" ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = super().to_dict() if isinstance(self.esmfold_config , lowercase ): __lowercase = self.esmfold_config.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : str = None lowercase__ : bool = True lowercase__ : bool = False lowercase__ : bool = False lowercase__ : bool = False lowercase__ : float = 0 lowercase__ : bool = True lowercase__ : bool = False lowercase__ : int = 128 lowercase__ : "TrunkConfig" = None def snake_case__ ( self : List[str] ) -> Any: """simple docstring""" if self.trunk is None: __lowercase = TrunkConfig() elif isinstance(self.trunk , lowercase ): __lowercase = TrunkConfig(**self.trunk ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.trunk.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 48 lowercase__ : int = 1_024 lowercase__ : int = 128 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : float = 0 lowercase__ : float = 0 lowercase__ : bool = False lowercase__ : int = 4 lowercase__ : Optional[int] = 128 lowercase__ : "StructureModuleConfig" = None def snake_case__ ( self : Tuple ) -> str: """simple docstring""" if self.structure_module is None: __lowercase = StructureModuleConfig() elif isinstance(self.structure_module , lowercase ): __lowercase = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(F"`max_recycles` should be positive, got {self.max_recycles}." ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( """`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got""" F" {self.sequence_state_dim} and {self.sequence_state_dim}." ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( """`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got""" F" {self.pairwise_state_dim} and {self.pairwise_state_dim}." ) __lowercase = self.sequence_state_dim // self.sequence_head_width __lowercase = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( """`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got""" F" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}." ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( """`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got""" F" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}." ) if self.pairwise_state_dim % 2 != 0: raise ValueError(F"`pairwise_state_dim` should be even, got {self.pairwise_state_dim}." ) if self.dropout >= 0.4: raise ValueError(F"`dropout` should not be greater than 0.4, got {self.dropout}." ) def snake_case__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.structure_module.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 384 lowercase__ : int = 128 lowercase__ : int = 16 lowercase__ : int = 128 lowercase__ : int = 12 lowercase__ : int = 4 lowercase__ : int = 8 lowercase__ : float = 0.1 lowercase__ : int = 8 lowercase__ : int = 1 lowercase__ : int = 2 lowercase__ : int = 7 lowercase__ : int = 10 lowercase__ : float = 1E-8 lowercase__ : float = 1E5 def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" return asdict(self ) def UpperCAmelCase__ ( ) -> List[Any]: return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
634
1
def UpperCAmelCase__ ( lowercase__ ) -> list[int]: if num <= 0: raise ValueError("""Input must be a positive integer""" ) __lowercase = [True] * (num + 1) __lowercase = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , lowercase__ ): __lowercase = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
634
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Tuple = LxmertTokenizer lowercase__ : List[str] = LxmertTokenizerFast lowercase__ : Optional[Any] = True lowercase__ : List[Any] = True def snake_case__ ( self : Tuple ) -> Tuple: """simple docstring""" super().setUp() __lowercase = [ """[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def snake_case__ ( self : Optional[int] , lowercase : int ) -> List[Any]: """simple docstring""" __lowercase = """UNwant\u00E9d,running""" __lowercase = """unwanted, running""" return input_text, output_text def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class(self.vocab_file ) __lowercase = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(lowercase , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , [7, 4, 5, 10, 8, 9] ) def snake_case__ ( self : Union[str, Any] ) -> int: """simple docstring""" if not self.test_rust_tokenizer: return __lowercase = self.get_tokenizer() __lowercase = self.get_rust_tokenizer() __lowercase = """I was born in 92000, and this is falsé.""" __lowercase = tokenizer.tokenize(lowercase ) __lowercase = rust_tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) __lowercase = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = self.get_rust_tokenizer() __lowercase = tokenizer.encode(lowercase ) __lowercase = rust_tokenizer.encode(lowercase ) self.assertListEqual(lowercase , lowercase )
634
1
def UpperCAmelCase__ ( lowercase__ ) -> bool: return str(lowercase__ ) == str(lowercase__ )[::-1] def UpperCAmelCase__ ( lowercase__ ) -> int: return int(lowercase__ ) + int(str(lowercase__ )[::-1] ) def UpperCAmelCase__ ( lowercase__ = 10_000 ) -> int: __lowercase = [] for num in range(1 , lowercase__ ): __lowercase = 0 __lowercase = num while iterations < 50: __lowercase = sum_reverse(lowercase__ ) iterations += 1 if is_palindrome(lowercase__ ): break else: lychrel_nums.append(lowercase__ ) return len(lowercase__ ) if __name__ == "__main__": print(F"""{solution() = }""")
634
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> bool: __lowercase = len(lowercase__ ) __lowercase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): __lowercase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): __lowercase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: __lowercase = subset[i - 1][j] if arr[i - 1] <= j: __lowercase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
634
1
import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import TimesformerConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, TimesformerForVideoClassification, TimesformerModel, ) from transformers.models.timesformer.modeling_timesformer import TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self : Union[str, Any] , lowercase : List[Any] , lowercase : List[Any]=13 , lowercase : Tuple=10 , lowercase : List[str]=3 , lowercase : Dict=2 , lowercase : Any=2 , lowercase : Optional[int]=True , lowercase : Optional[int]=True , lowercase : Union[str, Any]=32 , lowercase : Any=5 , lowercase : Optional[Any]=4 , lowercase : Union[str, Any]=37 , lowercase : int="gelu" , lowercase : Optional[int]=0.1 , lowercase : int=0.1 , lowercase : List[str]=10 , lowercase : int=0.02 , lowercase : int="divided_space_time" , lowercase : int=None , ) -> Tuple: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = patch_size __lowercase = num_frames __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = attention_type __lowercase = initializer_range __lowercase = scope __lowercase = num_labels # in TimeSformer, the number of spatial tokens equals num_frames * num_patches per frame + 1 CLS token __lowercase = (image_size // patch_size) ** 2 __lowercase = (num_frames) * self.num_patches_per_frame + 1 def snake_case__ ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" __lowercase = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def snake_case__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = TimesformerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , attention_type=self.attention_type , ) __lowercase = self.num_labels return config def snake_case__ ( self : Dict , lowercase : str , lowercase : Tuple , lowercase : str ) -> Union[str, Any]: """simple docstring""" __lowercase = TimesformerModel(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : str , lowercase : List[str] , lowercase : List[Any] , lowercase : Optional[int] ) -> Any: """simple docstring""" __lowercase = TimesformerForVideoClassification(lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) # verify the logits shape __lowercase = torch.Size((self.batch_size, self.num_labels) ) self.parent.assertEqual(result.logits.shape , lowercase ) def snake_case__ ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : str = (TimesformerModel, TimesformerForVideoClassification) if is_torch_available() else () lowercase__ : Any = ( {"""feature-extraction""": TimesformerModel, """video-classification""": TimesformerForVideoClassification} if is_torch_available() else {} ) lowercase__ : List[str] = False lowercase__ : Any = False lowercase__ : Union[str, Any] = False lowercase__ : Optional[int] = False def snake_case__ ( self : Dict ) -> Dict: """simple docstring""" __lowercase = TimesformerModelTester(self ) __lowercase = ConfigTester( self , config_class=lowercase , has_text_modality=lowercase , hidden_size=37 ) def snake_case__ ( self : Optional[int] , lowercase : Union[str, Any] , lowercase : Union[str, Any] , lowercase : Tuple=False ) -> Optional[Any]: """simple docstring""" __lowercase = copy.deepcopy(lowercase ) if return_labels: if model_class in get_values(lowercase ): __lowercase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowercase ) return inputs_dict def snake_case__ ( self : List[Any] ) -> int: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""TimeSformer does not use inputs_embeds""" ) def snake_case__ ( self : List[str] ) -> Any: """simple docstring""" pass def snake_case__ ( self : int ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def snake_case__ ( self : str ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase ) def snake_case__ ( self : Tuple ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def snake_case__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_video_classification(*lowercase ) @slow def snake_case__ ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" for model_name in TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = TimesformerModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" if not self.has_attentions: pass else: __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = True for model_class in self.all_model_classes: __lowercase = self.model_tester.seq_length __lowercase = self.model_tester.num_frames __lowercase = True __lowercase = False __lowercase = True __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) __lowercase = outputs.attentions self.assertEqual(len(lowercase ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __lowercase = True __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) __lowercase = outputs.attentions self.assertEqual(len(lowercase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) __lowercase = len(lowercase ) # Check attention is always last and order is fine __lowercase = True __lowercase = True __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) self.assertEqual(out_len + 1 , len(lowercase ) ) __lowercase = outputs.attentions self.assertEqual(len(lowercase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) def snake_case__ ( self : Dict ) -> List[Any]: """simple docstring""" def check_hidden_states_output(lowercase : List[Any] , lowercase : Dict , lowercase : Union[str, Any] ): __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) __lowercase = outputs.hidden_states __lowercase = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(lowercase ) , lowercase ) __lowercase = self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase , lowercase , lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase , lowercase , lowercase ) def UpperCAmelCase__ ( ) -> List[Any]: __lowercase = hf_hub_download( repo_id="""hf-internal-testing/spaghetti-video""" , filename="""eating_spaghetti.npy""" , repo_type="""dataset""" ) __lowercase = np.load(lowercase__ ) return list(lowercase__ ) @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def snake_case__ ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = TimesformerForVideoClassification.from_pretrained("""facebook/timesformer-base-finetuned-k400""" ).to( lowercase ) __lowercase = self.default_image_processor __lowercase = prepare_video() __lowercase = image_processor(video[:8] , return_tensors="""pt""" ).to(lowercase ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase ) # verify the logits __lowercase = torch.Size((1, 400) ) self.assertEqual(outputs.logits.shape , lowercase ) __lowercase = torch.tensor([-0.3016, -0.7713, -0.4205] ).to(lowercase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase , atol=1E-4 ) )
634
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "hustvl/yolos-small": "https://huggingface.co/hustvl/yolos-small/resolve/main/config.json", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """yolos""" def __init__( self : Optional[int] , lowercase : Any=768 , lowercase : Tuple=12 , lowercase : Tuple=12 , lowercase : str=3_072 , lowercase : Optional[Any]="gelu" , lowercase : Union[str, Any]=0.0 , lowercase : Dict=0.0 , lowercase : Optional[int]=0.02 , lowercase : Optional[Any]=1E-1_2 , lowercase : Tuple=[512, 864] , lowercase : Optional[int]=16 , lowercase : Dict=3 , lowercase : Optional[Any]=True , lowercase : Optional[int]=100 , lowercase : Optional[int]=True , lowercase : Any=False , lowercase : Any=1 , lowercase : Any=5 , lowercase : List[str]=2 , lowercase : Union[str, Any]=5 , lowercase : str=2 , lowercase : Tuple=0.1 , **lowercase : str , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase ) __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = qkv_bias __lowercase = num_detection_tokens __lowercase = use_mid_position_embeddings __lowercase = auxiliary_loss # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = eos_coefficient class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Dict = version.parse("""1.11""" ) @property def snake_case__ ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def snake_case__ ( self : int ) -> float: """simple docstring""" return 1E-4 @property def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" return 12
634
1
import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Optional[Any] = CpmAntTokenizer lowercase__ : Tuple = False def snake_case__ ( self : Optional[int] ) -> str: """simple docstring""" super().setUp() __lowercase = [ """<d>""", """</d>""", """<s>""", """</s>""", """</_>""", """<unk>""", """<pad>""", """</n>""", """我""", """是""", """C""", """P""", """M""", """A""", """n""", """t""", ] __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) @tooslow def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" __lowercase = CpmAntTokenizer.from_pretrained("""openbmb/cpm-ant-10b""" ) __lowercase = """今天天气真好!""" __lowercase = ["""今天""", """天气""", """真""", """好""", """!"""] __lowercase = tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = """今天天气真好!""" __lowercase = [tokenizer.bos_token] + tokens __lowercase = [6, 9_802, 14_962, 2_082, 831, 244] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , lowercase ) __lowercase = tokenizer.decode(lowercase ) self.assertEqual(lowercase , lowercase )
634
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = IFImgaImgSuperResolutionPipeline lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) lowercase__ : Tuple = PipelineTesterMixin.required_optional_params - {"""latents"""} def snake_case__ ( self : Tuple ) -> Any: """simple docstring""" return self._get_superresolution_dummy_components() def snake_case__ ( self : List[str] , lowercase : Optional[int] , lowercase : Optional[Any]=0 ) -> Union[str, Any]: """simple docstring""" if str(lowercase ).startswith("""mps""" ): __lowercase = torch.manual_seed(lowercase ) else: __lowercase = torch.Generator(device=lowercase ).manual_seed(lowercase ) __lowercase = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """original_image""": original_image, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def snake_case__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1E-1 ) def snake_case__ ( self : Dict ) -> int: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" self._test_save_load_local() def snake_case__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
634
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "microsoft/markuplm-base": "https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json", "microsoft/markuplm-large": "https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json", } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Optional[int] = """markuplm""" def __init__( self : Dict , lowercase : Union[str, Any]=30_522 , lowercase : Dict=768 , lowercase : List[Any]=12 , lowercase : List[Any]=12 , lowercase : Any=3_072 , lowercase : str="gelu" , lowercase : Any=0.1 , lowercase : Tuple=0.1 , lowercase : Dict=512 , lowercase : Any=2 , lowercase : List[str]=0.02 , lowercase : str=1E-1_2 , lowercase : Optional[Any]=0 , lowercase : Tuple=0 , lowercase : Union[str, Any]=2 , lowercase : Any=256 , lowercase : Optional[int]=1_024 , lowercase : List[Any]=216 , lowercase : Dict=1_001 , lowercase : List[Any]=32 , lowercase : str=50 , lowercase : List[str]="absolute" , lowercase : Optional[int]=True , lowercase : Dict=None , **lowercase : int , ) -> Any: """simple docstring""" super().__init__( pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , **lowercase , ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = classifier_dropout # additional properties __lowercase = max_depth __lowercase = max_xpath_tag_unit_embeddings __lowercase = max_xpath_subs_unit_embeddings __lowercase = tag_pad_id __lowercase = subs_pad_id __lowercase = xpath_unit_hidden_size
634
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ = { "configuration_luke": ["LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP", "LukeConfig"], "tokenization_luke": ["LukeTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "LUKE_PRETRAINED_MODEL_ARCHIVE_LIST", "LukeForEntityClassification", "LukeForEntityPairClassification", "LukeForEntitySpanClassification", "LukeForMultipleChoice", "LukeForQuestionAnswering", "LukeForSequenceClassification", "LukeForTokenClassification", "LukeForMaskedLM", "LukeModel", "LukePreTrainedModel", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
1
from __future__ import annotations def UpperCAmelCase__ ( lowercase__ ) -> list[int]: __lowercase = [True] * limit __lowercase = False __lowercase = False __lowercase = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): __lowercase = i * 2 while index < limit: __lowercase = False __lowercase = index + i __lowercase = [2] for i in range(3 , lowercase__ , 2 ): if is_prime[i]: primes.append(lowercase__ ) return primes def UpperCAmelCase__ ( lowercase__ = 1_000_000 ) -> int: __lowercase = prime_sieve(lowercase__ ) __lowercase = 0 __lowercase = 0 for i in range(len(lowercase__ ) ): for j in range(i + length , len(lowercase__ ) ): __lowercase = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: __lowercase = j - i __lowercase = sol return largest if __name__ == "__main__": print(F"""{solution() = }""")
634
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> int: if len(lowercase__ ) != len(lowercase__ ): raise ValueError("""String lengths must match!""" ) __lowercase = 0 for chara, chara in zip(lowercase__ , lowercase__ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
634
1
from sympy import diff, lambdify, symbols from sympy.functions import * # noqa: F403 def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ = "x" , lowercase__ = 10**-10 , lowercase__ = 1 , ) -> complex: __lowercase = symbols(lowercase__ ) __lowercase = lambdify(lowercase__ , lowercase__ ) __lowercase = lambdify(lowercase__ , diff(lowercase__ , lowercase__ ) ) __lowercase = starting_point while True: if diff_function(lowercase__ ) != 0: __lowercase = prev_guess - multiplicity * func(lowercase__ ) / diff_function( lowercase__ ) else: raise ZeroDivisionError("""Could not find root""" ) from None # Precision is checked by comparing the difference of consecutive guesses if abs(next_guess - prev_guess ) < precision: return next_guess __lowercase = next_guess # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F"""The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}""") # Find root of polynomial # Find fourth Root of 5 print(F"""The root of x**4 - 5 = 0 is {newton_raphson("x**4 -5", 0.4 +5j)}""") # Find value of e print( "The root of log(y) - 1 = 0 is ", F"""{newton_raphson("log(y) - 1", 2, variable="y")}""", ) # Exponential Roots print( "The root of exp(x) - 1 = 0 is", F"""{newton_raphson("exp(x) - 1", 10, precision=0.005)}""", ) # Find root of cos(x) print(F"""The root of cos(x) = 0 is {newton_raphson("cos(x)", 0)}""")
634
from __future__ import annotations from collections.abc import Callable UpperCamelCase__ = list[list[float | int]] def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Matrix: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(size + 1 )] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for row in range(lowercase__ ): for col in range(lowercase__ ): __lowercase = matrix[row][col] __lowercase = vector[row][0] __lowercase = 0 __lowercase = 0 while row < size and col < size: # pivoting __lowercase = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase__ , lowercase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: __lowercase , __lowercase = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase__ ): __lowercase = augmented[rowa][col] / augmented[row][col] __lowercase = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase__ ): for row in range(lowercase__ ): __lowercase = augmented[row][col] / augmented[col][col] for cola in range(lowercase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase__ ) ] def UpperCAmelCase__ ( lowercase__ ) -> Callable[[int], int]: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(lowercase__ )] for _ in range(lowercase__ )] __lowercase = [[0] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for x_val, y_val in enumerate(lowercase__ ): for col in range(lowercase__ ): __lowercase = (x_val + 1) ** (size - col - 1) __lowercase = y_val __lowercase = solve(lowercase__ , lowercase__ ) def interpolated_func(lowercase__ ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase__ ) ) return interpolated_func def UpperCAmelCase__ ( lowercase__ ) -> int: return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCAmelCase__ ( lowercase__ = question_function , lowercase__ = 10 ) -> int: __lowercase = [func(lowercase__ ) for x_val in range(1 , order + 1 )] __lowercase = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] __lowercase = 0 __lowercase = 42 __lowercase = 42 for poly in polynomials: __lowercase = 1 while func(lowercase__ ) == poly(lowercase__ ): x_val += 1 ret += poly(lowercase__ ) return ret if __name__ == "__main__": print(F"""{solution() = }""")
634
1
def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> int: if exponent == 1: return base if exponent % 2 == 0: __lowercase = _modexpt(lowercase__ , exponent // 2 , lowercase__ ) % modulo_value return (x * x) % modulo_value else: return (base * _modexpt(lowercase__ , exponent - 1 , lowercase__ )) % modulo_value def UpperCAmelCase__ ( lowercase__ = 1_777 , lowercase__ = 1_855 , lowercase__ = 8 ) -> int: __lowercase = base for _ in range(1 , lowercase__ ): __lowercase = _modexpt(lowercase__ , lowercase__ , 10**digits ) return result if __name__ == "__main__": print(F"""{solution() = }""")
634
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ = { "configuration_ctrl": ["CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP", "CTRLConfig"], "tokenization_ctrl": ["CTRLTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "CTRLForSequenceClassification", "CTRLLMHeadModel", "CTRLModel", "CTRLPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCTRLForSequenceClassification", "TFCTRLLMHeadModel", "TFCTRLModel", "TFCTRLPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class _lowerCAmelCase : """simple docstring""" def __init__( self : Union[str, Any] , lowercase : Tuple , lowercase : int=13 , lowercase : Dict=7 , lowercase : Optional[Any]=True , lowercase : str=True , lowercase : Any=True , lowercase : List[Any]=99 , lowercase : List[str]=32 , lowercase : str=5 , lowercase : List[str]=4 , lowercase : Tuple=37 , lowercase : Union[str, Any]="gelu" , lowercase : Optional[int]=0.1 , lowercase : Tuple=0.1 , lowercase : List[str]=512 , lowercase : Union[str, Any]=16 , lowercase : str=2 , lowercase : Tuple=0.02 , lowercase : int=3 , lowercase : Tuple=4 , lowercase : Optional[Any]=None , ) -> List[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = self.vocab_size - 1 def snake_case__ ( self : Union[str, Any] ) -> int: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) __lowercase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def snake_case__ ( self : Any , lowercase : int , lowercase : Optional[Any] , lowercase : Any , lowercase : List[str] , *lowercase : Any ) -> int: """simple docstring""" __lowercase = OpenAIGPTModel(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase , token_type_ids=lowercase , head_mask=lowercase ) __lowercase = model(lowercase , token_type_ids=lowercase ) __lowercase = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : Tuple , lowercase : List[str] , lowercase : Tuple , lowercase : Any , lowercase : Optional[int] , *lowercase : str ) -> List[Any]: """simple docstring""" __lowercase = OpenAIGPTLMHeadModel(lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self : List[str] , lowercase : Union[str, Any] , lowercase : List[Any] , lowercase : List[Any] , lowercase : List[str] , *lowercase : Dict ) -> str: """simple docstring""" __lowercase = OpenAIGPTDoubleHeadsModel(lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self : List[Any] , lowercase : Dict , lowercase : Dict , lowercase : List[str] , lowercase : Any , *lowercase : List[str] ) -> Any: """simple docstring""" __lowercase = self.num_labels __lowercase = OpenAIGPTForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = model(lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask, } return config, inputs_dict @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Union[str, Any] = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) lowercase__ : Tuple = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly lowercase__ : Optional[Any] = ( { """feature-extraction""": OpenAIGPTModel, """text-classification""": OpenAIGPTForSequenceClassification, """text-generation""": OpenAIGPTLMHeadModel, """zero-shot""": OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def snake_case__ ( self : Any , lowercase : Union[str, Any] , lowercase : List[Any] , lowercase : Union[str, Any] , lowercase : List[Any] , lowercase : Tuple ) -> Tuple: """simple docstring""" if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def snake_case__ ( self : Union[str, Any] , lowercase : Tuple , lowercase : Union[str, Any] , lowercase : Dict=False ) -> int: """simple docstring""" __lowercase = super()._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": __lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=lowercase , ) __lowercase = inputs_dict["""labels"""] __lowercase = inputs_dict["""labels"""] __lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=lowercase , ) __lowercase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowercase ) return inputs_dict def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = OpenAIGPTModelTester(self ) __lowercase = ConfigTester(self , config_class=lowercase , n_embd=37 ) def snake_case__ ( self : Tuple ) -> str: """simple docstring""" self.config_tester.run_common_tests() def snake_case__ ( self : Any ) -> Any: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*lowercase ) def snake_case__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*lowercase ) def snake_case__ ( self : str ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*lowercase ) def snake_case__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*lowercase ) @slow def snake_case__ ( self : Optional[Any] ) -> List[str]: """simple docstring""" for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = OpenAIGPTModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) @require_torch class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase = OpenAIGPTLMHeadModel.from_pretrained("""openai-gpt""" ) model.to(lowercase ) __lowercase = torch.tensor([[481, 4_735, 544]] , dtype=torch.long , device=lowercase ) # the president is __lowercase = [ 481, 4_735, 544, 246, 963, 870, 762, 239, 244, 40_477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the __lowercase = model.generate(lowercase , do_sample=lowercase ) self.assertListEqual(output_ids[0].tolist() , lowercase )
634
import unittest import numpy as np def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = None , ) -> np.ndarray: __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) if shape_a[0] != shape_b[0]: __lowercase = ( """Expected the same number of rows for A and B. """ F"Instead found A of size {shape_a} and B of size {shape_b}" ) raise ValueError(lowercase__ ) if shape_b[1] != shape_c[1]: __lowercase = ( """Expected the same number of columns for B and C. """ F"Instead found B of size {shape_b} and C of size {shape_c}" ) raise ValueError(lowercase__ ) __lowercase = pseudo_inv if a_inv is None: try: __lowercase = np.linalg.inv(lowercase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Dict ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) __lowercase = schur_complement(lowercase , lowercase , lowercase ) __lowercase = np.block([[a, b], [b.T, c]] ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) self.assertAlmostEqual(lowercase , det_a * det_s ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
634
1
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Any = ["""image_processor""", """tokenizer"""] lowercase__ : int = """BlipImageProcessor""" lowercase__ : Optional[Any] = """AutoTokenizer""" def __init__( self : Tuple , lowercase : Any , lowercase : Any ) -> Any: """simple docstring""" __lowercase = False super().__init__(lowercase , lowercase ) __lowercase = self.image_processor def __call__( self : int , lowercase : ImageInput = None , lowercase : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , lowercase : bool = True , lowercase : Union[bool, str, PaddingStrategy] = False , lowercase : Union[bool, str, TruncationStrategy] = None , lowercase : Optional[int] = None , lowercase : int = 0 , lowercase : Optional[int] = None , lowercase : Optional[bool] = None , lowercase : bool = False , lowercase : bool = False , lowercase : bool = False , lowercase : bool = False , lowercase : bool = False , lowercase : bool = True , lowercase : Optional[Union[str, TensorType]] = None , **lowercase : Dict , ) -> BatchEncoding: """simple docstring""" if images is None and text is None: raise ValueError("""You have to specify either images or text.""" ) # Get only text if images is None: __lowercase = self.tokenizer __lowercase = self.tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) return text_encoding # add pixel_values __lowercase = self.image_processor(lowercase , return_tensors=lowercase ) if text is not None: __lowercase = self.tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) else: __lowercase = None if text_encoding is not None: encoding_image_processor.update(lowercase ) return encoding_image_processor def snake_case__ ( self : List[Any] , *lowercase : Optional[int] , **lowercase : List[Any] ) -> str: """simple docstring""" return self.tokenizer.batch_decode(*lowercase , **lowercase ) def snake_case__ ( self : Optional[int] , *lowercase : List[Any] , **lowercase : str ) -> str: """simple docstring""" return self.tokenizer.decode(*lowercase , **lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def snake_case__ ( self : Optional[int] ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer.model_input_names __lowercase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
634
import random def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ = False ) -> dict: __lowercase = {i: [] for i in range(lowercase__ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(lowercase__ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(lowercase__ ): for j in range(i + 1 , lowercase__ ): if random.random() < probability: graph[i].append(lowercase__ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(lowercase__ ) return graph def UpperCAmelCase__ ( lowercase__ ) -> dict: return { i: [j for j in range(lowercase__ ) if i != j] for i in range(lowercase__ ) } if __name__ == "__main__": import doctest doctest.testmod()
634
1
import argparse from collections import defaultdict import yaml UpperCamelCase__ = "docs/source/en/_toctree.yml" def UpperCAmelCase__ ( lowercase__ ) -> str: __lowercase = defaultdict(lowercase__ ) for doc in model_doc: counts[doc["local"]] += 1 __lowercase = [key for key, value in counts.items() if value > 1] __lowercase = [] for duplicate_key in duplicates: __lowercase = list({doc["""title"""] for doc in model_doc if doc["""local"""] == duplicate_key} ) if len(lowercase__ ) > 1: raise ValueError( F"{duplicate_key} is present several times in the documentation table of content at " """`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the """ """others.""" ) # Only add this once new_doc.append({"""local""": duplicate_key, """title""": titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc["""local"""]] == 1] ) # Sort return sorted(lowercase__ , key=lambda lowercase__ : s["title"].lower() ) def UpperCAmelCase__ ( lowercase__=False ) -> Any: with open(lowercase__ , encoding="""utf-8""" ) as f: __lowercase = yaml.safe_load(f.read() ) # Get to the API doc __lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 __lowercase = content[api_idx]["""sections"""] # Then to the model doc __lowercase = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 __lowercase = api_doc[model_idx]["""sections"""] __lowercase = [(idx, section) for idx, section in enumerate(lowercase__ ) if """sections""" in section] __lowercase = False for idx, modality_doc in modalities_docs: __lowercase = modality_doc["""sections"""] __lowercase = clean_model_doc_toc(lowercase__ ) if old_modality_doc != new_modality_doc: __lowercase = True if overwrite: __lowercase = new_modality_doc if diff: if overwrite: __lowercase = model_doc __lowercase = api_doc with open(lowercase__ , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(lowercase__ , allow_unicode=lowercase__ ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") UpperCamelCase__ = parser.parse_args() check_model_doc(args.fix_and_overwrite)
634
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor UpperCamelCase__ = random.Random() def UpperCAmelCase__ ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ) -> str: if rng is None: __lowercase = global_rng __lowercase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : List[Any] , lowercase : Tuple , lowercase : Union[str, Any]=7 , lowercase : List[Any]=400 , lowercase : Any=2_000 , lowercase : Optional[int]=24 , lowercase : Any=24 , lowercase : List[str]=0.0 , lowercase : Dict=16_000 , lowercase : Union[str, Any]=True , lowercase : Dict=True , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = min_seq_length __lowercase = max_seq_length __lowercase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowercase = feature_size __lowercase = num_mel_bins __lowercase = padding_value __lowercase = sampling_rate __lowercase = return_attention_mask __lowercase = do_normalize def snake_case__ ( self : Optional[int] ) -> Any: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self : List[str] , lowercase : Tuple=False , lowercase : int=False ) -> Optional[Any]: """simple docstring""" def _flatten(lowercase : Optional[Any] ): return list(itertools.chain(*lowercase ) ) if equal_length: __lowercase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowercase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowercase = [np.asarray(lowercase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = SpeechaTextFeatureExtractor if is_speech_available() else None def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = SpeechaTextFeatureExtractionTester(self ) def snake_case__ ( self : Tuple , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" self.assertTrue(np.all(np.mean(lowercase , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowercase , axis=0 ) - 1 ) < 1E-3 ) ) def snake_case__ ( self : List[Any] ) -> str: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = [np.asarray(lowercase ) for speech_input in speech_inputs] # Test feature size __lowercase = feature_extractor(lowercase , padding=lowercase , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input __lowercase = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features __lowercase = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) # Test batched __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(lowercase , lowercase ): self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __lowercase = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowercase = np.asarray(lowercase ) __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(lowercase , lowercase ): self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) def snake_case__ ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(lowercase , lowercase ): __lowercase = feature_extractor( lowercase , padding=lowercase , max_length=lowercase , return_attention_mask=lowercase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(lowercase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def snake_case__ ( self : Any ) -> Dict: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(lowercase , lowercase ): __lowercase = feature_extractor( lowercase , max_length=lowercase , padding=lowercase , return_tensors="""np""" , return_attention_mask=lowercase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(lowercase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def snake_case__ ( self : str ) -> List[Any]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""max_length""" , max_length=4 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""longest""" , max_length=4 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""longest""" , max_length=16 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" import torch __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = np.random.rand(100 , 32 ).astype(np.floataa ) __lowercase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self : Optional[int] , lowercase : Union[str, Any] ) -> int: """simple docstring""" from datasets import load_dataset __lowercase = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __lowercase = ds.sort("""id""" ).select(range(lowercase ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on __lowercase = self._load_datasamples(1 ) __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = feature_extractor(lowercase , return_tensors="""pt""" ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , lowercase , atol=1E-4 ) )
634
1
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_UpperCAmelCase ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : str = field(default="""summarization""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowercase__ : ClassVar[Features] = Features({"""text""": Value("""string""" )} ) lowercase__ : ClassVar[Features] = Features({"""summary""": Value("""string""" )} ) lowercase__ : str = "text" lowercase__ : str = "summary" @property def snake_case__ ( self : List[Any] ) -> Dict[str, str]: """simple docstring""" return {self.text_column: "text", self.summary_column: "summary"}
634
def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: __lowercase = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def UpperCAmelCase__ ( ) -> List[str]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
634
1
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def UpperCAmelCase__ ( lowercase__ ) -> Optional[int]: monkeypatch.setattr("""datasets.utils.deprecation_utils._emitted_deprecation_warnings""" , set() ) @pytest.fixture def UpperCAmelCase__ ( lowercase__ ) -> Dict: class _lowerCAmelCase : """simple docstring""" def __init__( self : Any , lowercase : str ) -> List[str]: """simple docstring""" __lowercase = metric_id class _lowerCAmelCase : """simple docstring""" lowercase__ : List[str] = [MetricMock(_UpperCAmelCase ) for metric_id in ["""accuracy""", """mse""", """precision""", """codeparrot/apps_metric"""]] def snake_case__ ( self : List[Any] ) -> List[str]: """simple docstring""" return self._metrics monkeypatch.setattr("""datasets.inspect.huggingface_hub""" , HfhMock() ) @pytest.mark.parametrize( """func, args""" , [(load_metric, ("""metrics/mse""",)), (list_metrics, ()), (inspect_metric, ("""metrics/mse""", """tmp_path"""))] ) def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) -> Optional[int]: if "tmp_path" in args: __lowercase = tuple(arg if arg != """tmp_path""" else tmp_path for arg in args ) with pytest.warns(lowercase__ , match="""https://huggingface.co/docs/evaluate""" ): func(*lowercase__ )
634
def UpperCAmelCase__ ( lowercase__ = 100 ) -> int: __lowercase = n * (n + 1) * (2 * n + 1) / 6 __lowercase = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F"""{solution() = }""")
634
1
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : List[str] , lowercase : List[str] , lowercase : Optional[int]=13 , lowercase : List[Any]=7 , lowercase : int=True , lowercase : Any=True , lowercase : int=True , lowercase : Dict=True , lowercase : Any=99 , lowercase : List[Any]=32 , lowercase : Optional[int]=5 , lowercase : Union[str, Any]=4 , lowercase : Union[str, Any]=37 , lowercase : Dict="gelu" , lowercase : Tuple=0.1 , lowercase : Optional[int]=0.1 , lowercase : Union[str, Any]=512 , lowercase : List[Any]=16 , lowercase : Optional[int]=2 , lowercase : Optional[Any]=0.02 , lowercase : int=4 , ) -> Optional[Any]: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_attention_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_choices def snake_case__ ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_attention_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowercase = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self : List[Any] ) -> Tuple: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def snake_case__ ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = True __lowercase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = True lowercase__ : Optional[int] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def snake_case__ ( self : int ) -> Tuple: """simple docstring""" __lowercase = FlaxRobertaModelTester(self ) @slow def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" for model_class_name in self.all_model_classes: __lowercase = model_class_name.from_pretrained("""roberta-base""" , from_pt=lowercase ) __lowercase = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowercase )
634
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets UpperCamelCase__ = datasets.logging.get_logger(__name__) UpperCamelCase__ = "\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = \"Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric\",\n author = \"Moosavi, Nafise Sadat and\n Strube, Michael\",\n booktitle = \"Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)\",\n month = aug,\n year = \"2016\",\n address = \"Berlin, Germany\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/P16-1060\",\n doi = \"10.18653/v1/P16-1060\",\n pages = \"632--642\",\n}\n\n" UpperCamelCase__ = "\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a \"-\"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section \"*_conll File Format\"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n" UpperCamelCase__ = "\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting 'keep_singletons=False', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the 'NP_only' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting 'min_span', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n 'mentions': mentions\n 'muc': MUC metric [Vilain et al, 1995]\n 'bcub': B-cubed [Bagga and Baldwin, 1998]\n 'ceafe': CEAFe [Luo et al., 2005]\n 'lea': LEA [Moosavi and Strube, 2016]\n 'conll_score': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric('coval')\n >>> words = ['bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -',\n ... 'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)',\n ... 'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)',\n ... 'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -',\n ... 'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -',\n ... 'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {'mentions/recall': 1.0,[...] 'conll_score': 100.0}\n" def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__=False , lowercase__=False , lowercase__=True , lowercase__=False , lowercase__="dummy_doc" ) -> str: __lowercase = {doc: key_lines} __lowercase = {doc: sys_lines} __lowercase = {} __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase , __lowercase = reader.get_doc_mentions(lowercase__ , key_doc_lines[doc] , lowercase__ ) key_singletons_num += singletons_num if NP_only or min_span: __lowercase = reader.set_annotated_parse_trees(lowercase__ , key_doc_lines[doc] , lowercase__ , lowercase__ ) __lowercase , __lowercase = reader.get_doc_mentions(lowercase__ , sys_doc_lines[doc] , lowercase__ ) sys_singletons_num += singletons_num if NP_only or min_span: __lowercase = reader.set_annotated_parse_trees(lowercase__ , key_doc_lines[doc] , lowercase__ , lowercase__ ) if remove_nested: __lowercase , __lowercase = reader.remove_nested_coref_mentions(lowercase__ , lowercase__ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters __lowercase , __lowercase = reader.remove_nested_coref_mentions(lowercase__ , lowercase__ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters __lowercase = reader.get_mention_assignments(lowercase__ , lowercase__ ) __lowercase = reader.get_mention_assignments(lowercase__ , lowercase__ ) __lowercase = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F"annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}" ) logger.info( """Number of resulting singleton clusters in the key """ F"annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}" ) if not keep_singletons: logger.info( F"{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system " """files, respectively""" ) return doc_coref_infos def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) -> Optional[Any]: __lowercase = get_coref_infos(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) __lowercase = {} __lowercase = 0 __lowercase = 0 for name, metric in metrics: __lowercase , __lowercase , __lowercase = evaluator.evaluate_documents(lowercase__ , lowercase__ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F"{name}/recall": recall, F"{name}/precision": precision, F"{name}/f1": fa} ) logger.info( name.ljust(10 ) , F"Recall: {recall * 100:.2f}" , F" Precision: {precision * 100:.2f}" , F" F1: {fa * 100:.2f}" , ) if conll_subparts_num == 3: __lowercase = (conll / 3) * 100 logger.info(F"CoNLL score: {conll:.2f}" ) output_scores.update({"""conll_score""": conll} ) return output_scores def UpperCAmelCase__ ( lowercase__ ) -> List[Any]: __lowercase = False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: __lowercase = line.split()[5] if not parse_col == "-": __lowercase = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def snake_case__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def snake_case__ ( self : Tuple , lowercase : Dict , lowercase : Optional[int] , lowercase : Dict=True , lowercase : List[str]=False , lowercase : int=False , lowercase : Dict=False ) -> str: """simple docstring""" __lowercase = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: __lowercase = util.check_gold_parse_annotation(lowercase ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" __lowercase = evaluate( key_lines=lowercase , sys_lines=lowercase , metrics=lowercase , NP_only=lowercase , remove_nested=lowercase , keep_singletons=lowercase , min_span=lowercase , ) return score
634
1
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ = logging.get_logger() def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = True ) -> Any: print(F"Converting {name}..." ) with torch.no_grad(): if hidden_sizes == 128: if name[-1] == "S": __lowercase = timm.create_model("""levit_128s""" , pretrained=lowercase__ ) else: __lowercase = timm.create_model("""levit_128""" , pretrained=lowercase__ ) if hidden_sizes == 192: __lowercase = timm.create_model("""levit_192""" , pretrained=lowercase__ ) if hidden_sizes == 256: __lowercase = timm.create_model("""levit_256""" , pretrained=lowercase__ ) if hidden_sizes == 384: __lowercase = timm.create_model("""levit_384""" , pretrained=lowercase__ ) from_model.eval() __lowercase = LevitForImageClassificationWithTeacher(lowercase__ ).eval() __lowercase = OrderedDict() __lowercase = from_model.state_dict() __lowercase = list(from_model.state_dict().keys() ) __lowercase = list(our_model.state_dict().keys() ) print(len(lowercase__ ) , len(lowercase__ ) ) for i in range(len(lowercase__ ) ): __lowercase = weights[og_keys[i]] our_model.load_state_dict(lowercase__ ) __lowercase = torch.randn((2, 3, 224, 224) ) __lowercase = from_model(lowercase__ ) __lowercase = our_model(lowercase__ ).logits assert torch.allclose(lowercase__ , lowercase__ ), "The model logits don't match the original one." __lowercase = name print(lowercase__ ) if push_to_hub: our_model.save_pretrained(save_directory / checkpoint_name ) __lowercase = LevitImageProcessor() image_processor.save_pretrained(save_directory / checkpoint_name ) print(F"Pushed {checkpoint_name}" ) def UpperCAmelCase__ ( lowercase__ , lowercase__ = None , lowercase__ = True ) -> Optional[int]: __lowercase = """imagenet-1k-id2label.json""" __lowercase = 1_000 __lowercase = (1, num_labels) __lowercase = """huggingface/label-files""" __lowercase = num_labels __lowercase = json.load(open(hf_hub_download(lowercase__ , lowercase__ , repo_type="""dataset""" ) , """r""" ) ) __lowercase = {int(lowercase__ ): v for k, v in idalabel.items()} __lowercase = idalabel __lowercase = {v: k for k, v in idalabel.items()} __lowercase = partial(lowercase__ , num_labels=lowercase__ , idalabel=lowercase__ , labelaid=lowercase__ ) __lowercase = { """levit-128S""": 128, """levit-128""": 128, """levit-192""": 192, """levit-256""": 256, """levit-384""": 384, } __lowercase = { """levit-128S""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 6, 8] , depths=[2, 3, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-128""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 8, 12] , depths=[4, 4, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-192""": ImageNetPreTrainedConfig( hidden_sizes=[192, 288, 384] , num_attention_heads=[3, 5, 6] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-256""": ImageNetPreTrainedConfig( hidden_sizes=[256, 384, 512] , num_attention_heads=[4, 6, 8] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-384""": ImageNetPreTrainedConfig( hidden_sizes=[384, 512, 768] , num_attention_heads=[6, 9, 12] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0.1 , ), } if model_name: convert_weight_and_push( names_to_hidden_sizes[model_name] , lowercase__ , names_to_config[model_name] , lowercase__ , lowercase__ ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(names_to_hidden_sizes[model_name] , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) return config, expected_shape if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default=None, type=str, help="The name of the model you wish to convert, it must be one of the supported Levit* architecture,", ) parser.add_argument( "--pytorch_dump_folder_path", default="levit-dump-folder/", type=Path, required=False, help="Path to the output PyTorch model directory.", ) parser.add_argument("--push_to_hub", action="store_true", help="Push model and image processor to the hub") parser.add_argument( "--no-push_to_hub", dest="push_to_hub", action="store_false", help="Do not push model and image processor to the hub", ) UpperCamelCase__ = parser.parse_args() UpperCamelCase__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
634
UpperCamelCase__ = { "joule": 1.0, "kilojoule": 10_00, "megajoule": 1_00_00_00, "gigajoule": 10_00_00_00_00, "wattsecond": 1.0, "watthour": 36_00, "kilowatthour": 3_60_00_00, "newtonmeter": 1.0, "calorie_nutr": 41_86.8, "kilocalorie_nutr": 4_18_68_00.00, "electronvolt": 1.6_0217_6634e-19, "britishthermalunit_it": 10_55.0_55_85, "footpound": 1.355_818, } def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: __lowercase = ( F"Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n" F"Valid values are: {', '.join(lowercase__ )}" ) raise ValueError(lowercase__ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
634
1
import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Dict = BioGptTokenizer lowercase__ : Union[str, Any] = False def snake_case__ ( self : str ) -> str: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowercase = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] __lowercase = dict(zip(lowercase , range(len(lowercase ) ) ) ) __lowercase = ["""l o 123""", """lo w 1456""", """e r</w> 1789""", """"""] __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(lowercase ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(lowercase ) ) def snake_case__ ( self : Tuple , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = """lower newer""" __lowercase = """lower newer""" return input_text, output_text def snake_case__ ( self : Union[str, Any] ) -> str: """simple docstring""" __lowercase = BioGptTokenizer(self.vocab_file , self.merges_file ) __lowercase = """lower""" __lowercase = ["""low""", """er</w>"""] __lowercase = tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = tokens + ["""<unk>"""] __lowercase = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , lowercase ) @slow def snake_case__ ( self : str ) -> str: """simple docstring""" __lowercase = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) __lowercase = tokenizer.encode("""sequence builders""" , add_special_tokens=lowercase ) __lowercase = tokenizer.encode("""multi-sequence build""" , add_special_tokens=lowercase ) __lowercase = tokenizer.build_inputs_with_special_tokens(lowercase ) __lowercase = tokenizer.build_inputs_with_special_tokens(lowercase , lowercase ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
634
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_UpperCAmelCase ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : str = field(default="""summarization""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowercase__ : ClassVar[Features] = Features({"""text""": Value("""string""" )} ) lowercase__ : ClassVar[Features] = Features({"""summary""": Value("""string""" )} ) lowercase__ : str = "text" lowercase__ : str = "summary" @property def snake_case__ ( self : List[Any] ) -> Dict[str, str]: """simple docstring""" return {self.text_column: "text", self.summary_column: "summary"}
634
1
from __future__ import annotations from collections import namedtuple def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> tuple: __lowercase = namedtuple("""result""" , """name value""" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("""Only one argument must be 0""" ) elif power < 0: raise ValueError( """Power cannot be negative in any electrical/electronics system""" ) elif voltage == 0: return result("""voltage""" , power / current ) elif current == 0: return result("""current""" , power / voltage ) elif power == 0: return result("""power""" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
634
def UpperCAmelCase__ ( lowercase__ ) -> Optional[int]: __lowercase = len(lowercase__ ) __lowercase = sum(lowercase__ ) __lowercase = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): __lowercase = True for i in range(1 , s + 1 ): __lowercase = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): __lowercase = dp[i][j - 1] if arr[i - 1] <= j: __lowercase = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: __lowercase = s - 2 * j break return diff
634
1
import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: UpperCamelCase__ = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : List[Any] , lowercase : List[Any] , lowercase : Any=7 , lowercase : Tuple=3 , lowercase : Optional[int]=18 , lowercase : List[str]=30 , lowercase : Union[str, Any]=400 , lowercase : Optional[int]=None , lowercase : List[Any]=True , lowercase : Any=True , lowercase : Optional[Any]=None , ) -> Union[str, Any]: """simple docstring""" __lowercase = size if size is not None else {"""height""": 20, """width""": 20} __lowercase = parent __lowercase = batch_size __lowercase = num_channels __lowercase = image_size __lowercase = min_resolution __lowercase = max_resolution __lowercase = size __lowercase = do_normalize __lowercase = do_convert_rgb __lowercase = [512, 1_024, 2_048, 4_096] __lowercase = patch_size if patch_size is not None else {"""height""": 16, """width""": 16} def snake_case__ ( self : int ) -> Union[str, Any]: """simple docstring""" return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def snake_case__ ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = """https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg""" __lowercase = Image.open(requests.get(lowercase , stream=lowercase ).raw ).convert("""RGB""" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : str = PixaStructImageProcessor if is_vision_available() else None def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase = PixaStructImageProcessingTester(self ) @property def snake_case__ ( self : int ) -> Tuple: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase , """do_normalize""" ) ) self.assertTrue(hasattr(lowercase , """do_convert_rgb""" ) ) def snake_case__ ( self : Any ) -> str: """simple docstring""" __lowercase = self.image_processor_tester.prepare_dummy_image() __lowercase = self.image_processing_class(**self.image_processor_dict ) __lowercase = 2_048 __lowercase = image_processor(lowercase , return_tensors="""pt""" , max_patches=lowercase ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() , torch.tensor(0.0606 ) , atol=1E-3 , rtol=1E-3 ) ) def snake_case__ ( self : Any ) -> str: """simple docstring""" __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , Image.Image ) # Test not batched input __lowercase = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input __lowercase = image_processor( image_inputs[0] , return_tensors="""pt""" , max_patches=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __lowercase = image_processor( lowercase , return_tensors="""pt""" , max_patches=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def snake_case__ ( self : Tuple ) -> Dict: """simple docstring""" __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , Image.Image ) # Test not batched input __lowercase = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 __lowercase = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(lowercase ): __lowercase = image_processor( image_inputs[0] , return_tensors="""pt""" , max_patches=lowercase ).flattened_patches __lowercase = """Hello""" __lowercase = image_processor( image_inputs[0] , return_tensors="""pt""" , max_patches=lowercase , header_text=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __lowercase = image_processor( lowercase , return_tensors="""pt""" , max_patches=lowercase , header_text=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowercase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , numpify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , np.ndarray ) __lowercase = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input __lowercase = image_processor( image_inputs[0] , return_tensors="""pt""" , max_patches=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __lowercase = image_processor( lowercase , return_tensors="""pt""" , max_patches=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowercase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , torchify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , torch.Tensor ) # Test not batched input __lowercase = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input __lowercase = image_processor( image_inputs[0] , return_tensors="""pt""" , max_patches=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __lowercase = image_processor( lowercase , return_tensors="""pt""" , max_patches=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason="""`Pix2StructImageProcessor` requires `torch>=1.11.0`.""" , ) @require_torch @require_vision class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Optional[int] = PixaStructImageProcessor if is_vision_available() else None def snake_case__ ( self : str ) -> int: """simple docstring""" __lowercase = PixaStructImageProcessingTester(self , num_channels=4 ) __lowercase = 3 @property def snake_case__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase , """do_normalize""" ) ) self.assertTrue(hasattr(lowercase , """do_convert_rgb""" ) ) def snake_case__ ( self : int ) -> Dict: """simple docstring""" __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , Image.Image ) # Test not batched input __lowercase = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input __lowercase = image_processor( image_inputs[0] , return_tensors="""pt""" , max_patches=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __lowercase = image_processor( lowercase , return_tensors="""pt""" , max_patches=lowercase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , )
634
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class _lowerCAmelCase : """simple docstring""" def __init__( self : List[Any] , lowercase : str , lowercase : Union[str, Any]=13 , lowercase : Tuple=32 , lowercase : Optional[Any]=2 , lowercase : Tuple=3 , lowercase : Tuple=16 , lowercase : Tuple=[1, 2, 1] , lowercase : Optional[Any]=[2, 2, 4] , lowercase : Dict=2 , lowercase : Optional[int]=2.0 , lowercase : List[Any]=True , lowercase : str=0.0 , lowercase : Any=0.0 , lowercase : Optional[int]=0.1 , lowercase : int="gelu" , lowercase : Tuple=False , lowercase : Optional[Any]=True , lowercase : int=0.02 , lowercase : Union[str, Any]=1E-5 , lowercase : Dict=True , lowercase : Any=None , lowercase : str=True , lowercase : str=10 , lowercase : Dict=8 , lowercase : int=["stage1", "stage2", "stage3"] , lowercase : Optional[int]=[1, 2, 3] , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def snake_case__ ( self : List[str] ) -> int: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def snake_case__ ( self : Any , lowercase : List[Any] , lowercase : Optional[int] , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = MaskFormerSwinModel(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def snake_case__ ( self : Any , lowercase : Tuple , lowercase : Any , lowercase : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(lowercase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=lowercase ) def snake_case__ ( self : int ) -> Any: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Optional[int] = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) lowercase__ : List[str] = {"""feature-extraction""": MaskFormerSwinModel} if is_torch_available() else {} lowercase__ : List[str] = False lowercase__ : int = False lowercase__ : int = False lowercase__ : Tuple = False lowercase__ : Optional[Any] = False def snake_case__ ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=lowercase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" pass def snake_case__ ( self : Tuple ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" return def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def snake_case__ ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowercase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def snake_case__ ( self : int ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def snake_case__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" pass def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def snake_case__ ( self : Any ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def snake_case__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" pass def snake_case__ ( self : Tuple , lowercase : Tuple , lowercase : Optional[int] , lowercase : Optional[int] , lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowercase ) , lowercase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def snake_case__ ( self : int ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , lowercase ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def snake_case__ ( self : Any ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def snake_case__ ( self : List[str] ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" pass def snake_case__ ( self : Tuple ) -> List[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(lowercase : Optional[int] ): __lowercase = 0 return t def check_equivalence(lowercase : Optional[int] , lowercase : str , lowercase : str , lowercase : Tuple={} ): with torch.no_grad(): __lowercase = model(**lowercase , return_dict=lowercase , **lowercase ) __lowercase = model(**lowercase , return_dict=lowercase , **lowercase ).to_tuple() def recursive_check(lowercase : int , lowercase : Optional[Any] ): if isinstance(lowercase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(lowercase , lowercase ): recursive_check(lowercase , lowercase ) elif isinstance(lowercase , lowercase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(lowercase , lowercase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(lowercase ) , set_nan_tensor_to_zero(lowercase ) , atol=1E-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:" F" {torch.isnan(lowercase ).any()} and `inf`: {torch.isinf(lowercase )}. Dict has" F" `nan`: {torch.isnan(lowercase ).any()} and `inf`: {torch.isinf(lowercase )}." ) , ) recursive_check(lowercase , lowercase ) for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() __lowercase = self._prepare_for_class(lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) check_equivalence(lowercase , lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) check_equivalence(lowercase , lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) check_equivalence(lowercase , lowercase , lowercase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) check_equivalence(lowercase , lowercase , lowercase , {"""output_hidden_states""": True} ) @require_torch class _lowerCAmelCase ( unittest.TestCase , _UpperCAmelCase ): """simple docstring""" lowercase__ : List[str] = (MaskFormerSwinBackbone,) if is_torch_available() else () lowercase__ : Any = MaskFormerSwinConfig def snake_case__ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def snake_case__ ( self : Any ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(lowercase ) backbone.to(lowercase ) backbone.eval() __lowercase = backbone(**lowercase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , lowercase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**lowercase , output_hidden_states=lowercase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**lowercase , output_attentions=lowercase ) self.assertIsNotNone(outputs.attentions )
634
1
import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = MgpstrTokenizer lowercase__ : Tuple = False lowercase__ : Optional[Any] = {} lowercase__ : Tuple = False def snake_case__ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" super().setUp() # fmt: off __lowercase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on __lowercase = dict(zip(lowercase , range(len(lowercase ) ) ) ) __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowercase ) + """\n""" ) def snake_case__ ( self : Tuple , **lowercase : Optional[int] ) -> Any: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **lowercase ) def snake_case__ ( self : str , lowercase : Optional[Any] ) -> str: """simple docstring""" __lowercase = """tester""" __lowercase = """tester""" return input_text, output_text @unittest.skip("""MGP-STR always lower cases letters.""" ) def snake_case__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" pass def snake_case__ ( self : str ) -> int: """simple docstring""" __lowercase = self.get_tokenizers(do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = """[SPECIAL_TOKEN]""" tokenizer.add_special_tokens({"""cls_token""": special_token} ) __lowercase = tokenizer.encode([special_token] , add_special_tokens=lowercase ) self.assertEqual(len(lowercase ) , 1 ) __lowercase = tokenizer.decode(lowercase , skip_special_tokens=lowercase ) self.assertTrue(special_token not in decoded ) def snake_case__ ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase , __lowercase = self.get_input_output_texts(lowercase ) __lowercase = tokenizer.tokenize(lowercase ) __lowercase = tokenizer.convert_tokens_to_ids(lowercase ) __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = tokenizer.convert_ids_to_tokens(lowercase ) self.assertNotEqual(len(lowercase ) , 0 ) __lowercase = tokenizer.decode(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(text_a.replace(""" """ , """""" ) , lowercase ) @unittest.skip("""MGP-STR tokenizer only handles one sequence.""" ) def snake_case__ ( self : Union[str, Any] ) -> str: """simple docstring""" pass @unittest.skip("""inputs cannot be pretokenized in MgpstrTokenizer""" ) def snake_case__ ( self : Dict ) -> Dict: """simple docstring""" pass
634
import importlib import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Union import torch from ..utils import BaseOutput UpperCamelCase__ = "scheduler_config.json" class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : List[Any] = 1 lowercase__ : Tuple = 2 lowercase__ : Union[str, Any] = 3 lowercase__ : Union[str, Any] = 4 lowercase__ : str = 5 lowercase__ : Any = 6 lowercase__ : Any = 7 lowercase__ : List[str] = 8 lowercase__ : Union[str, Any] = 9 lowercase__ : int = 10 lowercase__ : List[str] = 11 lowercase__ : List[Any] = 12 lowercase__ : str = 13 lowercase__ : Optional[int] = 14 @dataclass class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : torch.FloatTensor class _lowerCAmelCase : """simple docstring""" lowercase__ : Optional[int] = SCHEDULER_CONFIG_NAME lowercase__ : int = [] lowercase__ : Dict = True @classmethod def snake_case__ ( cls : str , lowercase : Dict[str, Any] = None , lowercase : Optional[str] = None , lowercase : Any=False , **lowercase : List[str] , ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase , __lowercase = cls.load_config( pretrained_model_name_or_path=lowercase , subfolder=lowercase , return_unused_kwargs=lowercase , return_commit_hash=lowercase , **lowercase , ) return cls.from_config(lowercase , return_unused_kwargs=lowercase , **lowercase ) def snake_case__ ( self : Dict , lowercase : Union[str, os.PathLike] , lowercase : bool = False , **lowercase : List[str] ) -> Optional[Any]: """simple docstring""" self.save_config(save_directory=lowercase , push_to_hub=lowercase , **lowercase ) @property def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" return self._get_compatibles() @classmethod def snake_case__ ( cls : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = list(set([cls.__name__] + cls._compatibles ) ) __lowercase = importlib.import_module(__name__.split(""".""" )[0] ) __lowercase = [ getattr(lowercase , lowercase ) for c in compatible_classes_str if hasattr(lowercase , lowercase ) ] return compatible_classes
634
1
def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> List[str]: if n == 0: return 1 elif n % 2 == 1: return (binary_exponentiation(lowercase__ , n - 1 , lowercase__ ) * a) % mod else: __lowercase = binary_exponentiation(lowercase__ , n / 2 , lowercase__ ) return (b * b) % mod # a prime number UpperCamelCase__ = 7_01 UpperCamelCase__ = 10_00_00_00_00 UpperCamelCase__ = 10 # using binary exponentiation function, O(log(p)): print((a / b) % p == (a * binary_exponentiation(b, p - 2, p)) % p) print((a / b) % p == (a * b ** (p - 2)) % p)
634
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "asapp/sew-tiny-100k": "https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json", # See all SEW models at https://huggingface.co/models?filter=sew } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """sew""" def __init__( self : List[Any] , lowercase : int=32 , lowercase : List[str]=768 , lowercase : Dict=12 , lowercase : str=12 , lowercase : str=3_072 , lowercase : Optional[int]=2 , lowercase : List[str]="gelu" , lowercase : List[str]=0.1 , lowercase : Tuple=0.1 , lowercase : Dict=0.1 , lowercase : Any=0.0 , lowercase : Dict=0.1 , lowercase : Optional[int]=0.1 , lowercase : List[str]=0.02 , lowercase : Dict=1E-5 , lowercase : Tuple="group" , lowercase : int="gelu" , lowercase : Any=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , lowercase : Optional[int]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowercase : Any=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowercase : List[str]=False , lowercase : Tuple=128 , lowercase : int=16 , lowercase : Union[str, Any]=True , lowercase : List[str]=0.05 , lowercase : Optional[int]=10 , lowercase : Any=2 , lowercase : Optional[Any]=0.0 , lowercase : Optional[Any]=10 , lowercase : int=0 , lowercase : Optional[int]="mean" , lowercase : List[Any]=False , lowercase : str=False , lowercase : int=256 , lowercase : str=0 , lowercase : List[Any]=1 , lowercase : List[Any]=2 , **lowercase : List[Any] , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase , pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase ) __lowercase = hidden_size __lowercase = feat_extract_norm __lowercase = feat_extract_activation __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = conv_bias __lowercase = num_conv_pos_embeddings __lowercase = num_conv_pos_embedding_groups __lowercase = len(self.conv_dim ) __lowercase = num_hidden_layers __lowercase = intermediate_size __lowercase = squeeze_factor __lowercase = hidden_act __lowercase = num_attention_heads __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = feat_proj_dropout __lowercase = final_dropout __lowercase = layerdrop __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" F"but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)" F"= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowercase = apply_spec_augment __lowercase = mask_time_prob __lowercase = mask_time_length __lowercase = mask_time_min_masks __lowercase = mask_feature_prob __lowercase = mask_feature_length __lowercase = mask_feature_min_masks # ctc loss __lowercase = ctc_loss_reduction __lowercase = ctc_zero_infinity # sequence classification __lowercase = use_weighted_layer_sum __lowercase = classifier_proj_size @property def snake_case__ ( self : Dict ) -> str: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
634
1
import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "facebook/encodec_24khz": "https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json", "facebook/encodec_48khz": "https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json", } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : int = """encodec""" def __init__( self : str , lowercase : str=[1.5, 3.0, 6.0, 12.0, 24.0] , lowercase : int=24_000 , lowercase : str=1 , lowercase : Optional[Any]=False , lowercase : Tuple=None , lowercase : List[str]=None , lowercase : Any=128 , lowercase : Tuple=32 , lowercase : Optional[Any]=1 , lowercase : Tuple=[8, 5, 4, 2] , lowercase : str="weight_norm" , lowercase : Union[str, Any]=7 , lowercase : Any=7 , lowercase : List[Any]=3 , lowercase : Optional[Any]=2 , lowercase : Dict=True , lowercase : Union[str, Any]="reflect" , lowercase : str=2 , lowercase : Union[str, Any]=2 , lowercase : Optional[Any]=1.0 , lowercase : Optional[int]=1_024 , lowercase : Tuple=None , lowercase : List[Any]=True , **lowercase : str , ) -> Optional[int]: """simple docstring""" __lowercase = target_bandwidths __lowercase = sampling_rate __lowercase = audio_channels __lowercase = normalize __lowercase = chunk_length_s __lowercase = overlap __lowercase = hidden_size __lowercase = num_filters __lowercase = num_residual_layers __lowercase = upsampling_ratios __lowercase = norm_type __lowercase = kernel_size __lowercase = last_kernel_size __lowercase = residual_kernel_size __lowercase = dilation_growth_rate __lowercase = use_causal_conv __lowercase = pad_mode __lowercase = compress __lowercase = num_lstm_layers __lowercase = trim_right_ratio __lowercase = codebook_size __lowercase = codebook_dim if codebook_dim is not None else hidden_size __lowercase = use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( F"self.norm_type must be one of `\"weight_norm\"`, `\"time_group_norm\"`), got {self.norm_type}" ) super().__init__(**lowercase ) @property def snake_case__ ( self : Tuple ) -> Optional[int]: """simple docstring""" if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def snake_case__ ( self : int ) -> Optional[int]: """simple docstring""" if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def snake_case__ ( self : List[str] ) -> int: """simple docstring""" __lowercase = np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" return int(1_000 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
634
import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizerOutput from transformers.testing_utils import require_phonemizer from ...test_tokenization_common import TokenizerTesterMixin @require_phonemizer class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = WavaVecaPhonemeCTCTokenizer lowercase__ : Optional[int] = False def snake_case__ ( self : str ) -> int: """simple docstring""" super().setUp() __lowercase = ( """<s> <pad> </s> <unk> n s t ə l a i k d m ɛ ɾ e ɪ p o ɐ z ð f j v b ɹ ʁ ʊ iː r w ʌ u ɡ æ aɪ ʃ h ɔ ɑː """ """ŋ ɚ eɪ β uː y ɑ̃ oʊ ᵻ eː θ aʊ ts oː ɔ̃ ɣ ɜ ɑ dʒ əl x ɜː ç ʒ tʃ ɔː ɑːɹ ɛ̃ ʎ ɔːɹ ʋ aː ɕ œ ø oːɹ ɲ yː """ """ʔ iə i5 s. tɕ ?? nʲ ɛː œ̃ ɭ ɔø ʑ tʲ ɨ ɛɹ ts. rʲ ɪɹ ɭʲ i.5 ɔɪ q sʲ u5 ʊɹ iɜ a5 iɛ5 øː ʕ ja əɜ th ɑ5 """ """oɪ dʲ ə5 tɕh ts.h mʲ ɯ dʑ vʲ e̞ tʃʲ ei5 o5 onɡ5 ɑu5 iɑ5 ai5 aɪɚ kh ə1 ʐ i2 ʉ ħ t[ aɪə ʲ ju ə2 u2 oɜ """ """pː iɛɜ ou5 y5 uɜ tː uo5 d[ uoɜ tsh ɑɜ ɵ i̪5 uei5 ɟ aɜ ɑɨ i.ɜ eʊ o2 ɐ̃ ä pʲ kʲ n̩ ɒ ph ɑu2 uɨ əɪ ɫ ɬ """ """yɜ bʲ ɑ2 s̪ aiɜ χ ɐ̃ʊ̃ 1 ə4 yæɜ a2 ɨː t̪ iouɜ ũ onɡɜ aɨ iɛ2 ɔɨ ɑuɜ o̞ ei2 iou2 c kː y2 ɖ oe dˤ yɛɜ """ """əʊ S ɡʲ onɡ2 u\" eiɜ ʈ ɯᵝ iou5 dZ r̝̊ i.2 tS s^ ʝ yə5 iɑɜ uə5 pf ɨu iɑ2 ou2 ər2 fʲ ai2 r̝ uəɜ ɳ əɨ """ """ua5 uɪ ɽ bː yu5 uo2 yɛ5 l̩ ɻ ərɜ ʂ i̪2 ouɜ uaɜ a. a.ː yæ5 dː r̩ ee ɪu ər5 i̪ ɜ æi u: i.ː t^ o1 ɪ^ """ """ai ueiɜ æː ɛɪ eə i. ɴ ie ua2 ɑ1 o4 tʃː o: ɑ: u1 N i̪1 au yæ2 u. qː yəɜ y: kʰ tʃʰ iʊ sx õ uo tʰ """ """uai5 bʰ u.ː uə2 ʊə d^ s̪ː yiɜ dʰ r. oe: i1 ɟː yu2 nʲʲ i̪4 uei2 tsʲ ɸ ĩ ɑ4 t̪ː eɑ u4 e: tsː ʈʰ ɡʰ """ """ɯɯ dʒʲ ʂʲ X ɵː uaiɜ tɕʲ ã t^ː ẽː yɛ2 cː i.1 ɛʊ dˤdˤ dʒː i4 ɡː yi ɕʲ ɟʰ pʰ dʑʲ yuɜ ua1 ua4 æiː ɐɐ """ """ui iou1 ʊː a1 iou4 cʰ iɛ1 yə2 ɖʰ ẽ ʒʲ ää ər4 iːː ɪː iɑ1 ər1 œː øi ɪuː cʰcʰ əː1 iː1 ũ kʰː o̞o̞ xʲ """ """ou1 iɛ4 e̞e̞ y1 dzː dʲʲ dʰː ɯᵝɯᵝ lː uo1 i.4 i: yɛ5ʲ a4""" ).split(""" """ ) __lowercase = dict(zip(lowercase , range(len(lowercase ) ) ) ) __lowercase = {"""pad_token""": """<pad>""", """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>"""} __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowercase ) + """\n""" ) def snake_case__ ( self : List[Any] , lowercase : Optional[Any] , lowercase : List[str]=False , lowercase : List[str]=20 , lowercase : str=5 ) -> Tuple[str, list]: """simple docstring""" __lowercase = [(i, tokenizer.decode([i] , clean_up_tokenization_spaces=lowercase )) for i in range(len(lowercase ) )] __lowercase = list(filter(lambda lowercase : [t[0]] == tokenizer.encode(t[1] , do_phonemize=lowercase ) , lowercase ) ) if max_length is not None and len(lowercase ) > max_length: __lowercase = toks[:max_length] if min_length is not None and len(lowercase ) < min_length and len(lowercase ) > 0: while len(lowercase ) < min_length: __lowercase = toks + toks # toks_str = [t[1] for t in toks] __lowercase = [t[0] for t in toks] # Ensure consistency __lowercase = tokenizer.decode(lowercase , clean_up_tokenization_spaces=lowercase ) if " " not in output_txt and len(lowercase ) > 1: __lowercase = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=lowercase ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=lowercase ) ) if with_prefix_space: __lowercase = """ """ + output_txt __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) return output_txt, output_ids def snake_case__ ( self : Tuple , **lowercase : int ) -> Union[str, Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return WavaVecaPhonemeCTCTokenizer.from_pretrained(self.tmpdirname , **lowercase ) def snake_case__ ( self : Any ) -> List[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) # check adding a single token tokenizer.add_tokens("""xxx""" ) __lowercase = tokenizer("""m xxx ɪ""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 392, 17] ) # xxx should be last token tokenizer.add_tokens(["""aaa""", """bbb""", """ccc"""] ) __lowercase = tokenizer("""m aaa ɪ ccc""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 393, 17, 395] ) # aaa and ccc should be after xxx and 2 after aaa __lowercase = tokenizer("""maɪ c""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [3, 200] ) # mai should be <unk> (=3) def snake_case__ ( self : int ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) def snake_case__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98], [24, 22, 5, 24, 22, 5, 77], ] __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> Optional[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ | h aʊ | ɑːɹ | j uː |""" ) def snake_case__ ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : List[Any] ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, tokenizer.word_delimiter_token_id, 15, 8, tokenizer.word_delimiter_token_id, 98], [tokenizer.word_delimiter_token_id, 24, 22, tokenizer.word_delimiter_token_id, 5, 24, 22, 5, 77], ] # fmt: on # decode with word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) # decode with no word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] , filter_word_delimiter_token=lowercase ) __lowercase = tokenizer.batch_decode(lowercase , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ | ɾ l | ɭʲ""", """| j ð | s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : Optional[int] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(""" """.join([p.strip() for p in phonemes.split(""" |""" )] ).strip() , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token=lowercase ) __lowercase = """Hello how are you""" __lowercase = tokenizer(lowercase , phonemizer_lang="""en-us""" ).input_ids __lowercase = tokenizer(lowercase , phonemizer_lang="""fr-fr""" ).input_ids self.assertNotEqual(lowercase , lowercase ) __lowercase = tokenizer.decode(lowercase ) __lowercase = tokenizer.decode(lowercase ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) self.assertEqual(lowercase , """ɛ l o h aʊ a ʁ j u""" ) def snake_case__ ( self : int ) -> int: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how Are you""" __lowercase = """hello how are you""" __lowercase = tokenizer(lowercase ).input_ids __lowercase = tokenizer(lowercase ).input_ids self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : int ) -> Union[str, Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) tokenizer.add_tokens(["""!""", """?"""] ) tokenizer.add_special_tokens({"""cls_token""": """$$$"""} ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98, 392, 392, 393, 392, 392, 393, 394, 394], [24, 22, 5, 24, 22, 5, 77, tokenizer.pad_token_id, 394, 394], ] # fmt: on __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ!?!? $$$""", """j ð s j ð s oːɹ $$$"""] ) @staticmethod def snake_case__ ( lowercase : List[str] , lowercase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = [d[key] for d in offsets] return retrieved_list def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off # ksssɾɾ|ɾɾ<pad>ɾɾ|<pad>ɾlll|ɭʲ -> k s ɾ ɾ | ɾ l | ɭʲ" __lowercase = [11, 5, 5, 5, 15, 15, tokenizer.pad_token_id, 15, 15, tokenizer.word_delimiter_token_id, tokenizer.pad_token_id, 15, 8, 8, 8, tokenizer.word_delimiter_token_id, 98] # fmt: on __lowercase = tokenizer.decode(lowercase , output_char_offsets=lowercase , filter_word_delimiter_token=lowercase ) # check Wav2Vec2CTCTokenizerOutput keys for char self.assertEqual(len(outputs.keys() ) , 2 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""char_offsets""" in outputs ) self.assertTrue(isinstance(lowercase , lowercase ) ) # check that order of chars is correct and identical for both outputs self.assertEqual(""" """.join(self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) ) , outputs.text ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) , ["""k""", """s""", """ɾ""", """ɾ""", """|""", """ɾ""", """l""", """|""", """ɭʲ"""] ) # check that offsets are actually correct for char # 0-1 is 11, 1-4 is 5, 4-6 is first 15, 6-7 is <pad> (thus not shown), 7-9 is second 15, 9-10 is word_delimiter_token, # 10-11 is <pad> (thus not shown), 11-12 is third 15, 12-15 is 8, 15-16 is word_delimiter_token, 16-17 is 98 self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """start_offset""" ) , [0, 1, 4, 7, 9, 11, 12, 15, 16] ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """end_offset""" ) , [1, 4, 6, 9, 10, 12, 15, 16, 17] ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) def check_list_tuples_equal(lowercase : List[str] , lowercase : Dict ): self.assertTrue(isinstance(lowercase , lowercase ) ) self.assertTrue(isinstance(outputs_list[0] , lowercase ) ) # transform list to ModelOutput __lowercase = WavaVecaPhonemeCTCTokenizerOutput( {k: [d[k] for d in outputs_list] for k in outputs_list[0]} ) self.assertListEqual(outputs_batch["""text"""] , outputs_batch_a["""text"""] ) def recursive_check(lowercase : List[Any] , lowercase : Optional[int] ): if isinstance(lowercase , lowercase ): [recursive_check(lowercase , lowercase ) for la, la in zip(lowercase , lowercase )] self.assertEqual(lowercase , lowercase ) if "char_offsets" in outputs_batch: recursive_check(outputs_batch["""char_offsets"""] , outputs_batch_a["""char_offsets"""] ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 4, 8, 98, 32, 32, 32, 32, 4, 33, tokenizer.word_delimiter_token_id, 32, 32, 33, 34, 34], [24, 22, 5, tokenizer.word_delimiter_token_id, tokenizer.word_delimiter_token_id, 24, 22, 22, 22, 4, 5, 77, tokenizer.pad_token_id, 22, 22, 4, 34, 34, 34, 34], ] # fmt: on # We assume that `decode` works as expected. All we will check now is # the output type is correct and the output is identical to `decode` # char __lowercase = tokenizer.batch_decode(lowercase , output_char_offsets=lowercase ) __lowercase = [tokenizer.decode(lowercase , output_char_offsets=lowercase ) for ids in sample_ids] check_list_tuples_equal(lowercase , lowercase ) @unittest.skip("""Wav2Vec2PhonemeTokenizer always lower cases letters to correctly map to phonemes""" ) def snake_case__ ( self : List[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeTokenizer always puts spaces between phonemes""" ) def snake_case__ ( self : List[str] ) -> List[str]: """simple docstring""" pass @unittest.skip("""encodes to text to ids, but decodes ids to phonemes -> not possible to have internal consistency""" ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeModel has no max model length => no testing""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass def snake_case__ ( self : Any ) -> int: """simple docstring""" __lowercase = self.get_tokenizers(do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) __lowercase = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] __lowercase = tokenizer.add_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size + len(lowercase ) ) __lowercase = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) __lowercase = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} __lowercase = tokenizer.add_special_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size_a + len(lowercase ) ) __lowercase = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" pass def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizers(fast=lowercase , do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = ["""ð""", """ɪ""", """s""", """ɪ""", """z""", """ɐ""", """t""", """ɛ""", """k""", """s""", """t"""] __lowercase = tokenizer.convert_tokens_to_string(lowercase ) self.assertIsInstance(output["""text"""] , lowercase )
634
1
from random import randint from tempfile import TemporaryFile import numpy as np def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> Dict: __lowercase = 0 if start < end: __lowercase = randint(lowercase__ , lowercase__ ) __lowercase = a[end] __lowercase = a[pivot] __lowercase = temp __lowercase , __lowercase = _in_place_partition(lowercase__ , lowercase__ , lowercase__ ) count += _in_place_quick_sort(lowercase__ , lowercase__ , p - 1 ) count += _in_place_quick_sort(lowercase__ , p + 1 , lowercase__ ) return count def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> Optional[int]: __lowercase = 0 __lowercase = randint(lowercase__ , lowercase__ ) __lowercase = a[end] __lowercase = a[pivot] __lowercase = temp __lowercase = start - 1 for index in range(lowercase__ , lowercase__ ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value __lowercase = new_pivot_index + 1 __lowercase = a[new_pivot_index] __lowercase = a[index] __lowercase = temp __lowercase = a[new_pivot_index + 1] __lowercase = a[end] __lowercase = temp return new_pivot_index + 1, count UpperCamelCase__ = TemporaryFile() UpperCamelCase__ = 1_00 # 1000 elements are to be sorted UpperCamelCase__ , UpperCamelCase__ = 0, 1 # mean and standard deviation UpperCamelCase__ = np.random.normal(mu, sigma, p) np.save(outfile, X) print("The array is") print(X) outfile.seek(0) # using the same array UpperCamelCase__ = np.load(outfile) UpperCamelCase__ = len(M) - 1 UpperCamelCase__ = _in_place_quick_sort(M, 0, r) print( "No of Comparisons for 100 elements selected from a standard normal distribution" "is :" ) print(z)
634
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) # TODO Update this UpperCamelCase__ = { "facebook/esm-1b": "https://huggingface.co/facebook/esm-1b/resolve/main/config.json", # See all ESM models at https://huggingface.co/models?filter=esm } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """esm""" def __init__( self : Any , lowercase : Optional[Any]=None , lowercase : Optional[int]=None , lowercase : List[Any]=None , lowercase : Optional[int]=768 , lowercase : str=12 , lowercase : Union[str, Any]=12 , lowercase : Dict=3_072 , lowercase : Optional[int]=0.1 , lowercase : str=0.1 , lowercase : Dict=1_026 , lowercase : Tuple=0.02 , lowercase : str=1E-1_2 , lowercase : Dict="absolute" , lowercase : Optional[Any]=True , lowercase : int=None , lowercase : int=False , lowercase : List[str]=False , lowercase : Tuple=None , lowercase : Tuple=None , **lowercase : Union[str, Any] , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase , mask_token_id=lowercase , **lowercase ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = emb_layer_norm_before __lowercase = token_dropout __lowercase = is_folding_model if is_folding_model: if esmfold_config is None: logger.info("""No esmfold_config supplied for folding model, using default values.""" ) __lowercase = EsmFoldConfig() elif isinstance(lowercase , lowercase ): __lowercase = EsmFoldConfig(**lowercase ) __lowercase = esmfold_config if vocab_list is None: logger.warning("""No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!""" ) __lowercase = get_default_vocab_list() else: __lowercase = vocab_list else: __lowercase = None __lowercase = None if self.esmfold_config is not None and getattr(self.esmfold_config , """use_esm_attn_map""" , lowercase ): raise ValueError("""The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!""" ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = super().to_dict() if isinstance(self.esmfold_config , lowercase ): __lowercase = self.esmfold_config.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : str = None lowercase__ : bool = True lowercase__ : bool = False lowercase__ : bool = False lowercase__ : bool = False lowercase__ : float = 0 lowercase__ : bool = True lowercase__ : bool = False lowercase__ : int = 128 lowercase__ : "TrunkConfig" = None def snake_case__ ( self : List[str] ) -> Any: """simple docstring""" if self.trunk is None: __lowercase = TrunkConfig() elif isinstance(self.trunk , lowercase ): __lowercase = TrunkConfig(**self.trunk ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.trunk.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 48 lowercase__ : int = 1_024 lowercase__ : int = 128 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : float = 0 lowercase__ : float = 0 lowercase__ : bool = False lowercase__ : int = 4 lowercase__ : Optional[int] = 128 lowercase__ : "StructureModuleConfig" = None def snake_case__ ( self : Tuple ) -> str: """simple docstring""" if self.structure_module is None: __lowercase = StructureModuleConfig() elif isinstance(self.structure_module , lowercase ): __lowercase = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(F"`max_recycles` should be positive, got {self.max_recycles}." ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( """`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got""" F" {self.sequence_state_dim} and {self.sequence_state_dim}." ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( """`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got""" F" {self.pairwise_state_dim} and {self.pairwise_state_dim}." ) __lowercase = self.sequence_state_dim // self.sequence_head_width __lowercase = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( """`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got""" F" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}." ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( """`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got""" F" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}." ) if self.pairwise_state_dim % 2 != 0: raise ValueError(F"`pairwise_state_dim` should be even, got {self.pairwise_state_dim}." ) if self.dropout >= 0.4: raise ValueError(F"`dropout` should not be greater than 0.4, got {self.dropout}." ) def snake_case__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.structure_module.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 384 lowercase__ : int = 128 lowercase__ : int = 16 lowercase__ : int = 128 lowercase__ : int = 12 lowercase__ : int = 4 lowercase__ : int = 8 lowercase__ : float = 0.1 lowercase__ : int = 8 lowercase__ : int = 1 lowercase__ : int = 2 lowercase__ : int = 7 lowercase__ : int = 10 lowercase__ : float = 1E-8 lowercase__ : float = 1E5 def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" return asdict(self ) def UpperCAmelCase__ ( ) -> List[Any]: return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
634
1
import argparse from collections import defaultdict import yaml UpperCamelCase__ = "docs/source/en/_toctree.yml" def UpperCAmelCase__ ( lowercase__ ) -> Dict: __lowercase = defaultdict(lowercase__ ) __lowercase = [] __lowercase = [] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({"""local""": doc["""local"""], """title""": doc["""title"""]} ) else: new_doc_list.append(lowercase__ ) __lowercase = new_doc_list __lowercase = [key for key, value in counts.items() if value > 1] __lowercase = [] for duplicate_key in duplicates: __lowercase = list({doc["""title"""] for doc in doc_list if doc["""local"""] == duplicate_key} ) if len(lowercase__ ) > 1: raise ValueError( F"{duplicate_key} is present several times in the documentation table of content at " """`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the """ """others.""" ) # Only add this once new_doc.append({"""local""": duplicate_key, """title""": titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if """local""" not in counts or counts[doc["""local"""]] == 1] ) __lowercase = sorted(lowercase__ , key=lambda lowercase__ : s["title"].lower() ) # "overview" gets special treatment and is always first if len(lowercase__ ) > 1: raise ValueError("""{doc_list} has two 'overview' docs which is not allowed.""" ) overview_doc.extend(lowercase__ ) # Sort return overview_doc def UpperCAmelCase__ ( lowercase__=False ) -> str: with open(lowercase__ , encoding="""utf-8""" ) as f: __lowercase = yaml.safe_load(f.read() ) # Get to the API doc __lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 __lowercase = content[api_idx]["""sections"""] # Then to the model doc __lowercase = 0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 __lowercase = api_doc[scheduler_idx]["""sections"""] __lowercase = clean_doc_toc(lowercase__ ) __lowercase = False if new_scheduler_doc != scheduler_doc: __lowercase = True if overwrite: __lowercase = new_scheduler_doc if diff: if overwrite: __lowercase = api_doc with open(lowercase__ , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(lowercase__ , allow_unicode=lowercase__ ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) def UpperCAmelCase__ ( lowercase__=False ) -> Any: with open(lowercase__ , encoding="""utf-8""" ) as f: __lowercase = yaml.safe_load(f.read() ) # Get to the API doc __lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 __lowercase = content[api_idx]["""sections"""] # Then to the model doc __lowercase = 0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 __lowercase = False __lowercase = api_doc[pipeline_idx]["""sections"""] __lowercase = [] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: __lowercase = pipeline_doc["""section"""] __lowercase = clean_doc_toc(lowercase__ ) if overwrite: __lowercase = new_sub_pipeline_doc new_pipeline_docs.append(lowercase__ ) # sort overall pipeline doc __lowercase = clean_doc_toc(lowercase__ ) if new_pipeline_docs != pipeline_docs: __lowercase = True if overwrite: __lowercase = new_pipeline_docs if diff: if overwrite: __lowercase = api_doc with open(lowercase__ , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(lowercase__ , allow_unicode=lowercase__ ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") UpperCamelCase__ = parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
634
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Tuple = LxmertTokenizer lowercase__ : List[str] = LxmertTokenizerFast lowercase__ : Optional[Any] = True lowercase__ : List[Any] = True def snake_case__ ( self : Tuple ) -> Tuple: """simple docstring""" super().setUp() __lowercase = [ """[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def snake_case__ ( self : Optional[int] , lowercase : int ) -> List[Any]: """simple docstring""" __lowercase = """UNwant\u00E9d,running""" __lowercase = """unwanted, running""" return input_text, output_text def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class(self.vocab_file ) __lowercase = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(lowercase , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , [7, 4, 5, 10, 8, 9] ) def snake_case__ ( self : Union[str, Any] ) -> int: """simple docstring""" if not self.test_rust_tokenizer: return __lowercase = self.get_tokenizer() __lowercase = self.get_rust_tokenizer() __lowercase = """I was born in 92000, and this is falsé.""" __lowercase = tokenizer.tokenize(lowercase ) __lowercase = rust_tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) __lowercase = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = self.get_rust_tokenizer() __lowercase = tokenizer.encode(lowercase ) __lowercase = rust_tokenizer.encode(lowercase ) self.assertListEqual(lowercase , lowercase )
634
1
import copy import tempfile import unittest from transformers import MaMaaaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.utils import cached_property from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaTokenizer from transformers.models.mam_aaa.modeling_mam_aaa import MaMaaaDecoder, MaMaaaEncoder def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , ) -> Optional[Any]: if attention_mask is None: __lowercase = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: __lowercase = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: __lowercase = torch.ones(config.encoder_layers , config.encoder_attention_heads , device=lowercase__ ) if decoder_head_mask is None: __lowercase = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=lowercase__ ) if cross_attn_head_mask is None: __lowercase = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=lowercase__ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } class _lowerCAmelCase : """simple docstring""" def __init__( self : Union[str, Any] , lowercase : Optional[Any] , lowercase : Any=13 , lowercase : Union[str, Any]=7 , lowercase : Any=True , lowercase : List[Any]=False , lowercase : List[str]=99 , lowercase : List[Any]=16 , lowercase : int=2 , lowercase : str=4 , lowercase : List[Any]=4 , lowercase : int="relu" , lowercase : Dict=0.1 , lowercase : str=0.1 , lowercase : List[Any]=0.0 , lowercase : Optional[int]=0.0 , lowercase : Tuple=20 , lowercase : str=2 , lowercase : Dict=1 , lowercase : List[Any]=0 , ) -> Tuple: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = encoder_layerdrop __lowercase = decoder_layerdrop __lowercase = max_position_embeddings __lowercase = eos_token_id __lowercase = pad_token_id __lowercase = bos_token_id def snake_case__ ( self : Optional[int] ) -> str: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = self.eos_token_id # Eos Token __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for M2M100 the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input __lowercase = input_ids.clamp(self.pad_token_id + 1 ) __lowercase = decoder_input_ids.clamp(self.pad_token_id + 1 ) __lowercase = self.get_config() __lowercase = prepare_mam_aaa_inputs_dict(lowercase , lowercase , lowercase ) return config, inputs_dict def snake_case__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" return MaMaaaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , encoder_layerdrop=self.encoder_layerdrop , decoder_layerdrop=self.decoder_layerdrop , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , ) def snake_case__ ( self : str ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.prepare_config_and_inputs() return config, inputs_dict def snake_case__ ( self : int , lowercase : Optional[int] , lowercase : Any ) -> List[Any]: """simple docstring""" __lowercase = MaMaaaModel(config=lowercase ).get_decoder().to(lowercase ).eval() __lowercase = inputs_dict["""input_ids"""] __lowercase = inputs_dict["""attention_mask"""] __lowercase = inputs_dict["""head_mask"""] # first forward pass __lowercase = model(lowercase , attention_mask=lowercase , head_mask=lowercase , use_cache=lowercase ) __lowercase , __lowercase = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids __lowercase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowercase = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and __lowercase = torch.cat([input_ids, next_tokens] , dim=-1 ) __lowercase = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) __lowercase = model(lowercase , attention_mask=lowercase )["""last_hidden_state"""] __lowercase = model(lowercase , attention_mask=lowercase , past_key_values=lowercase )[ """last_hidden_state""" ] # select random slice __lowercase = ids_tensor((1,) , output_from_past.shape[-1] ).item() __lowercase = output_from_no_past[:, -3:, random_slice_idx].detach() __lowercase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-2 ) ) def snake_case__ ( self : Tuple , lowercase : Optional[Any] , lowercase : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = MaMaaaModel(config=lowercase ).to(lowercase ).eval() __lowercase = model(**lowercase ) __lowercase = outputs.encoder_last_hidden_state __lowercase = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = model.get_encoder() encoder.save_pretrained(lowercase ) __lowercase = MaMaaaEncoder.from_pretrained(lowercase ).to(lowercase ) __lowercase = encoder(inputs_dict["""input_ids"""] , attention_mask=inputs_dict["""attention_mask"""] )[ 0 ] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1E-3 ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = model.get_decoder() decoder.save_pretrained(lowercase ) __lowercase = MaMaaaDecoder.from_pretrained(lowercase ).to(lowercase ) __lowercase = decoder( input_ids=inputs_dict["""decoder_input_ids"""] , attention_mask=inputs_dict["""decoder_attention_mask"""] , encoder_hidden_states=lowercase , encoder_attention_mask=inputs_dict["""attention_mask"""] , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1E-3 ) @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : str = ( ( MaMaaaModel, MaMaaaForConditionalGeneration, ) if is_torch_available() else () ) lowercase__ : Union[str, Any] = (MaMaaaForConditionalGeneration,) if is_torch_available() else () lowercase__ : Union[str, Any] = ( { """conversational""": MaMaaaForConditionalGeneration, """feature-extraction""": MaMaaaModel, """summarization""": MaMaaaForConditionalGeneration, """text2text-generation""": MaMaaaForConditionalGeneration, """translation""": MaMaaaForConditionalGeneration, } if is_torch_available() else {} ) lowercase__ : Optional[Any] = True lowercase__ : List[Any] = True lowercase__ : Optional[int] = False lowercase__ : List[Any] = False def snake_case__ ( self : Tuple , lowercase : List[str] , lowercase : Any , lowercase : Any , lowercase : Dict , lowercase : Tuple ) -> Optional[Any]: """simple docstring""" if pipeline_test_casse_name == "TranslationPipelineTests": # Get `ValueError: Translation requires a `src_lang` and a `tgt_lang` for this model`. # `M2M100Config` was never used in pipeline tests: cannot create a simple tokenizer. return True return False def snake_case__ ( self : Any ) -> Any: """simple docstring""" __lowercase = MaMaaaModelTester(self ) __lowercase = ConfigTester(self , config_class=lowercase ) def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" self.config_tester.run_common_tests() def snake_case__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase ) __lowercase , __lowercase = model_class.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertEqual(info["""missing_keys"""] , [] ) def snake_case__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(*lowercase ) def snake_case__ ( self : Optional[Any] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*lowercase ) def snake_case__ ( self : Any ) -> List[str]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in (MaMaaaModel, MaMaaaForConditionalGeneration): __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() __lowercase = copy.deepcopy(self._prepare_for_class(lowercase , lowercase ) ) if not self.is_encoder_decoder: __lowercase = inputs["""input_ids"""] del inputs["input_ids"] else: __lowercase = inputs["""input_ids"""] __lowercase = inputs.get("""decoder_input_ids""" , lowercase ) del inputs["input_ids"] inputs.pop("""decoder_input_ids""" , lowercase ) __lowercase = model.get_input_embeddings() if not self.is_encoder_decoder: __lowercase = wte(lowercase ) else: __lowercase = wte(lowercase ) __lowercase = wte(lowercase ) with torch.no_grad(): model(**lowercase )[0] def snake_case__ ( self : Optional[Any] ) -> Any: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs() __lowercase = input_dict["""input_ids"""] __lowercase = input_ids.ne(1 ).to(lowercase ) __lowercase = MaMaaaForConditionalGeneration(lowercase ).eval().to(lowercase ) if torch_device == "cuda": model.half() model.generate(lowercase , attention_mask=lowercase ) model.generate(num_beams=4 , do_sample=lowercase , early_stopping=lowercase , num_return_sequences=3 ) def UpperCAmelCase__ ( lowercase__ ) -> Dict: return torch.tensor(lowercase__ , dtype=torch.long , device=lowercase__ ) UpperCamelCase__ = 1e-4 @require_torch @require_sentencepiece @require_tokenizers @slow class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case__ ( self : int ) -> Optional[Any]: """simple docstring""" return MaMaaaTokenizer.from_pretrained("""facebook/m2m100_418M""" ) def snake_case__ ( self : List[str] ) -> str: """simple docstring""" __lowercase = MaMaaaModel.from_pretrained("""facebook/m2m100_418M""" ).to(lowercase ) __lowercase = _long_tensor([[128_028, 98, 12, 30_527, 2_732, 159, 7_755, 61_904, 39_144, 38, 2]] ) __lowercase = _long_tensor([[2, 128_028, 98, 12, 30_527, 2_732, 159, 7_755, 61_904, 39_144, 38]] ) __lowercase = prepare_mam_aaa_inputs_dict(model.config , lowercase , lowercase ) with torch.no_grad(): __lowercase = model(**lowercase )[0] __lowercase = torch.Size((1, 11, 1_024) ) self.assertEqual(output.shape , lowercase ) # change to expected output here __lowercase = torch.tensor( [[-0.7780, -0.1676, 0.1038], [-6.7556, -1.3992, 0.0567], [-7.5383, -0.5920, -0.2779]] , device=lowercase ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase , atol=lowercase ) ) def snake_case__ ( self : Tuple ) -> str: """simple docstring""" __lowercase = MaMaaaForConditionalGeneration.from_pretrained("""facebook/m2m100_418M""" ).to(lowercase ) # change to intended input __lowercase = _long_tensor([[128_028, 98, 12, 30_527, 2_732, 159, 7_755, 61_904, 39_144, 38, 2]] ) __lowercase = _long_tensor([[2, 128_028, 98, 12, 30_527, 2_732, 159, 7_755, 61_904, 39_144, 38]] ) __lowercase = prepare_mam_aaa_inputs_dict(model.config , lowercase , lowercase ) with torch.no_grad(): __lowercase = model(**lowercase )[0] __lowercase = torch.Size((1, 11, model.config.vocab_size) ) self.assertEqual(output.shape , lowercase ) # change to expected output here __lowercase = torch.tensor( [[-1.0448, -1.0411, 3.7992], [-3.2191, -3.2386, -1.3451], [-3.6210, -3.5993, 0.4925]] , device=lowercase ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase , atol=lowercase ) ) def snake_case__ ( self : Any ) -> List[str]: """simple docstring""" __lowercase = MaMaaaForConditionalGeneration.from_pretrained("""facebook/m2m100_418M""" ).to(lowercase ) __lowercase = MaMaaaTokenizer.from_pretrained("""facebook/m2m100_418M""" , src_lang="""fr""" , tgt_lang="""en""" ) __lowercase = [ """L'affaire NSA souligne l'absence totale de débat sur le renseignement""", """Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.""", """Lorsque François Hollande téléphone à Barack Obama ou quand le ministre des affaires étrangères Laurent""" """ Fabius convoque l'ambassadeur des Etats-Unis, ils réagissent à une vraie découverte, qui est celle de""" """ l'ampleur de la surveillance américaine sur l'ensemble des communications en France.""", ] # The below article tests that we don't add any hypotheses outside of the top n_beams __lowercase = tokenizer(lowercase , padding=lowercase , return_tensors="""pt""" ) __lowercase = model.generate( input_ids=dct["""input_ids"""].to(lowercase ) , attention_mask=dct["""attention_mask"""].to(lowercase ) , num_beams=5 , forced_bos_token_id=tokenizer.get_lang_id("""en""" ) , ) __lowercase = [ """The NSA case highlights the total absence of intelligence debate""", """I think there are two levels of response from the French government.""", """When François Hollande calls Barack Obama or when Foreign Minister Laurent Fabius calls the U.S.""" """ Ambassador, they respond to a real discovery, which is that of the scale of U.S. surveillance on all""" """ communications in France.""", ] __lowercase = tokenizer.batch_decode( hypotheses_batch.tolist() , clean_up_tokenization_spaces=lowercase , skip_special_tokens=lowercase ) assert generated == expected_en
634
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> bool: __lowercase = len(lowercase__ ) __lowercase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): __lowercase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): __lowercase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: __lowercase = subset[i - 1][j] if arr[i - 1] <= j: __lowercase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
634
1
import argparse import struct import unittest class _lowerCAmelCase : """simple docstring""" def __init__( self : str , lowercase : bytes ) -> None: """simple docstring""" __lowercase = data # Initialize hash values __lowercase = [ 0X6a09_e667, 0Xbb67_ae85, 0X3c6e_f372, 0Xa54f_f53a, 0X510e_527f, 0X9b05_688c, 0X1f83_d9ab, 0X5be0_cd19, ] # Initialize round constants __lowercase = [ 0X428a_2f98, 0X7137_4491, 0Xb5c0_fbcf, 0Xe9b5_dba5, 0X3956_c25b, 0X59f1_11f1, 0X923f_82a4, 0Xab1c_5ed5, 0Xd807_aa98, 0X1283_5b01, 0X2431_85be, 0X550c_7dc3, 0X72be_5d74, 0X80de_b1fe, 0X9bdc_06a7, 0Xc19b_f174, 0Xe49b_69c1, 0Xefbe_4786, 0X0fc1_9dc6, 0X240c_a1cc, 0X2de9_2c6f, 0X4a74_84aa, 0X5cb0_a9dc, 0X76f9_88da, 0X983e_5152, 0Xa831_c66d, 0Xb003_27c8, 0Xbf59_7fc7, 0Xc6e0_0bf3, 0Xd5a7_9147, 0X06ca_6351, 0X1429_2967, 0X27b7_0a85, 0X2e1b_2138, 0X4d2c_6dfc, 0X5338_0d13, 0X650a_7354, 0X766a_0abb, 0X81c2_c92e, 0X9272_2c85, 0Xa2bf_e8a1, 0Xa81a_664b, 0Xc24b_8b70, 0Xc76c_51a3, 0Xd192_e819, 0Xd699_0624, 0Xf40e_3585, 0X106a_a070, 0X19a4_c116, 0X1e37_6c08, 0X2748_774c, 0X34b0_bcb5, 0X391c_0cb3, 0X4ed8_aa4a, 0X5b9c_ca4f, 0X682e_6ff3, 0X748f_82ee, 0X78a5_636f, 0X84c8_7814, 0X8cc7_0208, 0X90be_fffa, 0Xa450_6ceb, 0Xbef9_a3f7, 0Xc671_78f2, ] __lowercase = self.preprocessing(self.data ) self.final_hash() @staticmethod def snake_case__ ( lowercase : bytes ) -> bytes: """simple docstring""" __lowercase = b"""\x80""" + (b"""\x00""" * (63 - (len(lowercase ) + 8) % 64)) __lowercase = struct.pack(""">Q""" , (len(lowercase ) * 8) ) return data + padding + big_endian_integer def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = [ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers __lowercase = list(struct.unpack(""">16L""" , lowercase ) ) # add 48 0-ed integers words += [0] * 48 __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array __lowercase = ( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) __lowercase = ( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) __lowercase = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X1_0000_0000 # Compression __lowercase = self.ror(lowercase , 6 ) ^ self.ror(lowercase , 11 ) ^ self.ror(lowercase , 25 ) __lowercase = (e & f) ^ ((~e & 0Xffff_ffff) & g) __lowercase = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X1_0000_0000 __lowercase = self.ror(lowercase , 2 ) ^ self.ror(lowercase , 13 ) ^ self.ror(lowercase , 22 ) __lowercase = (a & b) ^ (a & c) ^ (b & c) __lowercase = (sa + maj) % 0X1_0000_0000 __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase = ( g, f, e, ((d + tempa) % 0X1_0000_0000), c, b, a, ((tempa + tempa) % 0X1_0000_0000), ) __lowercase = [a, b, c, d, e, f, g, h] # Modify final values __lowercase = [ ((element + mutated_hash_values[index]) % 0X1_0000_0000) for index, element in enumerate(self.hashes ) ] __lowercase = """""".join([hex(lowercase )[2:].zfill(8 ) for value in self.hashes] ) def snake_case__ ( self : Union[str, Any] , lowercase : int , lowercase : int ) -> int: """simple docstring""" return 0Xffff_ffff & (value << (32 - rotations)) | (value >> rotations) class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Dict ) -> None: """simple docstring""" import hashlib __lowercase = bytes("""Test String""" , """utf-8""" ) self.assertEqual(SHAaaa(lowercase ).hash , hashlib.shaaaa(lowercase ).hexdigest() ) def UpperCAmelCase__ ( ) -> None: import doctest doctest.testmod() __lowercase = argparse.ArgumentParser() parser.add_argument( """-s""" , """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument( """-f""" , """--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) __lowercase = parser.parse_args() __lowercase = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: __lowercase = f.read() else: __lowercase = bytes(lowercase__ , """utf-8""" ) print(SHAaaa(lowercase__ ).hash ) if __name__ == "__main__": main()
634
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "hustvl/yolos-small": "https://huggingface.co/hustvl/yolos-small/resolve/main/config.json", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """yolos""" def __init__( self : Optional[int] , lowercase : Any=768 , lowercase : Tuple=12 , lowercase : Tuple=12 , lowercase : str=3_072 , lowercase : Optional[Any]="gelu" , lowercase : Union[str, Any]=0.0 , lowercase : Dict=0.0 , lowercase : Optional[int]=0.02 , lowercase : Optional[Any]=1E-1_2 , lowercase : Tuple=[512, 864] , lowercase : Optional[int]=16 , lowercase : Dict=3 , lowercase : Optional[Any]=True , lowercase : Optional[int]=100 , lowercase : Optional[int]=True , lowercase : Any=False , lowercase : Any=1 , lowercase : Any=5 , lowercase : List[str]=2 , lowercase : Union[str, Any]=5 , lowercase : str=2 , lowercase : Tuple=0.1 , **lowercase : str , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase ) __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = qkv_bias __lowercase = num_detection_tokens __lowercase = use_mid_position_embeddings __lowercase = auxiliary_loss # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = eos_coefficient class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Dict = version.parse("""1.11""" ) @property def snake_case__ ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def snake_case__ ( self : int ) -> float: """simple docstring""" return 1E-4 @property def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" return 12
634
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "hustvl/yolos-small": "https://huggingface.co/hustvl/yolos-small/resolve/main/config.json", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """yolos""" def __init__( self : Optional[int] , lowercase : Any=768 , lowercase : Tuple=12 , lowercase : Tuple=12 , lowercase : str=3_072 , lowercase : Optional[Any]="gelu" , lowercase : Union[str, Any]=0.0 , lowercase : Dict=0.0 , lowercase : Optional[int]=0.02 , lowercase : Optional[Any]=1E-1_2 , lowercase : Tuple=[512, 864] , lowercase : Optional[int]=16 , lowercase : Dict=3 , lowercase : Optional[Any]=True , lowercase : Optional[int]=100 , lowercase : Optional[int]=True , lowercase : Any=False , lowercase : Any=1 , lowercase : Any=5 , lowercase : List[str]=2 , lowercase : Union[str, Any]=5 , lowercase : str=2 , lowercase : Tuple=0.1 , **lowercase : str , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase ) __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = qkv_bias __lowercase = num_detection_tokens __lowercase = use_mid_position_embeddings __lowercase = auxiliary_loss # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = eos_coefficient class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Dict = version.parse("""1.11""" ) @property def snake_case__ ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def snake_case__ ( self : int ) -> float: """simple docstring""" return 1E-4 @property def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" return 12
634
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = IFImgaImgSuperResolutionPipeline lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) lowercase__ : Tuple = PipelineTesterMixin.required_optional_params - {"""latents"""} def snake_case__ ( self : Tuple ) -> Any: """simple docstring""" return self._get_superresolution_dummy_components() def snake_case__ ( self : List[str] , lowercase : Optional[int] , lowercase : Optional[Any]=0 ) -> Union[str, Any]: """simple docstring""" if str(lowercase ).startswith("""mps""" ): __lowercase = torch.manual_seed(lowercase ) else: __lowercase = torch.Generator(device=lowercase ).manual_seed(lowercase ) __lowercase = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """original_image""": original_image, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def snake_case__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1E-1 ) def snake_case__ ( self : Dict ) -> int: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" self._test_save_load_local() def snake_case__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
634
1
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor UpperCamelCase__ = logging.get_logger(__name__) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" def __init__( self : List[Any] , *lowercase : Any , **lowercase : Any ) -> None: """simple docstring""" warnings.warn( """The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use DPTImageProcessor instead.""" , lowercase , ) super().__init__(*lowercase , **lowercase )
634
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ = { "configuration_luke": ["LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP", "LukeConfig"], "tokenization_luke": ["LukeTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "LUKE_PRETRAINED_MODEL_ARCHIVE_LIST", "LukeForEntityClassification", "LukeForEntityPairClassification", "LukeForEntitySpanClassification", "LukeForMultipleChoice", "LukeForQuestionAnswering", "LukeForSequenceClassification", "LukeForTokenClassification", "LukeForMaskedLM", "LukeModel", "LukePreTrainedModel", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
1
import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) UpperCamelCase__ = logging.getLogger(__name__) UpperCamelCase__ = tf.data.AUTOTUNE def UpperCAmelCase__ ( ) -> int: __lowercase = argparse.ArgumentParser(description="""Train a masked language model on TPU.""" ) parser.add_argument( """--pretrained_model_config""" , type=lowercase__ , default="""roberta-base""" , help="""The model config to use. Note that we don't copy the model's weights, only the config!""" , ) parser.add_argument( """--tokenizer""" , type=lowercase__ , default="""unigram-tokenizer-wikitext""" , help="""The name of the tokenizer to load. We use the pretrained tokenizer to initialize the model's vocab size.""" , ) parser.add_argument( """--per_replica_batch_size""" , type=lowercase__ , default=8 , help="""Batch size per TPU core.""" , ) parser.add_argument( """--no_tpu""" , action="""store_true""" , help="""If set, run on CPU and don't try to initialize a TPU. Useful for debugging on non-TPU instances.""" , ) parser.add_argument( """--tpu_name""" , type=lowercase__ , help="""Name of TPU resource to initialize. Should be blank on Colab, and 'local' on TPU VMs.""" , default="""local""" , ) parser.add_argument( """--tpu_zone""" , type=lowercase__ , help="""Google cloud zone that TPU resource is located in. Only used for non-Colab TPU nodes.""" , ) parser.add_argument( """--gcp_project""" , type=lowercase__ , help="""Google cloud project name. Only used for non-Colab TPU nodes.""" ) parser.add_argument( """--bfloat16""" , action="""store_true""" , help="""Use mixed-precision bfloat16 for training. This is the recommended lower-precision format for TPU.""" , ) parser.add_argument( """--train_dataset""" , type=lowercase__ , help="""Path to training dataset to load. If the path begins with `gs://`""" """ then the dataset will be loaded from a Google Cloud Storage bucket.""" , ) parser.add_argument( """--shuffle_buffer_size""" , type=lowercase__ , default=2**18 , help="""Size of the shuffle buffer (in samples)""" , ) parser.add_argument( """--eval_dataset""" , type=lowercase__ , help="""Path to evaluation dataset to load. If the path begins with `gs://`""" """ then the dataset will be loaded from a Google Cloud Storage bucket.""" , ) parser.add_argument( """--num_epochs""" , type=lowercase__ , default=1 , help="""Number of epochs to train for.""" , ) parser.add_argument( """--learning_rate""" , type=lowercase__ , default=1E-4 , help="""Learning rate to use for training.""" , ) parser.add_argument( """--weight_decay_rate""" , type=lowercase__ , default=1E-3 , help="""Weight decay rate to use for training.""" , ) parser.add_argument( """--max_length""" , type=lowercase__ , default=512 , help="""Maximum length of tokenized sequences. Should match the setting used in prepare_tfrecord_shards.py""" , ) parser.add_argument( """--mlm_probability""" , type=lowercase__ , default=0.15 , help="""Fraction of tokens to mask during training.""" , ) parser.add_argument("""--output_dir""" , type=lowercase__ , required=lowercase__ , help="""Path to save model checkpoints to.""" ) parser.add_argument("""--hub_model_id""" , type=lowercase__ , help="""Model ID to upload to on the Hugging Face Hub.""" ) __lowercase = parser.parse_args() return args def UpperCAmelCase__ ( lowercase__ ) -> List[str]: try: if args.tpu_name: __lowercase = tf.distribute.cluster_resolver.TPUClusterResolver( args.tpu_name , zone=args.tpu_zone , project=args.gcp_project ) else: __lowercase = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: raise RuntimeError( """Couldn't connect to TPU! Most likely you need to specify --tpu_name, --tpu_zone, or """ """--gcp_project. When running on a TPU VM, use --tpu_name local.""" ) tf.config.experimental_connect_to_cluster(lowercase__ ) tf.tpu.experimental.initialize_tpu_system(lowercase__ ) return tpu def UpperCAmelCase__ ( lowercase__ ) -> int: __lowercase = 0 for file in file_list: __lowercase = file.split("""/""" )[-1] __lowercase = re.search(r"""-\d+-(\d+)\.tfrecord""" , lowercase__ ).group(1 ) __lowercase = int(lowercase__ ) num_samples += sample_count return num_samples def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__=None ) -> Dict: __lowercase = count_samples(lowercase__ ) __lowercase = tf.data.Dataset.from_tensor_slices(lowercase__ ) if shuffle: __lowercase = dataset.shuffle(len(lowercase__ ) ) __lowercase = tf.data.TFRecordDataset(lowercase__ , num_parallel_reads=lowercase__ ) # TF can't infer the total sample count because it doesn't read all the records yet, so we assert it here __lowercase = dataset.apply(tf.data.experimental.assert_cardinality(lowercase__ ) ) __lowercase = dataset.map(lowercase__ , num_parallel_calls=lowercase__ ) if shuffle: assert shuffle_buffer_size is not None __lowercase = dataset.shuffle(args.shuffle_buffer_size ) __lowercase = dataset.batch(lowercase__ , drop_remainder=lowercase__ ) __lowercase = dataset.map(lowercase__ , num_parallel_calls=lowercase__ ) __lowercase = dataset.prefetch(lowercase__ ) return dataset def UpperCAmelCase__ ( lowercase__ ) -> List[Any]: if not args.no_tpu: __lowercase = initialize_tpu(lowercase__ ) __lowercase = tf.distribute.TPUStrategy(lowercase__ ) else: __lowercase = tf.distribute.OneDeviceStrategy(device="""/gpu:0""" ) if args.bfloataa: tf.keras.mixed_precision.set_global_policy("""mixed_bfloat16""" ) __lowercase = AutoTokenizer.from_pretrained(args.tokenizer ) __lowercase = AutoConfig.from_pretrained(args.pretrained_model_config ) __lowercase = tokenizer.vocab_size __lowercase = tf.io.gfile.glob(os.path.join(args.train_dataset , """*.tfrecord""" ) ) if not training_records: raise ValueError(F"No .tfrecord files found in {args.train_dataset}." ) __lowercase = tf.io.gfile.glob(os.path.join(args.eval_dataset , """*.tfrecord""" ) ) if not eval_records: raise ValueError(F"No .tfrecord files found in {args.eval_dataset}." ) __lowercase = count_samples(lowercase__ ) __lowercase = num_train_samples // (args.per_replica_batch_size * strategy.num_replicas_in_sync) __lowercase = steps_per_epoch * args.num_epochs with strategy.scope(): __lowercase = TFAutoModelForMaskedLM.from_config(lowercase__ ) model(model.dummy_inputs ) # Pass some dummy inputs through the model to ensure all the weights are built __lowercase , __lowercase = create_optimizer( num_train_steps=lowercase__ , num_warmup_steps=total_train_steps // 20 , init_lr=args.learning_rate , weight_decay_rate=args.weight_decay_rate , ) # Transformers models compute the right loss for their task by default when labels are passed, and will # use this for training unless you specify your own loss function in compile(). model.compile(optimizer=lowercase__ , metrics=["""accuracy"""] ) def decode_fn(lowercase__ ): __lowercase = { """input_ids""": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), """attention_mask""": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), } return tf.io.parse_single_example(lowercase__ , lowercase__ ) # Many of the data collators in Transformers are TF-compilable when return_tensors == "tf", so we can # use their methods in our data pipeline. __lowercase = DataCollatorForLanguageModeling( tokenizer=lowercase__ , mlm_probability=args.mlm_probability , mlm=lowercase__ , return_tensors="""tf""" ) def mask_with_collator(lowercase__ ): # TF really needs an isin() function __lowercase = ( ~tf.cast(batch["""attention_mask"""] , tf.bool ) | (batch["""input_ids"""] == tokenizer.cls_token_id) | (batch["""input_ids"""] == tokenizer.sep_token_id) ) __lowercase , __lowercase = data_collator.tf_mask_tokens( batch["""input_ids"""] , vocab_size=len(lowercase__ ) , mask_token_id=tokenizer.mask_token_id , special_tokens_mask=lowercase__ , ) return batch __lowercase = args.per_replica_batch_size * strategy.num_replicas_in_sync __lowercase = prepare_dataset( lowercase__ , decode_fn=lowercase__ , mask_fn=lowercase__ , batch_size=lowercase__ , shuffle=lowercase__ , shuffle_buffer_size=args.shuffle_buffer_size , ) __lowercase = prepare_dataset( lowercase__ , decode_fn=lowercase__ , mask_fn=lowercase__ , batch_size=lowercase__ , shuffle=lowercase__ , ) __lowercase = [] if args.hub_model_id: callbacks.append( PushToHubCallback(output_dir=args.output_dir , hub_model_id=args.hub_model_id , tokenizer=lowercase__ ) ) model.fit( lowercase__ , validation_data=lowercase__ , epochs=args.num_epochs , callbacks=lowercase__ , ) model.save_pretrained(args.output_dir ) if __name__ == "__main__": UpperCamelCase__ = parse_args() main(args)
634
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> int: if len(lowercase__ ) != len(lowercase__ ): raise ValueError("""String lengths must match!""" ) __lowercase = 0 for chara, chara in zip(lowercase__ , lowercase__ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
634
1
def UpperCAmelCase__ ( lowercase__ = 1_000_000 ) -> int: __lowercase = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , lowercase__ ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
634
from __future__ import annotations from collections.abc import Callable UpperCamelCase__ = list[list[float | int]] def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Matrix: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(size + 1 )] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for row in range(lowercase__ ): for col in range(lowercase__ ): __lowercase = matrix[row][col] __lowercase = vector[row][0] __lowercase = 0 __lowercase = 0 while row < size and col < size: # pivoting __lowercase = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase__ , lowercase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: __lowercase , __lowercase = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase__ ): __lowercase = augmented[rowa][col] / augmented[row][col] __lowercase = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase__ ): for row in range(lowercase__ ): __lowercase = augmented[row][col] / augmented[col][col] for cola in range(lowercase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase__ ) ] def UpperCAmelCase__ ( lowercase__ ) -> Callable[[int], int]: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(lowercase__ )] for _ in range(lowercase__ )] __lowercase = [[0] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for x_val, y_val in enumerate(lowercase__ ): for col in range(lowercase__ ): __lowercase = (x_val + 1) ** (size - col - 1) __lowercase = y_val __lowercase = solve(lowercase__ , lowercase__ ) def interpolated_func(lowercase__ ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase__ ) ) return interpolated_func def UpperCAmelCase__ ( lowercase__ ) -> int: return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCAmelCase__ ( lowercase__ = question_function , lowercase__ = 10 ) -> int: __lowercase = [func(lowercase__ ) for x_val in range(1 , order + 1 )] __lowercase = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] __lowercase = 0 __lowercase = 42 __lowercase = 42 for poly in polynomials: __lowercase = 1 while func(lowercase__ ) == poly(lowercase__ ): x_val += 1 ret += poly(lowercase__ ) return ret if __name__ == "__main__": print(F"""{solution() = }""")
634
1
from __future__ import annotations from collections.abc import Callable UpperCamelCase__ = list[list[float | int]] def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Matrix: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(size + 1 )] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for row in range(lowercase__ ): for col in range(lowercase__ ): __lowercase = matrix[row][col] __lowercase = vector[row][0] __lowercase = 0 __lowercase = 0 while row < size and col < size: # pivoting __lowercase = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase__ , lowercase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: __lowercase , __lowercase = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase__ ): __lowercase = augmented[rowa][col] / augmented[row][col] __lowercase = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase__ ): for row in range(lowercase__ ): __lowercase = augmented[row][col] / augmented[col][col] for cola in range(lowercase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase__ ) ] def UpperCAmelCase__ ( lowercase__ ) -> Callable[[int], int]: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(lowercase__ )] for _ in range(lowercase__ )] __lowercase = [[0] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for x_val, y_val in enumerate(lowercase__ ): for col in range(lowercase__ ): __lowercase = (x_val + 1) ** (size - col - 1) __lowercase = y_val __lowercase = solve(lowercase__ , lowercase__ ) def interpolated_func(lowercase__ ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase__ ) ) return interpolated_func def UpperCAmelCase__ ( lowercase__ ) -> int: return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCAmelCase__ ( lowercase__ = question_function , lowercase__ = 10 ) -> int: __lowercase = [func(lowercase__ ) for x_val in range(1 , order + 1 )] __lowercase = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] __lowercase = 0 __lowercase = 42 __lowercase = 42 for poly in polynomials: __lowercase = 1 while func(lowercase__ ) == poly(lowercase__ ): x_val += 1 ret += poly(lowercase__ ) return ret if __name__ == "__main__": print(F"""{solution() = }""")
634
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ = { "configuration_ctrl": ["CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP", "CTRLConfig"], "tokenization_ctrl": ["CTRLTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "CTRLForSequenceClassification", "CTRLLMHeadModel", "CTRLModel", "CTRLPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCTRLForSequenceClassification", "TFCTRLLMHeadModel", "TFCTRLModel", "TFCTRLPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
1
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self : Dict , lowercase : List[Any] , lowercase : Dict=100 , lowercase : Dict=13 , lowercase : int=30 , lowercase : List[str]=2 , lowercase : Any=3 , lowercase : List[str]=True , lowercase : List[str]=True , lowercase : Optional[int]=32 , lowercase : Optional[int]=4 , lowercase : Tuple=4 , lowercase : str=37 , lowercase : int="gelu" , lowercase : Union[str, Any]=0.1 , lowercase : Dict=0.1 , lowercase : Tuple=10 , lowercase : int=0.02 , lowercase : int=3 , lowercase : int=None , lowercase : Optional[Any]=[0, 1, 2, 3] , ) -> str: """simple docstring""" __lowercase = parent __lowercase = 100 __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = scope __lowercase = out_indices __lowercase = num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __lowercase = (image_size // patch_size) ** 2 __lowercase = num_patches + 1 def snake_case__ ( self : Any ) -> Optional[Any]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels, pixel_labels def snake_case__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" return BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase , initializer_range=self.initializer_range , out_indices=self.out_indices , ) def snake_case__ ( self : Any , lowercase : int , lowercase : str , lowercase : Optional[int] , lowercase : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = BeitModel(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : int , lowercase : Dict , lowercase : Union[str, Any] , lowercase : Optional[Any] , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = BeitForMaskedImageModeling(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def snake_case__ ( self : List[Any] , lowercase : Union[str, Any] , lowercase : Any , lowercase : List[str] , lowercase : Any ) -> Any: """simple docstring""" __lowercase = self.type_sequence_label_size __lowercase = BeitForImageClassification(lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowercase = 1 __lowercase = BeitForImageClassification(lowercase ) model.to(lowercase ) model.eval() __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def snake_case__ ( self : List[str] , lowercase : List[str] , lowercase : Optional[int] , lowercase : Tuple , lowercase : List[str] ) -> str: """simple docstring""" __lowercase = self.num_labels __lowercase = BeitForSemanticSegmentation(lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) __lowercase = model(lowercase , labels=lowercase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def snake_case__ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : List[Any] = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) lowercase__ : int = ( { """feature-extraction""": BeitModel, """image-classification""": BeitForImageClassification, """image-segmentation""": BeitForSemanticSegmentation, } if is_torch_available() else {} ) lowercase__ : Dict = False lowercase__ : Union[str, Any] = False lowercase__ : List[str] = False def snake_case__ ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase = BeitModelTester(self ) __lowercase = ConfigTester(self , config_class=lowercase , has_text_modality=lowercase , hidden_size=37 ) def snake_case__ ( self : List[Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""BEiT does not use inputs_embeds""" ) def snake_case__ ( self : Optional[int] ) -> List[str]: """simple docstring""" pass @require_torch_multi_gpu @unittest.skip(reason="""BEiT has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def snake_case__ ( self : Any ) -> List[Any]: """simple docstring""" pass def snake_case__ ( self : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def snake_case__ ( self : Optional[int] ) -> int: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase ) def snake_case__ ( self : Dict ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def snake_case__ ( self : Union[str, Any] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase ) def snake_case__ ( self : Dict ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowercase ) def snake_case__ ( self : Any ) -> Dict: """simple docstring""" if not self.model_tester.is_training: return __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(lowercase ), BeitForMaskedImageModeling]: continue __lowercase = model_class(lowercase ) model.to(lowercase ) model.train() __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = model(**lowercase ).loss loss.backward() def snake_case__ ( self : List[Any] ) -> int: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __lowercase = False __lowercase = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(lowercase ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue __lowercase = model_class(lowercase ) model.gradient_checkpointing_enable() model.to(lowercase ) model.train() __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = model(**lowercase ).loss loss.backward() def snake_case__ ( self : Optional[Any] ) -> List[str]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = _config_zero_init(lowercase ) for model_class in self.all_model_classes: __lowercase = model_class(config=lowercase ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F"Parameter {name} of model {model_class} seems not properly initialized" , ) @slow def snake_case__ ( self : str ) -> str: """simple docstring""" for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = BeitModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def UpperCAmelCase__ ( ) -> int: __lowercase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case__ ( self : List[str] ) -> int: """simple docstring""" return BeitImageProcessor.from_pretrained("""microsoft/beit-base-patch16-224""" ) if is_vision_available() else None @slow def snake_case__ ( self : int ) -> Optional[int]: """simple docstring""" __lowercase = BeitForMaskedImageModeling.from_pretrained("""microsoft/beit-base-patch16-224-pt22k""" ).to(lowercase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase , return_tensors="""pt""" ).pixel_values.to(lowercase ) # prepare bool_masked_pos __lowercase = torch.ones((1, 196) , dtype=torch.bool ).to(lowercase ) # forward pass with torch.no_grad(): __lowercase = model(pixel_values=lowercase , bool_masked_pos=lowercase ) __lowercase = outputs.logits # verify the logits __lowercase = torch.Size((1, 196, 8_192) ) self.assertEqual(logits.shape , lowercase ) __lowercase = torch.tensor( [[-3.2437, 0.5072, -13.9174], [-3.2456, 0.4948, -13.9401], [-3.2033, 0.5121, -13.8550]] ).to(lowercase ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] , lowercase , atol=1E-2 ) ) @slow def snake_case__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" __lowercase = BeitForImageClassification.from_pretrained("""microsoft/beit-base-patch16-224""" ).to(lowercase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase , return_tensors="""pt""" ).to(lowercase ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase ) __lowercase = outputs.logits # verify the logits __lowercase = torch.Size((1, 1_000) ) self.assertEqual(logits.shape , lowercase ) __lowercase = torch.tensor([-1.2385, -1.0987, -1.0108] ).to(lowercase ) self.assertTrue(torch.allclose(logits[0, :3] , lowercase , atol=1E-4 ) ) __lowercase = 281 self.assertEqual(logits.argmax(-1 ).item() , lowercase ) @slow def snake_case__ ( self : List[Any] ) -> List[str]: """simple docstring""" __lowercase = BeitForImageClassification.from_pretrained("""microsoft/beit-large-patch16-224-pt22k-ft22k""" ).to( lowercase ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase , return_tensors="""pt""" ).to(lowercase ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase ) __lowercase = outputs.logits # verify the logits __lowercase = torch.Size((1, 21_841) ) self.assertEqual(logits.shape , lowercase ) __lowercase = torch.tensor([1.6881, -0.2787, 0.5901] ).to(lowercase ) self.assertTrue(torch.allclose(logits[0, :3] , lowercase , atol=1E-4 ) ) __lowercase = 2_396 self.assertEqual(logits.argmax(-1 ).item() , lowercase ) @slow def snake_case__ ( self : Tuple ) -> List[Any]: """simple docstring""" __lowercase = BeitForSemanticSegmentation.from_pretrained("""microsoft/beit-base-finetuned-ade-640-640""" ) __lowercase = model.to(lowercase ) __lowercase = BeitImageProcessor(do_resize=lowercase , size=640 , do_center_crop=lowercase ) __lowercase = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) __lowercase = Image.open(ds[0]["""file"""] ) __lowercase = image_processor(images=lowercase , return_tensors="""pt""" ).to(lowercase ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase ) __lowercase = outputs.logits # verify the logits __lowercase = torch.Size((1, 150, 160, 160) ) self.assertEqual(logits.shape , lowercase ) __lowercase = version.parse(PIL.__version__ ) < version.parse("""9.0.0""" ) if is_pillow_less_than_a: __lowercase = torch.tensor( [ [[-4.9225, -2.3954, -3.0522], [-2.8822, -1.0046, -1.7561], [-2.9549, -1.3228, -2.1347]], [[-5.8168, -3.4129, -4.0778], [-3.8651, -2.2214, -3.0277], [-3.8356, -2.4643, -3.3535]], [[-0.0078, 3.9952, 4.0754], [2.9856, 4.6944, 5.0035], [3.2413, 4.7813, 4.9969]], ] , device=lowercase , ) else: __lowercase = torch.tensor( [ [[-4.8960, -2.3688, -3.0355], [-2.8478, -0.9836, -1.7418], [-2.9449, -1.3332, -2.1456]], [[-5.8081, -3.4124, -4.1006], [-3.8561, -2.2081, -3.0323], [-3.8365, -2.4601, -3.3669]], [[-0.0309, 3.9868, 4.0540], [2.9640, 4.6877, 4.9976], [3.2081, 4.7690, 4.9942]], ] , device=lowercase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowercase , atol=1E-4 ) ) @slow def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = BeitForSemanticSegmentation.from_pretrained("""microsoft/beit-base-finetuned-ade-640-640""" ) __lowercase = model.to(lowercase ) __lowercase = BeitImageProcessor(do_resize=lowercase , size=640 , do_center_crop=lowercase ) __lowercase = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) __lowercase = Image.open(ds[0]["""file"""] ) __lowercase = image_processor(images=lowercase , return_tensors="""pt""" ).to(lowercase ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase ) __lowercase = outputs.logits.detach().cpu() __lowercase = image_processor.post_process_semantic_segmentation(outputs=lowercase , target_sizes=[(500, 300)] ) __lowercase = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , lowercase ) __lowercase = image_processor.post_process_semantic_segmentation(outputs=lowercase ) __lowercase = torch.Size((160, 160) ) self.assertEqual(segmentation[0].shape , lowercase )
634
import unittest import numpy as np def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = None , ) -> np.ndarray: __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) if shape_a[0] != shape_b[0]: __lowercase = ( """Expected the same number of rows for A and B. """ F"Instead found A of size {shape_a} and B of size {shape_b}" ) raise ValueError(lowercase__ ) if shape_b[1] != shape_c[1]: __lowercase = ( """Expected the same number of columns for B and C. """ F"Instead found B of size {shape_b} and C of size {shape_c}" ) raise ValueError(lowercase__ ) __lowercase = pseudo_inv if a_inv is None: try: __lowercase = np.linalg.inv(lowercase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Dict ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) __lowercase = schur_complement(lowercase , lowercase , lowercase ) __lowercase = np.block([[a, b], [b.T, c]] ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) self.assertAlmostEqual(lowercase , det_a * det_s ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
634
1
import random def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> List[Any]: __lowercase = a[left_index] __lowercase = left_index + 1 for j in range(left_index + 1 , lowercase__ ): if a[j] < pivot: __lowercase , __lowercase = a[i], a[j] i += 1 __lowercase , __lowercase = a[i - 1], a[left_index] return i - 1 def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> Tuple: if left < right: __lowercase = random.randint(lowercase__ , right - 1 ) __lowercase , __lowercase = ( a[left], a[pivot], ) # switches the pivot with the left most bound __lowercase = partition(lowercase__ , lowercase__ , lowercase__ ) quick_sort_random( lowercase__ , lowercase__ , lowercase__ ) # recursive quicksort to the left of the pivot point quick_sort_random( lowercase__ , pivot_index + 1 , lowercase__ ) # recursive quicksort to the right of the pivot point def UpperCAmelCase__ ( ) -> Optional[int]: __lowercase = input("""Enter numbers separated by a comma:\n""" ).strip() __lowercase = [int(lowercase__ ) for item in user_input.split(""",""" )] quick_sort_random(lowercase__ , 0 , len(lowercase__ ) ) print(lowercase__ ) if __name__ == "__main__": main()
634
import random def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ = False ) -> dict: __lowercase = {i: [] for i in range(lowercase__ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(lowercase__ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(lowercase__ ): for j in range(i + 1 , lowercase__ ): if random.random() < probability: graph[i].append(lowercase__ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(lowercase__ ) return graph def UpperCAmelCase__ ( lowercase__ ) -> dict: return { i: [j for j in range(lowercase__ ) if i != j] for i in range(lowercase__ ) } if __name__ == "__main__": import doctest doctest.testmod()
634
1
from collections import defaultdict from graphs.minimum_spanning_tree_prims import prisms_algorithm as mst def UpperCAmelCase__ ( ) -> List[str]: __lowercase , __lowercase = 9, 14 # noqa: F841 __lowercase = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] __lowercase = defaultdict(lowercase__ ) for nodea, nodea, cost in edges: adjancency[nodea].append([nodea, cost] ) adjancency[nodea].append([nodea, cost] ) __lowercase = mst(lowercase__ ) __lowercase = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] for answer in expected: __lowercase = tuple(answer[:2] ) __lowercase = tuple(edge[::-1] ) assert edge in result or reverse in result
634
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor UpperCamelCase__ = random.Random() def UpperCAmelCase__ ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ) -> str: if rng is None: __lowercase = global_rng __lowercase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : List[Any] , lowercase : Tuple , lowercase : Union[str, Any]=7 , lowercase : List[Any]=400 , lowercase : Any=2_000 , lowercase : Optional[int]=24 , lowercase : Any=24 , lowercase : List[str]=0.0 , lowercase : Dict=16_000 , lowercase : Union[str, Any]=True , lowercase : Dict=True , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = min_seq_length __lowercase = max_seq_length __lowercase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowercase = feature_size __lowercase = num_mel_bins __lowercase = padding_value __lowercase = sampling_rate __lowercase = return_attention_mask __lowercase = do_normalize def snake_case__ ( self : Optional[int] ) -> Any: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self : List[str] , lowercase : Tuple=False , lowercase : int=False ) -> Optional[Any]: """simple docstring""" def _flatten(lowercase : Optional[Any] ): return list(itertools.chain(*lowercase ) ) if equal_length: __lowercase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowercase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowercase = [np.asarray(lowercase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = SpeechaTextFeatureExtractor if is_speech_available() else None def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = SpeechaTextFeatureExtractionTester(self ) def snake_case__ ( self : Tuple , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" self.assertTrue(np.all(np.mean(lowercase , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowercase , axis=0 ) - 1 ) < 1E-3 ) ) def snake_case__ ( self : List[Any] ) -> str: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = [np.asarray(lowercase ) for speech_input in speech_inputs] # Test feature size __lowercase = feature_extractor(lowercase , padding=lowercase , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input __lowercase = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features __lowercase = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) # Test batched __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(lowercase , lowercase ): self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __lowercase = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowercase = np.asarray(lowercase ) __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(lowercase , lowercase ): self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) def snake_case__ ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(lowercase , lowercase ): __lowercase = feature_extractor( lowercase , padding=lowercase , max_length=lowercase , return_attention_mask=lowercase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(lowercase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def snake_case__ ( self : Any ) -> Dict: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(lowercase , lowercase ): __lowercase = feature_extractor( lowercase , max_length=lowercase , padding=lowercase , return_tensors="""np""" , return_attention_mask=lowercase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(lowercase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def snake_case__ ( self : str ) -> List[Any]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""max_length""" , max_length=4 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""longest""" , max_length=4 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""longest""" , max_length=16 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" import torch __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = np.random.rand(100 , 32 ).astype(np.floataa ) __lowercase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self : Optional[int] , lowercase : Union[str, Any] ) -> int: """simple docstring""" from datasets import load_dataset __lowercase = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __lowercase = ds.sort("""id""" ).select(range(lowercase ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on __lowercase = self._load_datasamples(1 ) __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = feature_extractor(lowercase , return_tensors="""pt""" ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , lowercase , atol=1E-4 ) )
634
1
import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging UpperCamelCase__ = logging.get_logger(__name__) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : str = ["""input_features""", """is_longer"""] def __init__( self : List[str] , lowercase : Tuple=64 , lowercase : Any=48_000 , lowercase : Tuple=480 , lowercase : str=10 , lowercase : Union[str, Any]=1_024 , lowercase : Optional[int]=0.0 , lowercase : Optional[int]=False , lowercase : float = 0 , lowercase : float = 14_000 , lowercase : int = None , lowercase : str = "fusion" , lowercase : str = "repeatpad" , **lowercase : int , ) -> Tuple: """simple docstring""" super().__init__( feature_size=lowercase , sampling_rate=lowercase , padding_value=lowercase , return_attention_mask=lowercase , **lowercase , ) __lowercase = top_db __lowercase = truncation __lowercase = padding __lowercase = fft_window_size __lowercase = (fft_window_size >> 1) + 1 __lowercase = hop_length __lowercase = max_length_s __lowercase = max_length_s * sampling_rate __lowercase = sampling_rate __lowercase = frequency_min __lowercase = frequency_max __lowercase = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=lowercase , min_frequency=lowercase , max_frequency=lowercase , sampling_rate=lowercase , norm=lowercase , mel_scale="""htk""" , ) __lowercase = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=lowercase , min_frequency=lowercase , max_frequency=lowercase , sampling_rate=lowercase , norm="""slaney""" , mel_scale="""slaney""" , ) def snake_case__ ( self : Optional[Any] ) -> Dict[str, Any]: """simple docstring""" __lowercase = copy.deepcopy(self.__dict__ ) __lowercase = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def snake_case__ ( self : str , lowercase : np.array , lowercase : Optional[np.array] = None ) -> np.ndarray: """simple docstring""" __lowercase = spectrogram( lowercase , window_function(self.fft_window_size , """hann""" ) , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=lowercase , log_mel="""dB""" , ) return log_mel_spectrogram.T def snake_case__ ( self : Union[str, Any] , lowercase : Union[str, Any] , lowercase : Any , lowercase : List[Any] ) -> str: """simple docstring""" __lowercase = np.array_split(list(range(0 , total_frames - chunk_frames + 1 ) ) , 3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk __lowercase = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk __lowercase = [0] # randomly choose index for each part __lowercase = np.random.choice(ranges[0] ) __lowercase = np.random.choice(ranges[1] ) __lowercase = np.random.choice(ranges[2] ) __lowercase = mel[idx_front : idx_front + chunk_frames, :] __lowercase = mel[idx_middle : idx_middle + chunk_frames, :] __lowercase = mel[idx_back : idx_back + chunk_frames, :] __lowercase = torch.tensor(mel[None, None, :] ) __lowercase = torch.nn.functional.interpolate( lowercase , size=[chunk_frames, 64] , mode="""bilinear""" , align_corners=lowercase ) __lowercase = mel_shrink[0][0].numpy() __lowercase = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0 ) return mel_fusion def snake_case__ ( self : Any , lowercase : np.array , lowercase : List[str] , lowercase : Dict , lowercase : Optional[Any] ) -> np.array: """simple docstring""" if waveform.shape[0] > max_length: if truncation == "rand_trunc": __lowercase = True # random crop to max_length (for compatibility) -> this should be handled by self.pad __lowercase = len(lowercase ) - max_length __lowercase = np.random.randint(0 , overflow + 1 ) __lowercase = waveform[idx : idx + max_length] __lowercase = self._np_extract_fbank_features(lowercase , self.mel_filters_slaney )[None, :] elif truncation == "fusion": __lowercase = self._np_extract_fbank_features(lowercase , self.mel_filters ) __lowercase = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed __lowercase = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. __lowercase = np.stack([mel, mel, mel, mel] , axis=0 ) __lowercase = False else: __lowercase = self._random_mel_fusion(lowercase , lowercase , lowercase ) __lowercase = True else: raise NotImplementedError(F"data_truncating {truncation} not implemented" ) else: __lowercase = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": __lowercase = int(max_length / len(lowercase ) ) __lowercase = np.stack(np.tile(lowercase , n_repeat + 1 ) )[:max_length] if padding == "repeatpad": __lowercase = int(max_length / len(lowercase ) ) __lowercase = np.stack(np.tile(lowercase , lowercase ) ) __lowercase = np.pad(lowercase , (0, max_length - waveform.shape[0]) , mode="""constant""" , constant_values=0 ) if truncation == "fusion": __lowercase = self._np_extract_fbank_features(lowercase , self.mel_filters ) __lowercase = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0 ) else: __lowercase = self._np_extract_fbank_features(lowercase , self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self : Any , lowercase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , lowercase : str = None , lowercase : Optional[str] = None , lowercase : Optional[int] = None , lowercase : Optional[int] = None , lowercase : Optional[Union[str, TensorType]] = None , **lowercase : Dict , ) -> BatchFeature: """simple docstring""" __lowercase = truncation if truncation is not None else self.truncation __lowercase = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" F" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" F" was sampled with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) __lowercase = isinstance(lowercase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}" ) __lowercase = is_batched_numpy or ( isinstance(lowercase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __lowercase = [np.asarray(lowercase , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowercase , np.ndarray ): __lowercase = np.asarray(lowercase , dtype=np.floataa ) elif isinstance(lowercase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __lowercase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __lowercase = [np.asarray(lowercase )] # convert to mel spectrogram, truncate and pad if needed. __lowercase = [ self._get_input_mel(lowercase , max_length if max_length else self.nb_max_samples , lowercase , lowercase ) for waveform in raw_speech ] __lowercase = [] __lowercase = [] for mel, longer in padded_inputs: input_mel.append(lowercase ) is_longer.append(lowercase ) if truncation == "fusion" and sum(lowercase ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer __lowercase = np.random.randint(0 , len(lowercase ) ) __lowercase = True if isinstance(input_mel[0] , lowercase ): __lowercase = [np.asarray(lowercase , dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool __lowercase = [[longer] for longer in is_longer] __lowercase = {"""input_features""": input_mel, """is_longer""": is_longer} __lowercase = BatchFeature(lowercase ) if return_tensors is not None: __lowercase = input_features.convert_to_tensors(lowercase ) return input_features
634
def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: __lowercase = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def UpperCAmelCase__ ( ) -> List[str]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
634
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..bit import BitConfig UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "Intel/dpt-large": "https://huggingface.co/Intel/dpt-large/resolve/main/config.json", # See all DPT models at https://huggingface.co/models?filter=dpt } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : str = """dpt""" def __init__( self : Dict , lowercase : Any=768 , lowercase : str=12 , lowercase : Any=12 , lowercase : Dict=3_072 , lowercase : Dict="gelu" , lowercase : Any=0.0 , lowercase : Any=0.0 , lowercase : Optional[int]=0.02 , lowercase : int=1E-1_2 , lowercase : List[Any]=384 , lowercase : List[str]=16 , lowercase : Tuple=3 , lowercase : Tuple=False , lowercase : int=True , lowercase : Any=[2, 5, 8, 11] , lowercase : Optional[int]="project" , lowercase : Optional[Any]=[4, 2, 1, 0.5] , lowercase : str=[96, 192, 384, 768] , lowercase : Tuple=256 , lowercase : Optional[Any]=-1 , lowercase : Union[str, Any]=False , lowercase : List[str]=True , lowercase : str=0.4 , lowercase : str=255 , lowercase : List[str]=0.1 , lowercase : Dict=[1, 1_024, 24, 24] , lowercase : List[Any]=[0, 1] , lowercase : List[str]=None , **lowercase : Tuple , ) -> str: """simple docstring""" super().__init__(**lowercase ) __lowercase = hidden_size __lowercase = is_hybrid if self.is_hybrid: if backbone_config is None: logger.info("""Initializing the config with a `BiT` backbone.""" ) __lowercase = { """global_padding""": """same""", """layer_type""": """bottleneck""", """depths""": [3, 4, 9], """out_features""": ["""stage1""", """stage2""", """stage3"""], """embedding_dynamic_padding""": True, } __lowercase = BitConfig(**lowercase ) elif isinstance(lowercase , lowercase ): logger.info("""Initializing the config with a `BiT` backbone.""" ) __lowercase = BitConfig(**lowercase ) elif isinstance(lowercase , lowercase ): __lowercase = backbone_config else: raise ValueError( F"backbone_config must be a dictionary or a `PretrainedConfig`, got {backbone_config.__class__}." ) __lowercase = backbone_featmap_shape __lowercase = neck_ignore_stages if readout_type != "project": raise ValueError("""Readout type must be 'project' when using `DPT-hybrid` mode.""" ) else: __lowercase = None __lowercase = None __lowercase = [] __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = qkv_bias __lowercase = backbone_out_indices if readout_type not in ["ignore", "add", "project"]: raise ValueError("""Readout_type must be one of ['ignore', 'add', 'project']""" ) __lowercase = readout_type __lowercase = reassemble_factors __lowercase = neck_hidden_sizes __lowercase = fusion_hidden_size __lowercase = head_in_index __lowercase = use_batch_norm_in_fusion_residual # auxiliary head attributes (semantic segmentation) __lowercase = use_auxiliary_head __lowercase = auxiliary_loss_weight __lowercase = semantic_loss_ignore_index __lowercase = semantic_classifier_dropout def snake_case__ ( self : str ) -> Dict: """simple docstring""" __lowercase = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: __lowercase = self.backbone_config.to_dict() __lowercase = self.__class__.model_type return output
634
def UpperCAmelCase__ ( lowercase__ = 100 ) -> int: __lowercase = n * (n + 1) * (2 * n + 1) / 6 __lowercase = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F"""{solution() = }""")
634
1
def UpperCAmelCase__ ( lowercase__ = 100 ) -> int: __lowercase = n * (n + 1) * (2 * n + 1) / 6 __lowercase = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F"""{solution() = }""")
634
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets UpperCamelCase__ = datasets.logging.get_logger(__name__) UpperCamelCase__ = "\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = \"Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric\",\n author = \"Moosavi, Nafise Sadat and\n Strube, Michael\",\n booktitle = \"Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)\",\n month = aug,\n year = \"2016\",\n address = \"Berlin, Germany\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/P16-1060\",\n doi = \"10.18653/v1/P16-1060\",\n pages = \"632--642\",\n}\n\n" UpperCamelCase__ = "\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a \"-\"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section \"*_conll File Format\"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n" UpperCamelCase__ = "\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting 'keep_singletons=False', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the 'NP_only' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting 'min_span', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n 'mentions': mentions\n 'muc': MUC metric [Vilain et al, 1995]\n 'bcub': B-cubed [Bagga and Baldwin, 1998]\n 'ceafe': CEAFe [Luo et al., 2005]\n 'lea': LEA [Moosavi and Strube, 2016]\n 'conll_score': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric('coval')\n >>> words = ['bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -',\n ... 'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)',\n ... 'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)',\n ... 'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -',\n ... 'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -',\n ... 'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {'mentions/recall': 1.0,[...] 'conll_score': 100.0}\n" def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__=False , lowercase__=False , lowercase__=True , lowercase__=False , lowercase__="dummy_doc" ) -> str: __lowercase = {doc: key_lines} __lowercase = {doc: sys_lines} __lowercase = {} __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase , __lowercase = reader.get_doc_mentions(lowercase__ , key_doc_lines[doc] , lowercase__ ) key_singletons_num += singletons_num if NP_only or min_span: __lowercase = reader.set_annotated_parse_trees(lowercase__ , key_doc_lines[doc] , lowercase__ , lowercase__ ) __lowercase , __lowercase = reader.get_doc_mentions(lowercase__ , sys_doc_lines[doc] , lowercase__ ) sys_singletons_num += singletons_num if NP_only or min_span: __lowercase = reader.set_annotated_parse_trees(lowercase__ , key_doc_lines[doc] , lowercase__ , lowercase__ ) if remove_nested: __lowercase , __lowercase = reader.remove_nested_coref_mentions(lowercase__ , lowercase__ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters __lowercase , __lowercase = reader.remove_nested_coref_mentions(lowercase__ , lowercase__ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters __lowercase = reader.get_mention_assignments(lowercase__ , lowercase__ ) __lowercase = reader.get_mention_assignments(lowercase__ , lowercase__ ) __lowercase = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F"annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}" ) logger.info( """Number of resulting singleton clusters in the key """ F"annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}" ) if not keep_singletons: logger.info( F"{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system " """files, respectively""" ) return doc_coref_infos def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) -> Optional[Any]: __lowercase = get_coref_infos(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) __lowercase = {} __lowercase = 0 __lowercase = 0 for name, metric in metrics: __lowercase , __lowercase , __lowercase = evaluator.evaluate_documents(lowercase__ , lowercase__ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F"{name}/recall": recall, F"{name}/precision": precision, F"{name}/f1": fa} ) logger.info( name.ljust(10 ) , F"Recall: {recall * 100:.2f}" , F" Precision: {precision * 100:.2f}" , F" F1: {fa * 100:.2f}" , ) if conll_subparts_num == 3: __lowercase = (conll / 3) * 100 logger.info(F"CoNLL score: {conll:.2f}" ) output_scores.update({"""conll_score""": conll} ) return output_scores def UpperCAmelCase__ ( lowercase__ ) -> List[Any]: __lowercase = False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: __lowercase = line.split()[5] if not parse_col == "-": __lowercase = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def snake_case__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def snake_case__ ( self : Tuple , lowercase : Dict , lowercase : Optional[int] , lowercase : Dict=True , lowercase : List[str]=False , lowercase : int=False , lowercase : Dict=False ) -> str: """simple docstring""" __lowercase = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: __lowercase = util.check_gold_parse_annotation(lowercase ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" __lowercase = evaluate( key_lines=lowercase , sys_lines=lowercase , metrics=lowercase , NP_only=lowercase , remove_nested=lowercase , keep_singletons=lowercase , min_span=lowercase , ) return score
634
1
from math import factorial UpperCamelCase__ = {str(d): factorial(d) for d in range(10)} def UpperCAmelCase__ ( lowercase__ ) -> int: return sum(DIGIT_FACTORIAL[d] for d in str(lowercase__ ) ) def UpperCAmelCase__ ( ) -> int: __lowercase = 7 * factorial(9 ) + 1 return sum(i for i in range(3 , lowercase__ ) if sum_of_digit_factorial(lowercase__ ) == i ) if __name__ == "__main__": print(F"""{solution() = }""")
634
UpperCamelCase__ = { "joule": 1.0, "kilojoule": 10_00, "megajoule": 1_00_00_00, "gigajoule": 10_00_00_00_00, "wattsecond": 1.0, "watthour": 36_00, "kilowatthour": 3_60_00_00, "newtonmeter": 1.0, "calorie_nutr": 41_86.8, "kilocalorie_nutr": 4_18_68_00.00, "electronvolt": 1.6_0217_6634e-19, "britishthermalunit_it": 10_55.0_55_85, "footpound": 1.355_818, } def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: __lowercase = ( F"Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n" F"Valid values are: {', '.join(lowercase__ )}" ) raise ValueError(lowercase__ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
634
1
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=_UpperCAmelCase ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : str = field(default="""audio-classification""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowercase__ : ClassVar[Features] = Features({"""audio""": Audio()} ) lowercase__ : ClassVar[Features] = Features({"""labels""": ClassLabel} ) lowercase__ : str = "audio" lowercase__ : str = "labels" def snake_case__ ( self : Any , lowercase : List[Any] ) -> Any: """simple docstring""" if self.label_column not in features: raise ValueError(F"Column {self.label_column} is not present in features." ) if not isinstance(features[self.label_column] , lowercase ): raise ValueError(F"Column {self.label_column} is not a ClassLabel." ) __lowercase = copy.deepcopy(self ) __lowercase = self.label_schema.copy() __lowercase = features[self.label_column] __lowercase = label_schema return task_template @property def snake_case__ ( self : Union[str, Any] ) -> Dict[str, str]: """simple docstring""" return { self.audio_column: "audio", self.label_column: "labels", }
634
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_UpperCAmelCase ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : str = field(default="""summarization""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowercase__ : ClassVar[Features] = Features({"""text""": Value("""string""" )} ) lowercase__ : ClassVar[Features] = Features({"""summary""": Value("""string""" )} ) lowercase__ : str = "text" lowercase__ : str = "summary" @property def snake_case__ ( self : List[Any] ) -> Dict[str, str]: """simple docstring""" return {self.text_column: "text", self.summary_column: "summary"}
634
1
def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: __lowercase = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def UpperCAmelCase__ ( ) -> List[str]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
634
def UpperCAmelCase__ ( lowercase__ ) -> Optional[int]: __lowercase = len(lowercase__ ) __lowercase = sum(lowercase__ ) __lowercase = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): __lowercase = True for i in range(1 , s + 1 ): __lowercase = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): __lowercase = dp[i][j - 1] if arr[i - 1] <= j: __lowercase = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: __lowercase = s - 2 * j break return diff
634
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "google/switch-base-8": "https://huggingface.co/google/switch-base-8/blob/main/config.json", } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Dict = """switch_transformers""" lowercase__ : int = ["""past_key_values"""] lowercase__ : Dict = {"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self : Any , lowercase : List[Any]=32_128 , lowercase : Optional[int]=768 , lowercase : Optional[int]=64 , lowercase : int=2_048 , lowercase : Union[str, Any]=64 , lowercase : List[Any]=12 , lowercase : Tuple=3 , lowercase : Tuple=12 , lowercase : Union[str, Any]=3 , lowercase : List[str]=12 , lowercase : int=8 , lowercase : Optional[Any]=False , lowercase : List[str]=0.01 , lowercase : str="float32" , lowercase : int=False , lowercase : Any=32 , lowercase : Optional[int]=128 , lowercase : Tuple=0.1 , lowercase : Union[str, Any]=1E-6 , lowercase : List[str]=0.001 , lowercase : Optional[Any]=0.001 , lowercase : Any=1.0 , lowercase : Dict="relu" , lowercase : List[Any]=True , lowercase : Any=False , lowercase : Optional[Any]=True , lowercase : Dict=0 , lowercase : List[Any]=1 , **lowercase : str , ) -> str: """simple docstring""" __lowercase = vocab_size __lowercase = d_model __lowercase = d_kv __lowercase = d_ff __lowercase = num_sparse_encoder_layers __lowercase = num_layers __lowercase = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry __lowercase = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: __lowercase = self.num_layers // self.num_sparse_encoder_layers else: __lowercase = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: __lowercase = self.num_decoder_layers // self.num_sparse_decoder_layers else: __lowercase = self.num_decoder_layers # HACK: this will create 0 sparse layers __lowercase = num_heads __lowercase = num_experts __lowercase = expert_capacity __lowercase = router_bias __lowercase = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}" ) __lowercase = router_dtype __lowercase = router_ignore_padding_tokens __lowercase = relative_attention_num_buckets __lowercase = relative_attention_max_distance __lowercase = dropout_rate __lowercase = layer_norm_epsilon __lowercase = initializer_factor __lowercase = feed_forward_proj __lowercase = use_cache __lowercase = add_router_probs __lowercase = router_z_loss_coef __lowercase = router_aux_loss_coef __lowercase = self.feed_forward_proj.split("""-""" ) __lowercase = act_info[-1] __lowercase = act_info[0] == """gated""" if len(lowercase ) > 1 and act_info[0] != "gated" or len(lowercase ) > 2: raise ValueError( F"`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer." """Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. """ """'gated-gelu' or 'relu'""" ) # for backwards compatibility if feed_forward_proj == "gated-gelu": __lowercase = """gelu_new""" super().__init__( pad_token_id=lowercase , eos_token_id=lowercase , is_encoder_decoder=lowercase , **lowercase , )
634
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class _lowerCAmelCase : """simple docstring""" def __init__( self : List[Any] , lowercase : str , lowercase : Union[str, Any]=13 , lowercase : Tuple=32 , lowercase : Optional[Any]=2 , lowercase : Tuple=3 , lowercase : Tuple=16 , lowercase : Tuple=[1, 2, 1] , lowercase : Optional[Any]=[2, 2, 4] , lowercase : Dict=2 , lowercase : Optional[int]=2.0 , lowercase : List[Any]=True , lowercase : str=0.0 , lowercase : Any=0.0 , lowercase : Optional[int]=0.1 , lowercase : int="gelu" , lowercase : Tuple=False , lowercase : Optional[Any]=True , lowercase : int=0.02 , lowercase : Union[str, Any]=1E-5 , lowercase : Dict=True , lowercase : Any=None , lowercase : str=True , lowercase : str=10 , lowercase : Dict=8 , lowercase : int=["stage1", "stage2", "stage3"] , lowercase : Optional[int]=[1, 2, 3] , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def snake_case__ ( self : List[str] ) -> int: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def snake_case__ ( self : Any , lowercase : List[Any] , lowercase : Optional[int] , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = MaskFormerSwinModel(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def snake_case__ ( self : Any , lowercase : Tuple , lowercase : Any , lowercase : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(lowercase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=lowercase ) def snake_case__ ( self : int ) -> Any: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Optional[int] = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) lowercase__ : List[str] = {"""feature-extraction""": MaskFormerSwinModel} if is_torch_available() else {} lowercase__ : List[str] = False lowercase__ : int = False lowercase__ : int = False lowercase__ : Tuple = False lowercase__ : Optional[Any] = False def snake_case__ ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=lowercase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" pass def snake_case__ ( self : Tuple ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" return def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def snake_case__ ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowercase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def snake_case__ ( self : int ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def snake_case__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" pass def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def snake_case__ ( self : Any ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def snake_case__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" pass def snake_case__ ( self : Tuple , lowercase : Tuple , lowercase : Optional[int] , lowercase : Optional[int] , lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowercase ) , lowercase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def snake_case__ ( self : int ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , lowercase ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def snake_case__ ( self : Any ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def snake_case__ ( self : List[str] ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" pass def snake_case__ ( self : Tuple ) -> List[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(lowercase : Optional[int] ): __lowercase = 0 return t def check_equivalence(lowercase : Optional[int] , lowercase : str , lowercase : str , lowercase : Tuple={} ): with torch.no_grad(): __lowercase = model(**lowercase , return_dict=lowercase , **lowercase ) __lowercase = model(**lowercase , return_dict=lowercase , **lowercase ).to_tuple() def recursive_check(lowercase : int , lowercase : Optional[Any] ): if isinstance(lowercase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(lowercase , lowercase ): recursive_check(lowercase , lowercase ) elif isinstance(lowercase , lowercase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(lowercase , lowercase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(lowercase ) , set_nan_tensor_to_zero(lowercase ) , atol=1E-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:" F" {torch.isnan(lowercase ).any()} and `inf`: {torch.isinf(lowercase )}. Dict has" F" `nan`: {torch.isnan(lowercase ).any()} and `inf`: {torch.isinf(lowercase )}." ) , ) recursive_check(lowercase , lowercase ) for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() __lowercase = self._prepare_for_class(lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) check_equivalence(lowercase , lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) check_equivalence(lowercase , lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) check_equivalence(lowercase , lowercase , lowercase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) check_equivalence(lowercase , lowercase , lowercase , {"""output_hidden_states""": True} ) @require_torch class _lowerCAmelCase ( unittest.TestCase , _UpperCAmelCase ): """simple docstring""" lowercase__ : List[str] = (MaskFormerSwinBackbone,) if is_torch_available() else () lowercase__ : Any = MaskFormerSwinConfig def snake_case__ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def snake_case__ ( self : Any ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(lowercase ) backbone.to(lowercase ) backbone.eval() __lowercase = backbone(**lowercase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , lowercase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**lowercase , output_hidden_states=lowercase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**lowercase , output_attentions=lowercase ) self.assertIsNotNone(outputs.attentions )
634
1
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : List[Any] , lowercase : str ) -> str: """simple docstring""" for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] , model_result["""ss"""] ): __lowercase = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(lowercase ) def snake_case__ ( self : str ) -> List[Any]: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : List[Any] ) -> int: """simple docstring""" __lowercase = """sgugger/tiny-distilbert-classification""" __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , only_pretrain_model=lowercase , ) __lowercase = PyTorchBenchmark(lowercase ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : Dict ) -> str: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , torchscript=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == """cpu""" , """Cant do half precision""" ) def snake_case__ ( self : Tuple ) -> Any: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , fpaa=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : Tuple ) -> Any: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" __lowercase = AutoConfig.from_pretrained(lowercase ) # set architectures equal to `None` __lowercase = None __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase , configs=[config] ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == """cpu""" , """Can't do half precision""" ) def snake_case__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , fpaa=lowercase , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def snake_case__ ( self : Optional[Any] ) -> List[str]: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" __lowercase = AutoConfig.from_pretrained(lowercase ) __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase , configs=[config] ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : str ) -> Optional[int]: """simple docstring""" __lowercase = """sshleifer/tinier_bart""" __lowercase = AutoConfig.from_pretrained(lowercase ) __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase , configs=[config] ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def snake_case__ ( self : List[str] ) -> str: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" __lowercase = AutoConfig.from_pretrained(lowercase ) __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase , configs=[config] ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase = """sshleifer/tinier_bart""" __lowercase = AutoConfig.from_pretrained(lowercase ) __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase , configs=[config] ) __lowercase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def snake_case__ ( self : Tuple ) -> List[Any]: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , save_to_csv=lowercase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(lowercase , """inf_time.csv""" ) , train_memory_csv_file=os.path.join(lowercase , """train_mem.csv""" ) , inference_memory_csv_file=os.path.join(lowercase , """inf_mem.csv""" ) , train_time_csv_file=os.path.join(lowercase , """train_time.csv""" ) , env_info_csv_file=os.path.join(lowercase , """env.csv""" ) , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase ) benchmark.run() self.assertTrue(Path(os.path.join(lowercase , """inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase , """train_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase , """inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase , """train_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase , """env.csv""" ) ).exists() ) def snake_case__ ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(lowercase : Optional[Any] ): self.assertTrue(hasattr(lowercase , """sequential""" ) ) self.assertTrue(hasattr(lowercase , """cumulative""" ) ) self.assertTrue(hasattr(lowercase , """current""" ) ) self.assertTrue(hasattr(lowercase , """total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: __lowercase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(lowercase , """log.txt""" ) , log_print=lowercase , trace_memory_line_by_line=lowercase , multi_process=lowercase , ) __lowercase = PyTorchBenchmark(lowercase ) __lowercase = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(lowercase , """log.txt""" ) ).exists() )
634
import importlib import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Union import torch from ..utils import BaseOutput UpperCamelCase__ = "scheduler_config.json" class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : List[Any] = 1 lowercase__ : Tuple = 2 lowercase__ : Union[str, Any] = 3 lowercase__ : Union[str, Any] = 4 lowercase__ : str = 5 lowercase__ : Any = 6 lowercase__ : Any = 7 lowercase__ : List[str] = 8 lowercase__ : Union[str, Any] = 9 lowercase__ : int = 10 lowercase__ : List[str] = 11 lowercase__ : List[Any] = 12 lowercase__ : str = 13 lowercase__ : Optional[int] = 14 @dataclass class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : torch.FloatTensor class _lowerCAmelCase : """simple docstring""" lowercase__ : Optional[int] = SCHEDULER_CONFIG_NAME lowercase__ : int = [] lowercase__ : Dict = True @classmethod def snake_case__ ( cls : str , lowercase : Dict[str, Any] = None , lowercase : Optional[str] = None , lowercase : Any=False , **lowercase : List[str] , ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase , __lowercase = cls.load_config( pretrained_model_name_or_path=lowercase , subfolder=lowercase , return_unused_kwargs=lowercase , return_commit_hash=lowercase , **lowercase , ) return cls.from_config(lowercase , return_unused_kwargs=lowercase , **lowercase ) def snake_case__ ( self : Dict , lowercase : Union[str, os.PathLike] , lowercase : bool = False , **lowercase : List[str] ) -> Optional[Any]: """simple docstring""" self.save_config(save_directory=lowercase , push_to_hub=lowercase , **lowercase ) @property def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" return self._get_compatibles() @classmethod def snake_case__ ( cls : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = list(set([cls.__name__] + cls._compatibles ) ) __lowercase = importlib.import_module(__name__.split(""".""" )[0] ) __lowercase = [ getattr(lowercase , lowercase ) for c in compatible_classes_str if hasattr(lowercase , lowercase ) ] return compatible_classes
634
1
import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging UpperCamelCase__ = logging.get_logger(__name__) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Any = ["""input_features"""] def __init__( self : Optional[int] , lowercase : List[Any]=80 , lowercase : List[str]=16_000 , lowercase : Optional[Any]=160 , lowercase : Optional[int]=30 , lowercase : Optional[Any]=400 , lowercase : Optional[Any]=0.0 , lowercase : List[str]=False , **lowercase : int , ) -> Any: """simple docstring""" super().__init__( feature_size=lowercase , sampling_rate=lowercase , padding_value=lowercase , return_attention_mask=lowercase , **lowercase , ) __lowercase = n_fft __lowercase = hop_length __lowercase = chunk_length __lowercase = chunk_length * sampling_rate __lowercase = self.n_samples // hop_length __lowercase = sampling_rate __lowercase = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=lowercase , min_frequency=0.0 , max_frequency=8000.0 , sampling_rate=lowercase , norm="""slaney""" , mel_scale="""slaney""" , ) def snake_case__ ( self : Dict , lowercase : np.array ) -> np.ndarray: """simple docstring""" __lowercase = spectrogram( lowercase , window_function(self.n_fft , """hann""" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel="""log10""" , ) __lowercase = log_spec[:, :-1] __lowercase = np.maximum(lowercase , log_spec.max() - 8.0 ) __lowercase = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def snake_case__ ( lowercase : List[np.ndarray] , lowercase : List[np.ndarray] , lowercase : float = 0.0 ) -> List[np.ndarray]: """simple docstring""" if attention_mask is not None: __lowercase = np.array(lowercase , np.intaa ) __lowercase = [] for vector, length in zip(lowercase , attention_mask.sum(-1 ) ): __lowercase = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: __lowercase = padding_value normed_input_values.append(lowercase ) else: __lowercase = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__( self : Any , lowercase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , lowercase : bool = True , lowercase : Optional[int] = None , lowercase : Optional[Union[str, TensorType]] = None , lowercase : Optional[bool] = None , lowercase : Optional[str] = "max_length" , lowercase : Optional[int] = None , lowercase : Optional[int] = None , lowercase : Optional[bool] = None , **lowercase : List[str] , ) -> BatchFeature: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" F" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" F" was sampled with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) __lowercase = isinstance(lowercase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}" ) __lowercase = is_batched_numpy or ( isinstance(lowercase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __lowercase = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(lowercase , np.ndarray ): __lowercase = np.asarray(lowercase , dtype=np.floataa ) elif isinstance(lowercase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __lowercase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __lowercase = [np.asarray([raw_speech] ).T] __lowercase = BatchFeature({"""input_features""": raw_speech} ) # convert into correct format for padding __lowercase = self.pad( lowercase , padding=lowercase , max_length=max_length if max_length else self.n_samples , truncation=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: __lowercase = self.zero_mean_unit_var_norm( padded_inputs["""input_features"""] , attention_mask=padded_inputs["""attention_mask"""] , padding_value=self.padding_value , ) __lowercase = np.stack(padded_inputs["""input_features"""] , axis=0 ) # make sure list is in array format __lowercase = padded_inputs.get("""input_features""" ).transpose(2 , 0 , 1 ) __lowercase = [self._np_extract_fbank_features(lowercase ) for waveform in input_features[0]] if isinstance(input_features[0] , lowercase ): __lowercase = [np.asarray(lowercase , dtype=np.floataa ) for feature in input_features] else: __lowercase = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) __lowercase = padded_inputs["""attention_mask"""][:, :: self.hop_length] if return_tensors is not None: __lowercase = padded_inputs.convert_to_tensors(lowercase ) return padded_inputs def snake_case__ ( self : List[Any] ) -> Dict[str, Any]: """simple docstring""" __lowercase = copy.deepcopy(self.__dict__ ) __lowercase = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
634
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "asapp/sew-tiny-100k": "https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json", # See all SEW models at https://huggingface.co/models?filter=sew } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """sew""" def __init__( self : List[Any] , lowercase : int=32 , lowercase : List[str]=768 , lowercase : Dict=12 , lowercase : str=12 , lowercase : str=3_072 , lowercase : Optional[int]=2 , lowercase : List[str]="gelu" , lowercase : List[str]=0.1 , lowercase : Tuple=0.1 , lowercase : Dict=0.1 , lowercase : Any=0.0 , lowercase : Dict=0.1 , lowercase : Optional[int]=0.1 , lowercase : List[str]=0.02 , lowercase : Dict=1E-5 , lowercase : Tuple="group" , lowercase : int="gelu" , lowercase : Any=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , lowercase : Optional[int]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowercase : Any=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowercase : List[str]=False , lowercase : Tuple=128 , lowercase : int=16 , lowercase : Union[str, Any]=True , lowercase : List[str]=0.05 , lowercase : Optional[int]=10 , lowercase : Any=2 , lowercase : Optional[Any]=0.0 , lowercase : Optional[Any]=10 , lowercase : int=0 , lowercase : Optional[int]="mean" , lowercase : List[Any]=False , lowercase : str=False , lowercase : int=256 , lowercase : str=0 , lowercase : List[Any]=1 , lowercase : List[Any]=2 , **lowercase : List[Any] , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase , pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase ) __lowercase = hidden_size __lowercase = feat_extract_norm __lowercase = feat_extract_activation __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = conv_bias __lowercase = num_conv_pos_embeddings __lowercase = num_conv_pos_embedding_groups __lowercase = len(self.conv_dim ) __lowercase = num_hidden_layers __lowercase = intermediate_size __lowercase = squeeze_factor __lowercase = hidden_act __lowercase = num_attention_heads __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = feat_proj_dropout __lowercase = final_dropout __lowercase = layerdrop __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" F"but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)" F"= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowercase = apply_spec_augment __lowercase = mask_time_prob __lowercase = mask_time_length __lowercase = mask_time_min_masks __lowercase = mask_feature_prob __lowercase = mask_feature_length __lowercase = mask_feature_min_masks # ctc loss __lowercase = ctc_loss_reduction __lowercase = ctc_zero_infinity # sequence classification __lowercase = use_weighted_layer_sum __lowercase = classifier_proj_size @property def snake_case__ ( self : Dict ) -> str: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
634
1
import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Union[str, Any] = GPTaTokenizer lowercase__ : str = GPTaTokenizerFast lowercase__ : List[Any] = True lowercase__ : List[str] = {"""add_prefix_space""": True} lowercase__ : Optional[Any] = False def snake_case__ ( self : int ) -> Optional[Any]: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowercase = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", """<|endoftext|>""", ] __lowercase = dict(zip(lowercase , range(len(lowercase ) ) ) ) __lowercase = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] __lowercase = {"""unk_token""": """<unk>"""} __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowercase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(lowercase ) ) def snake_case__ ( self : Union[str, Any] , **lowercase : Tuple ) -> Optional[Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **lowercase ) def snake_case__ ( self : Dict , **lowercase : Optional[int] ) -> str: """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **lowercase ) def snake_case__ ( self : List[str] , lowercase : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = """lower newer""" __lowercase = """lower newer""" return input_text, output_text def snake_case__ ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) __lowercase = """lower newer""" __lowercase = ["""\u0120low""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] __lowercase = tokenizer.tokenize(lowercase , add_prefix_space=lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = tokens + [tokenizer.unk_token] __lowercase = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , lowercase ) def snake_case__ ( self : Any ) -> str: """simple docstring""" if not self.test_rust_tokenizer: return __lowercase = self.get_tokenizer() __lowercase = self.get_rust_tokenizer(add_prefix_space=lowercase ) __lowercase = """lower newer""" # Testing tokenization __lowercase = tokenizer.tokenize(lowercase , add_prefix_space=lowercase ) __lowercase = rust_tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) # Testing conversion to ids without special tokens __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase , add_prefix_space=lowercase ) __lowercase = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase ) self.assertListEqual(lowercase , lowercase ) # Testing conversion to ids with special tokens __lowercase = self.get_rust_tokenizer(add_prefix_space=lowercase ) __lowercase = tokenizer.encode(lowercase , add_prefix_space=lowercase ) __lowercase = rust_tokenizer.encode(lowercase ) self.assertListEqual(lowercase , lowercase ) # Testing the unknown token __lowercase = tokens + [rust_tokenizer.unk_token] __lowercase = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(lowercase ) , lowercase ) def snake_case__ ( self : List[str] , *lowercase : Tuple , **lowercase : int ) -> List[str]: """simple docstring""" pass def snake_case__ ( self : Tuple , lowercase : Optional[Any]=15 ) -> Union[str, Any]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): __lowercase = self.rust_tokenizer_class.from_pretrained(lowercase , **lowercase ) # Simple input __lowercase = """This is a simple input""" __lowercase = ["""This is a simple input 1""", """This is a simple input 2"""] __lowercase = ("""This is a simple input""", """This is a pair""") __lowercase = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(lowercase , tokenizer_r.encode , lowercase , max_length=lowercase , padding="""max_length""" ) # Simple input self.assertRaises(lowercase , tokenizer_r.encode_plus , lowercase , max_length=lowercase , padding="""max_length""" ) # Simple input self.assertRaises( lowercase , tokenizer_r.batch_encode_plus , lowercase , max_length=lowercase , padding="""max_length""" , ) # Pair input self.assertRaises(lowercase , tokenizer_r.encode , lowercase , max_length=lowercase , padding="""max_length""" ) # Pair input self.assertRaises(lowercase , tokenizer_r.encode_plus , lowercase , max_length=lowercase , padding="""max_length""" ) # Pair input self.assertRaises( lowercase , tokenizer_r.batch_encode_plus , lowercase , max_length=lowercase , padding="""max_length""" , ) def snake_case__ ( self : Any ) -> List[str]: """simple docstring""" __lowercase = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token="""<pad>""" ) # Simple input __lowercase = """This is a simple input""" __lowercase = ["""This is a simple input looooooooong""", """This is a simple input"""] __lowercase = ("""This is a simple input""", """This is a pair""") __lowercase = [ ("""This is a simple input loooooong""", """This is a simple input"""), ("""This is a simple pair loooooong""", """This is a simple pair"""), ] __lowercase = tokenizer.pad_token_id __lowercase = tokenizer(lowercase , padding="""max_length""" , max_length=30 , return_tensors="""np""" ) __lowercase = tokenizer(lowercase , padding=lowercase , truncate=lowercase , return_tensors="""np""" ) __lowercase = tokenizer(*lowercase , padding="""max_length""" , max_length=60 , return_tensors="""np""" ) __lowercase = tokenizer(lowercase , padding=lowercase , truncate=lowercase , return_tensors="""np""" ) # s # test single string max_length padding self.assertEqual(out_s["""input_ids"""].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s["""input_ids"""] ) self.assertTrue(0 in out_s["""attention_mask"""] ) # s2 # test automatic padding self.assertEqual(out_sa["""input_ids"""].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["""input_ids"""][0] ) self.assertFalse(0 in out_sa["""attention_mask"""][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["""input_ids"""][1] ) self.assertTrue(0 in out_sa["""attention_mask"""][1] ) # p # test single pair max_length padding self.assertEqual(out_p["""input_ids"""].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p["""input_ids"""] ) self.assertTrue(0 in out_p["""attention_mask"""] ) # p2 # test automatic padding pair self.assertEqual(out_pa["""input_ids"""].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["""input_ids"""][0] ) self.assertFalse(0 in out_pa["""attention_mask"""][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["""input_ids"""][1] ) self.assertTrue(0 in out_pa["""attention_mask"""][1] ) def snake_case__ ( self : Any ) -> Optional[Any]: """simple docstring""" __lowercase = """$$$""" __lowercase = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=lowercase , add_bos_token=lowercase ) __lowercase = """This is a simple input""" __lowercase = ["""This is a simple input 1""", """This is a simple input 2"""] __lowercase = tokenizer.bos_token_id __lowercase = tokenizer(lowercase ) __lowercase = tokenizer(lowercase ) self.assertEqual(out_s.input_ids[0] , lowercase ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) __lowercase = tokenizer.decode(out_s.input_ids ) __lowercase = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , lowercase ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def snake_case__ ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass def snake_case__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = [self.get_tokenizer(do_lower_case=lowercase , add_bos_token=lowercase )] for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = """Encode this.""" __lowercase = """This one too please.""" __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) encoded_sequence += tokenizer.encode(lowercase , add_special_tokens=lowercase ) __lowercase = tokenizer.encode_plus( lowercase , lowercase , add_special_tokens=lowercase , return_special_tokens_mask=lowercase , ) __lowercase = encoded_sequence_dict["""input_ids"""] __lowercase = encoded_sequence_dict["""special_tokens_mask"""] self.assertEqual(len(lowercase ) , len(lowercase ) ) __lowercase = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(lowercase ) ] __lowercase = [x for x in filtered_sequence if x is not None] self.assertEqual(lowercase , lowercase ) @require_tokenizers class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : int ) -> Union[str, Any]: """simple docstring""" __lowercase = AutoTokenizer.from_pretrained("""facebook/opt-350m""" , from_slow=lowercase ) __lowercase = """A photo of a cat""" __lowercase = tokenizer.encode( lowercase , ) self.assertEqual(lowercase , [2, 250, 1_345, 9, 10, 4_758] ) tokenizer.save_pretrained("""test_opt""" ) __lowercase = AutoTokenizer.from_pretrained("""./test_opt""" ) __lowercase = tokenizer.encode( lowercase , ) self.assertEqual(lowercase , [2, 250, 1_345, 9, 10, 4_758] ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = AutoTokenizer.from_pretrained("""facebook/opt-350m""" , use_slow=lowercase ) __lowercase = """A photo of a cat""" __lowercase = tokenizer.encode( lowercase , ) # Same as above self.assertEqual(lowercase , [2, 250, 1_345, 9, 10, 4_758] ) @unittest.skip("""This test is failing because of a bug in the fast tokenizer""" ) def snake_case__ ( self : Any ) -> List[str]: """simple docstring""" __lowercase = AutoTokenizer.from_pretrained("""facebook/opt-350m""" , from_slow=lowercase ) __lowercase = """bos""" __lowercase = tokenizer.get_vocab()["""bos"""] __lowercase = """A photo of a cat""" __lowercase = tokenizer.encode( lowercase , ) # We changed the bos token self.assertEqual(lowercase , [31_957, 250, 1_345, 9, 10, 4_758] ) tokenizer.save_pretrained("""./tok""" ) __lowercase = AutoTokenizer.from_pretrained("""./tok""" ) self.assertTrue(tokenizer.is_fast ) __lowercase = tokenizer.encode( lowercase , ) self.assertEqual(lowercase , [31_957, 250, 1_345, 9, 10, 4_758] )
634
import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizerOutput from transformers.testing_utils import require_phonemizer from ...test_tokenization_common import TokenizerTesterMixin @require_phonemizer class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = WavaVecaPhonemeCTCTokenizer lowercase__ : Optional[int] = False def snake_case__ ( self : str ) -> int: """simple docstring""" super().setUp() __lowercase = ( """<s> <pad> </s> <unk> n s t ə l a i k d m ɛ ɾ e ɪ p o ɐ z ð f j v b ɹ ʁ ʊ iː r w ʌ u ɡ æ aɪ ʃ h ɔ ɑː """ """ŋ ɚ eɪ β uː y ɑ̃ oʊ ᵻ eː θ aʊ ts oː ɔ̃ ɣ ɜ ɑ dʒ əl x ɜː ç ʒ tʃ ɔː ɑːɹ ɛ̃ ʎ ɔːɹ ʋ aː ɕ œ ø oːɹ ɲ yː """ """ʔ iə i5 s. tɕ ?? nʲ ɛː œ̃ ɭ ɔø ʑ tʲ ɨ ɛɹ ts. rʲ ɪɹ ɭʲ i.5 ɔɪ q sʲ u5 ʊɹ iɜ a5 iɛ5 øː ʕ ja əɜ th ɑ5 """ """oɪ dʲ ə5 tɕh ts.h mʲ ɯ dʑ vʲ e̞ tʃʲ ei5 o5 onɡ5 ɑu5 iɑ5 ai5 aɪɚ kh ə1 ʐ i2 ʉ ħ t[ aɪə ʲ ju ə2 u2 oɜ """ """pː iɛɜ ou5 y5 uɜ tː uo5 d[ uoɜ tsh ɑɜ ɵ i̪5 uei5 ɟ aɜ ɑɨ i.ɜ eʊ o2 ɐ̃ ä pʲ kʲ n̩ ɒ ph ɑu2 uɨ əɪ ɫ ɬ """ """yɜ bʲ ɑ2 s̪ aiɜ χ ɐ̃ʊ̃ 1 ə4 yæɜ a2 ɨː t̪ iouɜ ũ onɡɜ aɨ iɛ2 ɔɨ ɑuɜ o̞ ei2 iou2 c kː y2 ɖ oe dˤ yɛɜ """ """əʊ S ɡʲ onɡ2 u\" eiɜ ʈ ɯᵝ iou5 dZ r̝̊ i.2 tS s^ ʝ yə5 iɑɜ uə5 pf ɨu iɑ2 ou2 ər2 fʲ ai2 r̝ uəɜ ɳ əɨ """ """ua5 uɪ ɽ bː yu5 uo2 yɛ5 l̩ ɻ ərɜ ʂ i̪2 ouɜ uaɜ a. a.ː yæ5 dː r̩ ee ɪu ər5 i̪ ɜ æi u: i.ː t^ o1 ɪ^ """ """ai ueiɜ æː ɛɪ eə i. ɴ ie ua2 ɑ1 o4 tʃː o: ɑ: u1 N i̪1 au yæ2 u. qː yəɜ y: kʰ tʃʰ iʊ sx õ uo tʰ """ """uai5 bʰ u.ː uə2 ʊə d^ s̪ː yiɜ dʰ r. oe: i1 ɟː yu2 nʲʲ i̪4 uei2 tsʲ ɸ ĩ ɑ4 t̪ː eɑ u4 e: tsː ʈʰ ɡʰ """ """ɯɯ dʒʲ ʂʲ X ɵː uaiɜ tɕʲ ã t^ː ẽː yɛ2 cː i.1 ɛʊ dˤdˤ dʒː i4 ɡː yi ɕʲ ɟʰ pʰ dʑʲ yuɜ ua1 ua4 æiː ɐɐ """ """ui iou1 ʊː a1 iou4 cʰ iɛ1 yə2 ɖʰ ẽ ʒʲ ää ər4 iːː ɪː iɑ1 ər1 œː øi ɪuː cʰcʰ əː1 iː1 ũ kʰː o̞o̞ xʲ """ """ou1 iɛ4 e̞e̞ y1 dzː dʲʲ dʰː ɯᵝɯᵝ lː uo1 i.4 i: yɛ5ʲ a4""" ).split(""" """ ) __lowercase = dict(zip(lowercase , range(len(lowercase ) ) ) ) __lowercase = {"""pad_token""": """<pad>""", """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>"""} __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowercase ) + """\n""" ) def snake_case__ ( self : List[Any] , lowercase : Optional[Any] , lowercase : List[str]=False , lowercase : List[str]=20 , lowercase : str=5 ) -> Tuple[str, list]: """simple docstring""" __lowercase = [(i, tokenizer.decode([i] , clean_up_tokenization_spaces=lowercase )) for i in range(len(lowercase ) )] __lowercase = list(filter(lambda lowercase : [t[0]] == tokenizer.encode(t[1] , do_phonemize=lowercase ) , lowercase ) ) if max_length is not None and len(lowercase ) > max_length: __lowercase = toks[:max_length] if min_length is not None and len(lowercase ) < min_length and len(lowercase ) > 0: while len(lowercase ) < min_length: __lowercase = toks + toks # toks_str = [t[1] for t in toks] __lowercase = [t[0] for t in toks] # Ensure consistency __lowercase = tokenizer.decode(lowercase , clean_up_tokenization_spaces=lowercase ) if " " not in output_txt and len(lowercase ) > 1: __lowercase = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=lowercase ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=lowercase ) ) if with_prefix_space: __lowercase = """ """ + output_txt __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) return output_txt, output_ids def snake_case__ ( self : Tuple , **lowercase : int ) -> Union[str, Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return WavaVecaPhonemeCTCTokenizer.from_pretrained(self.tmpdirname , **lowercase ) def snake_case__ ( self : Any ) -> List[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) # check adding a single token tokenizer.add_tokens("""xxx""" ) __lowercase = tokenizer("""m xxx ɪ""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 392, 17] ) # xxx should be last token tokenizer.add_tokens(["""aaa""", """bbb""", """ccc"""] ) __lowercase = tokenizer("""m aaa ɪ ccc""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 393, 17, 395] ) # aaa and ccc should be after xxx and 2 after aaa __lowercase = tokenizer("""maɪ c""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [3, 200] ) # mai should be <unk> (=3) def snake_case__ ( self : int ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) def snake_case__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98], [24, 22, 5, 24, 22, 5, 77], ] __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> Optional[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ | h aʊ | ɑːɹ | j uː |""" ) def snake_case__ ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : List[Any] ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, tokenizer.word_delimiter_token_id, 15, 8, tokenizer.word_delimiter_token_id, 98], [tokenizer.word_delimiter_token_id, 24, 22, tokenizer.word_delimiter_token_id, 5, 24, 22, 5, 77], ] # fmt: on # decode with word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) # decode with no word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] , filter_word_delimiter_token=lowercase ) __lowercase = tokenizer.batch_decode(lowercase , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ | ɾ l | ɭʲ""", """| j ð | s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : Optional[int] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(""" """.join([p.strip() for p in phonemes.split(""" |""" )] ).strip() , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token=lowercase ) __lowercase = """Hello how are you""" __lowercase = tokenizer(lowercase , phonemizer_lang="""en-us""" ).input_ids __lowercase = tokenizer(lowercase , phonemizer_lang="""fr-fr""" ).input_ids self.assertNotEqual(lowercase , lowercase ) __lowercase = tokenizer.decode(lowercase ) __lowercase = tokenizer.decode(lowercase ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) self.assertEqual(lowercase , """ɛ l o h aʊ a ʁ j u""" ) def snake_case__ ( self : int ) -> int: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how Are you""" __lowercase = """hello how are you""" __lowercase = tokenizer(lowercase ).input_ids __lowercase = tokenizer(lowercase ).input_ids self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : int ) -> Union[str, Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) tokenizer.add_tokens(["""!""", """?"""] ) tokenizer.add_special_tokens({"""cls_token""": """$$$"""} ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98, 392, 392, 393, 392, 392, 393, 394, 394], [24, 22, 5, 24, 22, 5, 77, tokenizer.pad_token_id, 394, 394], ] # fmt: on __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ!?!? $$$""", """j ð s j ð s oːɹ $$$"""] ) @staticmethod def snake_case__ ( lowercase : List[str] , lowercase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = [d[key] for d in offsets] return retrieved_list def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off # ksssɾɾ|ɾɾ<pad>ɾɾ|<pad>ɾlll|ɭʲ -> k s ɾ ɾ | ɾ l | ɭʲ" __lowercase = [11, 5, 5, 5, 15, 15, tokenizer.pad_token_id, 15, 15, tokenizer.word_delimiter_token_id, tokenizer.pad_token_id, 15, 8, 8, 8, tokenizer.word_delimiter_token_id, 98] # fmt: on __lowercase = tokenizer.decode(lowercase , output_char_offsets=lowercase , filter_word_delimiter_token=lowercase ) # check Wav2Vec2CTCTokenizerOutput keys for char self.assertEqual(len(outputs.keys() ) , 2 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""char_offsets""" in outputs ) self.assertTrue(isinstance(lowercase , lowercase ) ) # check that order of chars is correct and identical for both outputs self.assertEqual(""" """.join(self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) ) , outputs.text ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) , ["""k""", """s""", """ɾ""", """ɾ""", """|""", """ɾ""", """l""", """|""", """ɭʲ"""] ) # check that offsets are actually correct for char # 0-1 is 11, 1-4 is 5, 4-6 is first 15, 6-7 is <pad> (thus not shown), 7-9 is second 15, 9-10 is word_delimiter_token, # 10-11 is <pad> (thus not shown), 11-12 is third 15, 12-15 is 8, 15-16 is word_delimiter_token, 16-17 is 98 self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """start_offset""" ) , [0, 1, 4, 7, 9, 11, 12, 15, 16] ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """end_offset""" ) , [1, 4, 6, 9, 10, 12, 15, 16, 17] ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) def check_list_tuples_equal(lowercase : List[str] , lowercase : Dict ): self.assertTrue(isinstance(lowercase , lowercase ) ) self.assertTrue(isinstance(outputs_list[0] , lowercase ) ) # transform list to ModelOutput __lowercase = WavaVecaPhonemeCTCTokenizerOutput( {k: [d[k] for d in outputs_list] for k in outputs_list[0]} ) self.assertListEqual(outputs_batch["""text"""] , outputs_batch_a["""text"""] ) def recursive_check(lowercase : List[Any] , lowercase : Optional[int] ): if isinstance(lowercase , lowercase ): [recursive_check(lowercase , lowercase ) for la, la in zip(lowercase , lowercase )] self.assertEqual(lowercase , lowercase ) if "char_offsets" in outputs_batch: recursive_check(outputs_batch["""char_offsets"""] , outputs_batch_a["""char_offsets"""] ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 4, 8, 98, 32, 32, 32, 32, 4, 33, tokenizer.word_delimiter_token_id, 32, 32, 33, 34, 34], [24, 22, 5, tokenizer.word_delimiter_token_id, tokenizer.word_delimiter_token_id, 24, 22, 22, 22, 4, 5, 77, tokenizer.pad_token_id, 22, 22, 4, 34, 34, 34, 34], ] # fmt: on # We assume that `decode` works as expected. All we will check now is # the output type is correct and the output is identical to `decode` # char __lowercase = tokenizer.batch_decode(lowercase , output_char_offsets=lowercase ) __lowercase = [tokenizer.decode(lowercase , output_char_offsets=lowercase ) for ids in sample_ids] check_list_tuples_equal(lowercase , lowercase ) @unittest.skip("""Wav2Vec2PhonemeTokenizer always lower cases letters to correctly map to phonemes""" ) def snake_case__ ( self : List[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeTokenizer always puts spaces between phonemes""" ) def snake_case__ ( self : List[str] ) -> List[str]: """simple docstring""" pass @unittest.skip("""encodes to text to ids, but decodes ids to phonemes -> not possible to have internal consistency""" ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeModel has no max model length => no testing""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass def snake_case__ ( self : Any ) -> int: """simple docstring""" __lowercase = self.get_tokenizers(do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) __lowercase = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] __lowercase = tokenizer.add_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size + len(lowercase ) ) __lowercase = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) __lowercase = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} __lowercase = tokenizer.add_special_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size_a + len(lowercase ) ) __lowercase = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" pass def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizers(fast=lowercase , do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = ["""ð""", """ɪ""", """s""", """ɪ""", """z""", """ɐ""", """t""", """ɛ""", """k""", """s""", """t"""] __lowercase = tokenizer.convert_tokens_to_string(lowercase ) self.assertIsInstance(output["""text"""] , lowercase )
634
1
def UpperCAmelCase__ ( lowercase__ ) -> Tuple: if not head: return True # split the list to two parts __lowercase , __lowercase = head.next, head while fast and fast.next: __lowercase = fast.next.next __lowercase = slow.next __lowercase = slow.next __lowercase = None # Don't forget here! But forget still works! # reverse the second part __lowercase = None while second: __lowercase = second.next __lowercase = node __lowercase = second __lowercase = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False __lowercase = node.next __lowercase = head.next return True def UpperCAmelCase__ ( lowercase__ ) -> str: if not head or not head.next: return True # 1. Get the midpoint (slow) __lowercase = __lowercase = __lowercase = head while fast and fast.next: __lowercase , __lowercase = fast.next.next, slow.next # 2. Push the second half into the stack __lowercase = [slow.val] while slow.next: __lowercase = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False __lowercase = cur.next return True def UpperCAmelCase__ ( lowercase__ ) -> Optional[int]: if not head or not head.next: return True __lowercase = {} __lowercase = 0 while head: if head.val in d: d[head.val].append(SCREAMING_SNAKE_CASE_ ) else: __lowercase = [pos] __lowercase = head.next pos += 1 __lowercase = pos - 1 __lowercase = 0 for v in d.values(): if len(SCREAMING_SNAKE_CASE_ ) % 2 != 0: middle += 1 else: __lowercase = 0 for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) ): if v[i] + v[len(SCREAMING_SNAKE_CASE_ ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
700
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) # TODO Update this UpperCamelCase__ = { "facebook/esm-1b": "https://huggingface.co/facebook/esm-1b/resolve/main/config.json", # See all ESM models at https://huggingface.co/models?filter=esm } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """esm""" def __init__( self : Any , lowercase : Optional[Any]=None , lowercase : Optional[int]=None , lowercase : List[Any]=None , lowercase : Optional[int]=768 , lowercase : str=12 , lowercase : Union[str, Any]=12 , lowercase : Dict=3_072 , lowercase : Optional[int]=0.1 , lowercase : str=0.1 , lowercase : Dict=1_026 , lowercase : Tuple=0.02 , lowercase : str=1E-1_2 , lowercase : Dict="absolute" , lowercase : Optional[Any]=True , lowercase : int=None , lowercase : int=False , lowercase : List[str]=False , lowercase : Tuple=None , lowercase : Tuple=None , **lowercase : Union[str, Any] , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase , mask_token_id=lowercase , **lowercase ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = emb_layer_norm_before __lowercase = token_dropout __lowercase = is_folding_model if is_folding_model: if esmfold_config is None: logger.info("""No esmfold_config supplied for folding model, using default values.""" ) __lowercase = EsmFoldConfig() elif isinstance(lowercase , lowercase ): __lowercase = EsmFoldConfig(**lowercase ) __lowercase = esmfold_config if vocab_list is None: logger.warning("""No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!""" ) __lowercase = get_default_vocab_list() else: __lowercase = vocab_list else: __lowercase = None __lowercase = None if self.esmfold_config is not None and getattr(self.esmfold_config , """use_esm_attn_map""" , lowercase ): raise ValueError("""The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!""" ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = super().to_dict() if isinstance(self.esmfold_config , lowercase ): __lowercase = self.esmfold_config.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : str = None lowercase__ : bool = True lowercase__ : bool = False lowercase__ : bool = False lowercase__ : bool = False lowercase__ : float = 0 lowercase__ : bool = True lowercase__ : bool = False lowercase__ : int = 128 lowercase__ : "TrunkConfig" = None def snake_case__ ( self : List[str] ) -> Any: """simple docstring""" if self.trunk is None: __lowercase = TrunkConfig() elif isinstance(self.trunk , lowercase ): __lowercase = TrunkConfig(**self.trunk ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.trunk.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 48 lowercase__ : int = 1_024 lowercase__ : int = 128 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : float = 0 lowercase__ : float = 0 lowercase__ : bool = False lowercase__ : int = 4 lowercase__ : Optional[int] = 128 lowercase__ : "StructureModuleConfig" = None def snake_case__ ( self : Tuple ) -> str: """simple docstring""" if self.structure_module is None: __lowercase = StructureModuleConfig() elif isinstance(self.structure_module , lowercase ): __lowercase = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(F"`max_recycles` should be positive, got {self.max_recycles}." ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( """`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got""" F" {self.sequence_state_dim} and {self.sequence_state_dim}." ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( """`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got""" F" {self.pairwise_state_dim} and {self.pairwise_state_dim}." ) __lowercase = self.sequence_state_dim // self.sequence_head_width __lowercase = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( """`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got""" F" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}." ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( """`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got""" F" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}." ) if self.pairwise_state_dim % 2 != 0: raise ValueError(F"`pairwise_state_dim` should be even, got {self.pairwise_state_dim}." ) if self.dropout >= 0.4: raise ValueError(F"`dropout` should not be greater than 0.4, got {self.dropout}." ) def snake_case__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.structure_module.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 384 lowercase__ : int = 128 lowercase__ : int = 16 lowercase__ : int = 128 lowercase__ : int = 12 lowercase__ : int = 4 lowercase__ : int = 8 lowercase__ : float = 0.1 lowercase__ : int = 8 lowercase__ : int = 1 lowercase__ : int = 2 lowercase__ : int = 7 lowercase__ : int = 10 lowercase__ : float = 1E-8 lowercase__ : float = 1E5 def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" return asdict(self ) def UpperCAmelCase__ ( ) -> List[Any]: return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
634
0
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> int: # Load checkpoint __lowercase = torch.load(__A , map_location="""cpu""" ) __lowercase = chkpt["""model"""] # We have the base model one level deeper than the original XLM repository __lowercase = {} for k, v in state_dict.items(): if "pred_layer" in k: __lowercase = v else: __lowercase = v __lowercase = chkpt["""params"""] __lowercase = {n: v for n, v in config.items() if not isinstance(__A , (torch.FloatTensor, numpy.ndarray) )} __lowercase = chkpt["""dico_word2id"""] __lowercase = {s + """</w>""" if s.find("""@@""" ) == -1 and i > 13 else s.replace("""@@""" , """""" ): i for s, i in vocab.items()} # Save pytorch-model __lowercase = pytorch_dump_folder_path + """/""" + WEIGHTS_NAME __lowercase = pytorch_dump_folder_path + """/""" + CONFIG_NAME __lowercase = pytorch_dump_folder_path + """/""" + VOCAB_FILES_NAMES["""vocab_file"""] print(F"Save PyTorch model to {pytorch_weights_dump_path}" ) torch.save(__A , __A ) print(F"Save configuration file to {pytorch_config_dump_path}" ) with open(__A , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(__A , indent=2 ) + """\n""" ) print(F"Save vocab file to {pytorch_config_dump_path}" ) with open(__A , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(__A , indent=2 ) + """\n""" ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--xlm_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) UpperCamelCase__ = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
701
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Tuple = LxmertTokenizer lowercase__ : List[str] = LxmertTokenizerFast lowercase__ : Optional[Any] = True lowercase__ : List[Any] = True def snake_case__ ( self : Tuple ) -> Tuple: """simple docstring""" super().setUp() __lowercase = [ """[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def snake_case__ ( self : Optional[int] , lowercase : int ) -> List[Any]: """simple docstring""" __lowercase = """UNwant\u00E9d,running""" __lowercase = """unwanted, running""" return input_text, output_text def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class(self.vocab_file ) __lowercase = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(lowercase , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , [7, 4, 5, 10, 8, 9] ) def snake_case__ ( self : Union[str, Any] ) -> int: """simple docstring""" if not self.test_rust_tokenizer: return __lowercase = self.get_tokenizer() __lowercase = self.get_rust_tokenizer() __lowercase = """I was born in 92000, and this is falsé.""" __lowercase = tokenizer.tokenize(lowercase ) __lowercase = rust_tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) __lowercase = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = self.get_rust_tokenizer() __lowercase = tokenizer.encode(lowercase ) __lowercase = rust_tokenizer.encode(lowercase ) self.assertListEqual(lowercase , lowercase )
634
0
import logging import os from .state import PartialState class _lowerCAmelCase ( logging.LoggerAdapter ): """simple docstring""" @staticmethod def snake_case__ ( lowercase : List[str] ) -> List[Any]: """simple docstring""" __lowercase = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def snake_case__ ( self : List[str] , lowercase : Any , lowercase : List[Any] , *lowercase : Any , **lowercase : int ) -> str: """simple docstring""" if PartialState._shared_state == {}: raise RuntimeError( """You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility.""" ) __lowercase = kwargs.pop("""main_process_only""" , UpperCAmelCase_ ) __lowercase = kwargs.pop("""in_order""" , UpperCAmelCase_ ) if self.isEnabledFor(UpperCAmelCase_ ): if self._should_log(UpperCAmelCase_ ): __lowercase , __lowercase = self.process(UpperCAmelCase_ , UpperCAmelCase_ ) self.logger.log(UpperCAmelCase_ , UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_ ) elif in_order: __lowercase = PartialState() for i in range(state.num_processes ): if i == state.process_index: __lowercase , __lowercase = self.process(UpperCAmelCase_ , UpperCAmelCase_ ) self.logger.log(UpperCAmelCase_ , UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_ ) state.wait_for_everyone() def UpperCAmelCase__ ( lowercase__ , lowercase__ = None ): if log_level is None: __lowercase = os.environ.get("""ACCELERATE_LOG_LEVEL""" , _SCREAMING_SNAKE_CASE ) __lowercase = logging.getLogger(_SCREAMING_SNAKE_CASE ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(_SCREAMING_SNAKE_CASE , {} )
702
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> bool: __lowercase = len(lowercase__ ) __lowercase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): __lowercase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): __lowercase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: __lowercase = subset[i - 1][j] if arr[i - 1] <= j: __lowercase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
634
0
import colorsys from PIL import Image # type: ignore def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: __lowercase = x __lowercase = y for step in range(lowercase__ ): # noqa: B007 __lowercase = a * a - b * b + x __lowercase = 2 * a * b + y __lowercase = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def UpperCAmelCase__ ( lowercase__ ) -> tuple: if distance == 1: return (0, 0, 0) else: return (255, 255, 255) def UpperCAmelCase__ ( lowercase__ ) -> tuple: if distance == 1: return (0, 0, 0) else: return tuple(round(i * 255 ) for i in colorsys.hsv_to_rgb(lowercase__ , 1 , 1 ) ) def UpperCAmelCase__ ( lowercase__ = 800 , lowercase__ = 600 , lowercase__ = -0.6 , lowercase__ = 0 , lowercase__ = 3.2 , lowercase__ = 50 , lowercase__ = True , ) -> Image.Image: __lowercase = Image.new("""RGB""" , (image_width, image_height) ) __lowercase = img.load() # loop through the image-coordinates for image_x in range(lowercase__ ): for image_y in range(lowercase__ ): # determine the figure-coordinates based on the image-coordinates __lowercase = figure_width / image_width * image_height __lowercase = figure_center_x + (image_x / image_width - 0.5) * figure_width __lowercase = figure_center_y + (image_y / image_height - 0.5) * figure_height __lowercase = get_distance(lowercase__ , lowercase__ , lowercase__ ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: __lowercase = get_color_coded_rgb(lowercase__ ) else: __lowercase = get_black_and_white_rgb(lowercase__ ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure UpperCamelCase__ = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
703
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "hustvl/yolos-small": "https://huggingface.co/hustvl/yolos-small/resolve/main/config.json", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """yolos""" def __init__( self : Optional[int] , lowercase : Any=768 , lowercase : Tuple=12 , lowercase : Tuple=12 , lowercase : str=3_072 , lowercase : Optional[Any]="gelu" , lowercase : Union[str, Any]=0.0 , lowercase : Dict=0.0 , lowercase : Optional[int]=0.02 , lowercase : Optional[Any]=1E-1_2 , lowercase : Tuple=[512, 864] , lowercase : Optional[int]=16 , lowercase : Dict=3 , lowercase : Optional[Any]=True , lowercase : Optional[int]=100 , lowercase : Optional[int]=True , lowercase : Any=False , lowercase : Any=1 , lowercase : Any=5 , lowercase : List[str]=2 , lowercase : Union[str, Any]=5 , lowercase : str=2 , lowercase : Tuple=0.1 , **lowercase : str , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase ) __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = qkv_bias __lowercase = num_detection_tokens __lowercase = use_mid_position_embeddings __lowercase = auxiliary_loss # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = eos_coefficient class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Dict = version.parse("""1.11""" ) @property def snake_case__ ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def snake_case__ ( self : int ) -> float: """simple docstring""" return 1E-4 @property def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" return 12
634
0
import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...image_processing_utils import ImageProcessingMixin from ...utils import CONFIG_NAME, IMAGE_PROCESSOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = OrderedDict( [ ("align", "EfficientNetImageProcessor"), ("beit", "BeitImageProcessor"), ("bit", "BitImageProcessor"), ("blip", "BlipImageProcessor"), ("blip-2", "BlipImageProcessor"), ("bridgetower", "BridgeTowerImageProcessor"), ("chinese_clip", "ChineseCLIPImageProcessor"), ("clip", "CLIPImageProcessor"), ("clipseg", "ViTImageProcessor"), ("conditional_detr", "ConditionalDetrImageProcessor"), ("convnext", "ConvNextImageProcessor"), ("convnextv2", "ConvNextImageProcessor"), ("cvt", "ConvNextImageProcessor"), ("data2vec-vision", "BeitImageProcessor"), ("deformable_detr", "DeformableDetrImageProcessor"), ("deit", "DeiTImageProcessor"), ("deta", "DetaImageProcessor"), ("detr", "DetrImageProcessor"), ("dinat", "ViTImageProcessor"), ("donut-swin", "DonutImageProcessor"), ("dpt", "DPTImageProcessor"), ("efficientformer", "EfficientFormerImageProcessor"), ("efficientnet", "EfficientNetImageProcessor"), ("flava", "FlavaImageProcessor"), ("focalnet", "BitImageProcessor"), ("git", "CLIPImageProcessor"), ("glpn", "GLPNImageProcessor"), ("groupvit", "CLIPImageProcessor"), ("imagegpt", "ImageGPTImageProcessor"), ("instructblip", "BlipImageProcessor"), ("layoutlmv2", "LayoutLMv2ImageProcessor"), ("layoutlmv3", "LayoutLMv3ImageProcessor"), ("levit", "LevitImageProcessor"), ("mask2former", "Mask2FormerImageProcessor"), ("maskformer", "MaskFormerImageProcessor"), ("mgp-str", "ViTImageProcessor"), ("mobilenet_v1", "MobileNetV1ImageProcessor"), ("mobilenet_v2", "MobileNetV2ImageProcessor"), ("mobilevit", "MobileViTImageProcessor"), ("mobilevit", "MobileViTImageProcessor"), ("mobilevitv2", "MobileViTImageProcessor"), ("nat", "ViTImageProcessor"), ("oneformer", "OneFormerImageProcessor"), ("owlvit", "OwlViTImageProcessor"), ("perceiver", "PerceiverImageProcessor"), ("pix2struct", "Pix2StructImageProcessor"), ("poolformer", "PoolFormerImageProcessor"), ("regnet", "ConvNextImageProcessor"), ("resnet", "ConvNextImageProcessor"), ("sam", "SamImageProcessor"), ("segformer", "SegformerImageProcessor"), ("swiftformer", "ViTImageProcessor"), ("swin", "ViTImageProcessor"), ("swin2sr", "Swin2SRImageProcessor"), ("swinv2", "ViTImageProcessor"), ("table-transformer", "DetrImageProcessor"), ("timesformer", "VideoMAEImageProcessor"), ("tvlt", "TvltImageProcessor"), ("upernet", "SegformerImageProcessor"), ("van", "ConvNextImageProcessor"), ("videomae", "VideoMAEImageProcessor"), ("vilt", "ViltImageProcessor"), ("vit", "ViTImageProcessor"), ("vit_hybrid", "ViTHybridImageProcessor"), ("vit_mae", "ViTImageProcessor"), ("vit_msn", "ViTImageProcessor"), ("xclip", "CLIPImageProcessor"), ("yolos", "YolosImageProcessor"), ] ) UpperCamelCase__ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, IMAGE_PROCESSOR_MAPPING_NAMES) def UpperCAmelCase__ ( lowercase__ ) -> str: for module_name, extractors in IMAGE_PROCESSOR_MAPPING_NAMES.items(): if class_name in extractors: __lowercase = model_type_to_module_name(_lowercase ) __lowercase = importlib.import_module(F".{module_name}" , """transformers.models""" ) try: return getattr(_lowercase , _lowercase ) except AttributeError: continue for _, extractor in IMAGE_PROCESSOR_MAPPING._extra_content.items(): if getattr(_lowercase , """__name__""" , _lowercase ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. __lowercase = importlib.import_module("""transformers""" ) if hasattr(_lowercase , _lowercase ): return getattr(_lowercase , _lowercase ) return None def UpperCAmelCase__ ( lowercase__ , lowercase__ = None , lowercase__ = False , lowercase__ = False , lowercase__ = None , lowercase__ = None , lowercase__ = None , lowercase__ = False , **lowercase__ , ) -> List[Any]: __lowercase = get_file_from_repo( _lowercase , _lowercase , cache_dir=_lowercase , force_download=_lowercase , resume_download=_lowercase , proxies=_lowercase , use_auth_token=_lowercase , revision=_lowercase , local_files_only=_lowercase , ) if resolved_config_file is None: logger.info( """Could not locate the image processor configuration file, will try to use the model config instead.""" ) return {} with open(_lowercase , encoding="""utf-8""" ) as reader: return json.load(_lowercase ) class _lowerCAmelCase : """simple docstring""" def __init__( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" raise EnvironmentError( """AutoImageProcessor is designed to be instantiated """ """using the `AutoImageProcessor.from_pretrained(pretrained_model_name_or_path)` method.""" ) @classmethod @replace_list_option_in_docstrings(UpperCamelCase_ ) def snake_case__ ( cls : Tuple , lowercase : str , **lowercase : List[Any] ) -> List[str]: """simple docstring""" __lowercase = kwargs.pop("""config""" , UpperCamelCase_ ) __lowercase = kwargs.pop("""trust_remote_code""" , UpperCamelCase_ ) __lowercase = True __lowercase = ImageProcessingMixin.get_image_processor_dict(UpperCamelCase_ , **UpperCamelCase_ ) __lowercase = config_dict.get("""image_processor_type""" , UpperCamelCase_ ) __lowercase = None if "AutoImageProcessor" in config_dict.get("""auto_map""" , {} ): __lowercase = config_dict['auto_map']['AutoImageProcessor'] # If we still don't have the image processor class, check if we're loading from a previous feature extractor config # and if so, infer the image processor class from there. if image_processor_class is None and image_processor_auto_map is None: __lowercase = config_dict.pop("""feature_extractor_type""" , UpperCamelCase_ ) if feature_extractor_class is not None: logger.warning( """Could not find image processor class in the image processor config or the model config. Loading""" """ based on pattern matching with the model\'s feature extractor configuration.""" ) __lowercase = feature_extractor_class.replace("""FeatureExtractor""" , """ImageProcessor""" ) if "AutoFeatureExtractor" in config_dict.get("""auto_map""" , {} ): __lowercase = config_dict['auto_map']['AutoFeatureExtractor'] __lowercase = feature_extractor_auto_map.replace("""FeatureExtractor""" , """ImageProcessor""" ) logger.warning( """Could not find image processor auto map in the image processor config or the model config.""" """ Loading based on pattern matching with the model\'s feature extractor configuration.""" ) # If we don't find the image processor class in the image processor config, let's try the model config. if image_processor_class is None and image_processor_auto_map is None: if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): __lowercase = AutoConfig.from_pretrained(UpperCamelCase_ , **UpperCamelCase_ ) # It could be in `config.image_processor_type`` __lowercase = getattr(UpperCamelCase_ , """image_processor_type""" , UpperCamelCase_ ) if hasattr(UpperCamelCase_ , """auto_map""" ) and "AutoImageProcessor" in config.auto_map: __lowercase = config.auto_map['AutoImageProcessor'] if image_processor_class is not None: __lowercase = image_processor_class_from_name(UpperCamelCase_ ) __lowercase = image_processor_auto_map is not None __lowercase = image_processor_class is not None or type(UpperCamelCase_ ) in IMAGE_PROCESSOR_MAPPING __lowercase = resolve_trust_remote_code( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) if has_remote_code and trust_remote_code: __lowercase = get_class_from_dynamic_module( UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_ ) __lowercase = kwargs.pop("""code_revision""" , UpperCamelCase_ ) if os.path.isdir(UpperCamelCase_ ): image_processor_class.register_for_auto_class() return image_processor_class.from_dict(UpperCamelCase_ , **UpperCamelCase_ ) elif image_processor_class is not None: return image_processor_class.from_dict(UpperCamelCase_ , **UpperCamelCase_ ) # Last try: we use the IMAGE_PROCESSOR_MAPPING. elif type(UpperCamelCase_ ) in IMAGE_PROCESSOR_MAPPING: __lowercase = IMAGE_PROCESSOR_MAPPING[type(UpperCamelCase_ )] return image_processor_class.from_dict(UpperCamelCase_ , **UpperCamelCase_ ) raise ValueError( F"Unrecognized image processor in {pretrained_model_name_or_path}. Should have a " F"`image_processor_type` key in its {IMAGE_PROCESSOR_NAME} of {CONFIG_NAME}, or one of the following " F"`model_type` keys in its {CONFIG_NAME}: {', '.join(c for c in IMAGE_PROCESSOR_MAPPING_NAMES.keys() )}" ) @staticmethod def snake_case__ ( lowercase : Dict , lowercase : int ) -> List[str]: """simple docstring""" IMAGE_PROCESSOR_MAPPING.register(UpperCamelCase_ , UpperCamelCase_ )
704
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = IFImgaImgSuperResolutionPipeline lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) lowercase__ : Tuple = PipelineTesterMixin.required_optional_params - {"""latents"""} def snake_case__ ( self : Tuple ) -> Any: """simple docstring""" return self._get_superresolution_dummy_components() def snake_case__ ( self : List[str] , lowercase : Optional[int] , lowercase : Optional[Any]=0 ) -> Union[str, Any]: """simple docstring""" if str(lowercase ).startswith("""mps""" ): __lowercase = torch.manual_seed(lowercase ) else: __lowercase = torch.Generator(device=lowercase ).manual_seed(lowercase ) __lowercase = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """original_image""": original_image, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def snake_case__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1E-1 ) def snake_case__ ( self : Dict ) -> int: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" self._test_save_load_local() def snake_case__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
634
0
from __future__ import annotations def UpperCAmelCase__ ( lowercase__ ) -> Tuple: return [ord(_A ) - 96 for elem in plain] def UpperCAmelCase__ ( lowercase__ ) -> int: return "".join(chr(elem + 96 ) for elem in encoded ) def UpperCAmelCase__ ( ) -> int: __lowercase = encode(input("""-> """ ).strip().lower() ) print("""Encoded: """ , _A ) print("""Decoded:""" , decode(_A ) ) if __name__ == "__main__": main()
705
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ = { "configuration_luke": ["LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP", "LukeConfig"], "tokenization_luke": ["LukeTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "LUKE_PRETRAINED_MODEL_ARCHIVE_LIST", "LukeForEntityClassification", "LukeForEntityPairClassification", "LukeForEntitySpanClassification", "LukeForMultipleChoice", "LukeForQuestionAnswering", "LukeForSequenceClassification", "LukeForTokenClassification", "LukeForMaskedLM", "LukeModel", "LukePreTrainedModel", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
0
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=__lowerCamelCase ) class _lowerCAmelCase ( __lowerCamelCase ): """simple docstring""" lowercase__ : Optional[Any] = field(default="""text-classification""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowercase__ : Union[str, Any] = Features({"""text""": Value("""string""" )} ) lowercase__ : List[str] = Features({"""labels""": ClassLabel} ) lowercase__ : Union[str, Any] = """text""" lowercase__ : Union[str, Any] = """labels""" def snake_case__ ( self : List[Any] , lowercase : Union[str, Any] ) -> int: """simple docstring""" if self.label_column not in features: raise ValueError(F"Column {self.label_column} is not present in features." ) if not isinstance(features[self.label_column] , UpperCAmelCase_ ): raise ValueError(F"Column {self.label_column} is not a ClassLabel." ) __lowercase = copy.deepcopy(self ) __lowercase = self.label_schema.copy() __lowercase = features[self.label_column] __lowercase = label_schema return task_template @property def snake_case__ ( self : Optional[int] ) -> Dict: """simple docstring""" return { self.text_column: "text", self.label_column: "labels", }
706
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> int: if len(lowercase__ ) != len(lowercase__ ): raise ValueError("""String lengths must match!""" ) __lowercase = 0 for chara, chara in zip(lowercase__ , lowercase__ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
634
0
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 UpperCamelCase__ = get_tests_dir("fixtures") UpperCamelCase__ = get_tests_dir("fixtures/dummy_feature_extractor_config.json") UpperCamelCase__ = get_tests_dir("fixtures/dummy-config.json") class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : List[Any] ) -> List[str]: """simple docstring""" __lowercase = 0 def snake_case__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = AutoFeatureExtractor.from_pretrained("""facebook/wav2vec2-base-960h""" ) self.assertIsInstance(_lowercase , _lowercase ) def snake_case__ ( self : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = AutoFeatureExtractor.from_pretrained(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def snake_case__ ( self : int ) -> Dict: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally __lowercase = AutoFeatureExtractor.from_pretrained(_lowercase ).to_dict() config_dict.pop("""feature_extractor_type""" ) __lowercase = WavaVecaFeatureExtractor(**_lowercase ) # save in new folder model_config.save_pretrained(_lowercase ) config.save_pretrained(_lowercase ) __lowercase = AutoFeatureExtractor.from_pretrained(_lowercase ) # make sure private variable is not incorrectly saved __lowercase = json.loads(config.to_json_string() ) self.assertTrue("""_processor_class""" not in dict_as_saved ) self.assertIsInstance(_lowercase , _lowercase ) def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = AutoFeatureExtractor.from_pretrained(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" with self.assertRaisesRegex( _lowercase , """bert-base is not a local folder and is not a valid model identifier""" ): __lowercase = AutoFeatureExtractor.from_pretrained("""bert-base""" ) def snake_case__ ( self : List[str] ) -> int: """simple docstring""" with self.assertRaisesRegex( _lowercase , r"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): __lowercase = AutoFeatureExtractor.from_pretrained(_lowercase , revision="""aaaaaa""" ) def snake_case__ ( self : Union[str, Any] ) -> Any: """simple docstring""" with self.assertRaisesRegex( _lowercase , """hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.""" , ): __lowercase = AutoFeatureExtractor.from_pretrained("""hf-internal-testing/config-no-model""" ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" with self.assertRaises(_lowercase ): __lowercase = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(_lowercase ): __lowercase = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=_lowercase ) __lowercase = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=_lowercase ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(_lowercase ) __lowercase = AutoFeatureExtractor.from_pretrained(_lowercase , trust_remote_code=_lowercase ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) def snake_case__ ( self : Any ) -> Optional[int]: """simple docstring""" try: AutoConfig.register("""custom""" , _lowercase ) AutoFeatureExtractor.register(_lowercase , _lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_lowercase ): AutoFeatureExtractor.register(_lowercase , _lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API __lowercase = CustomFeatureExtractor.from_pretrained(_lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(_lowercase ) __lowercase = AutoFeatureExtractor.from_pretrained(_lowercase ) self.assertIsInstance(_lowercase , _lowercase ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def snake_case__ ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" class _lowerCAmelCase ( UpperCAmelCase_ ): """simple docstring""" lowercase__ : Tuple = True try: AutoConfig.register("""custom""" , _lowercase ) AutoFeatureExtractor.register(_lowercase , _lowercase ) # If remote code is not set, the default is to use local __lowercase = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. __lowercase = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=_lowercase ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub __lowercase = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=_lowercase ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(not hasattr(_lowercase , """is_local""" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
707
from __future__ import annotations from collections.abc import Callable UpperCamelCase__ = list[list[float | int]] def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Matrix: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(size + 1 )] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for row in range(lowercase__ ): for col in range(lowercase__ ): __lowercase = matrix[row][col] __lowercase = vector[row][0] __lowercase = 0 __lowercase = 0 while row < size and col < size: # pivoting __lowercase = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase__ , lowercase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: __lowercase , __lowercase = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase__ ): __lowercase = augmented[rowa][col] / augmented[row][col] __lowercase = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase__ ): for row in range(lowercase__ ): __lowercase = augmented[row][col] / augmented[col][col] for cola in range(lowercase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase__ ) ] def UpperCAmelCase__ ( lowercase__ ) -> Callable[[int], int]: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(lowercase__ )] for _ in range(lowercase__ )] __lowercase = [[0] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for x_val, y_val in enumerate(lowercase__ ): for col in range(lowercase__ ): __lowercase = (x_val + 1) ** (size - col - 1) __lowercase = y_val __lowercase = solve(lowercase__ , lowercase__ ) def interpolated_func(lowercase__ ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase__ ) ) return interpolated_func def UpperCAmelCase__ ( lowercase__ ) -> int: return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCAmelCase__ ( lowercase__ = question_function , lowercase__ = 10 ) -> int: __lowercase = [func(lowercase__ ) for x_val in range(1 , order + 1 )] __lowercase = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] __lowercase = 0 __lowercase = 42 __lowercase = 42 for poly in polynomials: __lowercase = 1 while func(lowercase__ ) == poly(lowercase__ ): x_val += 1 ret += poly(lowercase__ ) return ret if __name__ == "__main__": print(F"""{solution() = }""")
634
0
import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class _lowerCAmelCase ( __A ): """simple docstring""" @require_torch def snake_case__ ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = ''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' __lowercase = ''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' __lowercase = ''' import socket def offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache __lowercase = '''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(lowercase ) BertModel.from_pretrained(lowercase ) BertTokenizer.from_pretrained(lowercase ) pipeline(task="""fill-mask""" , model=lowercase ) # baseline - just load from_pretrained with normal network __lowercase = [sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed __lowercase = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files __lowercase = '''1''' __lowercase = subprocess.run(lowercase , env=lowercase , check=lowercase , capture_output=lowercase ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = ''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' __lowercase = ''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' __lowercase = ''' import socket def offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache __lowercase = '''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(lowercase ) BertModel.from_pretrained(lowercase ) BertTokenizer.from_pretrained(lowercase ) pipeline(task="""fill-mask""" , model=lowercase ) # baseline - just load from_pretrained with normal network __lowercase = [sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed __lowercase = self.get_env() __lowercase = subprocess.run(lowercase , env=lowercase , check=lowercase , capture_output=lowercase ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def snake_case__ ( self : Any ) -> Optional[Any]: """simple docstring""" __lowercase = ''' from transformers import BertConfig, BertModel, BertTokenizer ''' __lowercase = ''' mname = "hf-internal-testing/tiny-random-bert-sharded" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) print("success") ''' __lowercase = ''' import socket def offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled") socket.socket = offline_socket ''' # baseline - just load from_pretrained with normal network __lowercase = [sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed __lowercase = self.get_env() __lowercase = subprocess.run(lowercase , env=lowercase , check=lowercase , capture_output=lowercase ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) # next emulate no network __lowercase = [sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files __lowercase = '''1''' __lowercase = subprocess.run(lowercase , env=lowercase , check=lowercase , capture_output=lowercase ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def snake_case__ ( self : Tuple ) -> Optional[int]: """simple docstring""" __lowercase = ''' from transformers import pipeline ''' __lowercase = ''' mname = "hf-internal-testing/tiny-random-bert" pipe = pipeline(model=mname) ''' __lowercase = ''' import socket def offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled") socket.socket = offline_socket ''' __lowercase = self.get_env() __lowercase = '''1''' __lowercase = [sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] __lowercase = subprocess.run(lowercase , env=lowercase , check=lowercase , capture_output=lowercase ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( """You cannot infer task automatically within `pipeline` when using offline mode""" , result.stderr.decode().replace("""\n""" , """""" ) , ) @require_torch def snake_case__ ( self : int ) -> Optional[int]: """simple docstring""" __lowercase = ''' from transformers import AutoModel ''' __lowercase = ''' mname = "hf-internal-testing/test_dynamic_model" AutoModel.from_pretrained(mname, trust_remote_code=True) print("success") ''' # baseline - just load from_pretrained with normal network __lowercase = [sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed __lowercase = self.get_env() __lowercase = subprocess.run(lowercase , env=lowercase , check=lowercase , capture_output=lowercase ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files __lowercase = '''1''' __lowercase = subprocess.run(lowercase , env=lowercase , check=lowercase , capture_output=lowercase ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() )
708
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ = { "configuration_ctrl": ["CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP", "CTRLConfig"], "tokenization_ctrl": ["CTRLTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "CTRLForSequenceClassification", "CTRLLMHeadModel", "CTRLModel", "CTRLPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCTRLForSequenceClassification", "TFCTRLLMHeadModel", "TFCTRLModel", "TFCTRLPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
0
import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class _lowerCAmelCase : """simple docstring""" @staticmethod def snake_case__ ( *lowercase : Optional[Any] , **lowercase : Optional[Any] ) -> List[Any]: """simple docstring""" pass @is_pipeline_test @require_vision @require_torch class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" lowercase__ : List[str] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def snake_case__ ( self : Tuple , lowercase : List[str] , lowercase : Optional[Any] , lowercase : Any ) -> Tuple: """simple docstring""" __lowercase = pipeline( """zero-shot-object-detection""" , model="""hf-internal-testing/tiny-random-owlvit-object-detection""" ) __lowercase = [ { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """candidate_labels""": ["""cat""", """remote""", """couch"""], } ] return object_detector, examples def snake_case__ ( self : Dict , lowercase : Any , lowercase : Any ) -> List[Any]: """simple docstring""" __lowercase = object_detector(examples[0] , threshold=0.0 ) __lowercase = len(_lowerCAmelCase ) self.assertGreater(_lowerCAmelCase , 0 ) self.assertEqual( _lowerCAmelCase , [ { """score""": ANY(_lowerCAmelCase ), """label""": ANY(_lowerCAmelCase ), """box""": {"""xmin""": ANY(_lowerCAmelCase ), """ymin""": ANY(_lowerCAmelCase ), """xmax""": ANY(_lowerCAmelCase ), """ymax""": ANY(_lowerCAmelCase )}, } for i in range(_lowerCAmelCase ) ] , ) @require_tf @unittest.skip("""Zero Shot Object Detection not implemented in TF""" ) def snake_case__ ( self : str ) -> Any: """simple docstring""" pass @require_torch def snake_case__ ( self : Union[str, Any] ) -> int: """simple docstring""" __lowercase = pipeline( """zero-shot-object-detection""" , model="""hf-internal-testing/tiny-random-owlvit-object-detection""" ) __lowercase = object_detector( """./tests/fixtures/tests_samples/COCO/000000039769.png""" , candidate_labels=["""cat""", """remote""", """couch"""] , threshold=0.64 , ) self.assertEqual( nested_simplify(_lowerCAmelCase , decimals=4 ) , [ {"""score""": 0.7235, """label""": """cat""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.7218, """label""": """remote""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.7184, """label""": """couch""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.6748, """label""": """remote""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6656, """label""": """cat""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6614, """label""": """couch""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6456, """label""": """remote""", """box""": {"""xmin""": 494, """ymin""": 105, """xmax""": 521, """ymax""": 127}}, {"""score""": 0.642, """label""": """remote""", """box""": {"""xmin""": 67, """ymin""": 274, """xmax""": 93, """ymax""": 297}}, {"""score""": 0.6419, """label""": """cat""", """box""": {"""xmin""": 494, """ymin""": 105, """xmax""": 521, """ymax""": 127}}, ] , ) __lowercase = object_detector( [ { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """candidate_labels""": ["""cat""", """remote""", """couch"""], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(_lowerCAmelCase , decimals=4 ) , [ [ {"""score""": 0.7235, """label""": """cat""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.7218, """label""": """remote""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.7184, """label""": """couch""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.6748, """label""": """remote""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6656, """label""": """cat""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6614, """label""": """couch""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6456, """label""": """remote""", """box""": {"""xmin""": 494, """ymin""": 105, """xmax""": 521, """ymax""": 127}}, {"""score""": 0.642, """label""": """remote""", """box""": {"""xmin""": 67, """ymin""": 274, """xmax""": 93, """ymax""": 297}}, {"""score""": 0.6419, """label""": """cat""", """box""": {"""xmin""": 494, """ymin""": 105, """xmax""": 521, """ymax""": 127}}, ] ] , ) @require_torch @slow def snake_case__ ( self : Any ) -> Any: """simple docstring""" __lowercase = pipeline("""zero-shot-object-detection""" ) __lowercase = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , ) self.assertEqual( nested_simplify(_lowerCAmelCase , decimals=4 ) , [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, {"""score""": 0.2537, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 55, """xmax""": 315, """ymax""": 472}}, {"""score""": 0.1474, """label""": """remote""", """box""": {"""xmin""": 335, """ymin""": 74, """xmax""": 371, """ymax""": 187}}, {"""score""": 0.1208, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 642, """ymax""": 476}}, ] , ) __lowercase = object_detector( [ { """image""": """http://images.cocodataset.org/val2017/000000039769.jpg""", """candidate_labels""": ["""cat""", """remote""", """couch"""], }, { """image""": """http://images.cocodataset.org/val2017/000000039769.jpg""", """candidate_labels""": ["""cat""", """remote""", """couch"""], }, ] , ) self.assertEqual( nested_simplify(_lowerCAmelCase , decimals=4 ) , [ [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, {"""score""": 0.2537, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 55, """xmax""": 315, """ymax""": 472}}, {"""score""": 0.1474, """label""": """remote""", """box""": {"""xmin""": 335, """ymin""": 74, """xmax""": 371, """ymax""": 187}}, {"""score""": 0.1208, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 642, """ymax""": 476}}, ], [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, {"""score""": 0.2537, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 55, """xmax""": 315, """ymax""": 472}}, {"""score""": 0.1474, """label""": """remote""", """box""": {"""xmin""": 335, """ymin""": 74, """xmax""": 371, """ymax""": 187}}, {"""score""": 0.1208, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 642, """ymax""": 476}}, ], ] , ) @require_tf @unittest.skip("""Zero Shot Object Detection not implemented in TF""" ) def snake_case__ ( self : Any ) -> Any: """simple docstring""" pass @require_torch @slow def snake_case__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase = 0.2 __lowercase = pipeline("""zero-shot-object-detection""" ) __lowercase = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , threshold=_lowerCAmelCase , ) self.assertEqual( nested_simplify(_lowerCAmelCase , decimals=4 ) , [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, {"""score""": 0.2537, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 55, """xmax""": 315, """ymax""": 472}}, ] , ) @require_torch @slow def snake_case__ ( self : str ) -> Tuple: """simple docstring""" __lowercase = 2 __lowercase = pipeline("""zero-shot-object-detection""" ) __lowercase = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , top_k=_lowerCAmelCase , ) self.assertEqual( nested_simplify(_lowerCAmelCase , decimals=4 ) , [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, ] , )
709
import unittest import numpy as np def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = None , ) -> np.ndarray: __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) if shape_a[0] != shape_b[0]: __lowercase = ( """Expected the same number of rows for A and B. """ F"Instead found A of size {shape_a} and B of size {shape_b}" ) raise ValueError(lowercase__ ) if shape_b[1] != shape_c[1]: __lowercase = ( """Expected the same number of columns for B and C. """ F"Instead found B of size {shape_b} and C of size {shape_c}" ) raise ValueError(lowercase__ ) __lowercase = pseudo_inv if a_inv is None: try: __lowercase = np.linalg.inv(lowercase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Dict ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) __lowercase = schur_complement(lowercase , lowercase , lowercase ) __lowercase = np.block([[a, b], [b.T, c]] ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) self.assertAlmostEqual(lowercase , det_a * det_s ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
634
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def UpperCAmelCase__ ( lowercase__ ) -> Union[str, Any]: __lowercase = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Union[str, Any] = StableDiffusionLatentUpscalePipeline lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "height", "width", "cross_attention_kwargs", "negative_prompt_embeds", "prompt_embeds", } lowercase__ : List[str] = PipelineTesterMixin.required_optional_params - {"num_images_per_prompt"} lowercase__ : Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase__ : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowercase__ : Any = frozenset([] ) lowercase__ : Optional[int] = True @property def snake_case__ ( self : List[Any] ) -> Tuple: """simple docstring""" __lowercase = 1 __lowercase = 4 __lowercase = (16, 16) __lowercase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(UpperCamelCase_ ) return image def snake_case__ ( self : Dict ) -> str: """simple docstring""" torch.manual_seed(0 ) __lowercase = UNetaDConditionModel( act_fn="""gelu""" , attention_head_dim=8 , norm_num_groups=UpperCamelCase_ , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=160 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( """KDownBlock2D""", """KCrossAttnDownBlock2D""", """KCrossAttnDownBlock2D""", """KCrossAttnDownBlock2D""", ) , in_channels=8 , mid_block_type=UpperCamelCase_ , only_cross_attention=UpperCamelCase_ , out_channels=5 , resnet_time_scale_shift="""scale_shift""" , time_embedding_type="""fourier""" , timestep_post_act="""gelu""" , up_block_types=("""KCrossAttnUpBlock2D""", """KCrossAttnUpBlock2D""", """KCrossAttnUpBlock2D""", """KUpBlock2D""") , ) __lowercase = AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ """DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D""", ] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) __lowercase = EulerDiscreteScheduler(prediction_type="""sample""" ) __lowercase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""quick_gelu""" , projection_dim=512 , ) __lowercase = CLIPTextModel(UpperCamelCase_ ) __lowercase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) __lowercase = { "unet": model.eval(), "vae": vae.eval(), "scheduler": scheduler, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def snake_case__ ( self : List[Any] , lowercase : List[Any] , lowercase : int=0 ) -> Optional[int]: """simple docstring""" if str(UpperCamelCase_ ).startswith("""mps""" ): __lowercase = torch.manual_seed(UpperCamelCase_ ) else: __lowercase = torch.Generator(device=UpperCamelCase_ ).manual_seed(UpperCamelCase_ ) __lowercase = { "prompt": "A painting of a squirrel eating a burger", "image": self.dummy_image.cpu(), "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def snake_case__ ( self : str ) -> Optional[int]: """simple docstring""" __lowercase = "cpu" __lowercase = self.get_dummy_components() __lowercase = self.pipeline_class(**UpperCamelCase_ ) pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) __lowercase = self.get_dummy_inputs(UpperCamelCase_ ) __lowercase = pipe(**UpperCamelCase_ ).images __lowercase = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 256, 256, 3) ) __lowercase = np.array( [0.4722_2412, 0.4192_1633, 0.4471_7434, 0.4687_4192, 0.4258_8258, 0.4615_0726, 0.467_7534, 0.4558_3832, 0.4857_9055] ) __lowercase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(UpperCamelCase_ , 1E-3 ) def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" super().test_attention_slicing_forward_pass(expected_max_diff=7E-3 ) def snake_case__ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" super().test_cpu_offload_forward_pass(expected_max_diff=3E-3 ) def snake_case__ ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def snake_case__ ( self : int ) -> List[str]: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=7E-3 ) def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3E-3 ) def snake_case__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" super().test_save_load_local(expected_max_difference=3E-3 ) def snake_case__ ( self : Dict ) -> List[Any]: """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3E-3 ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase = [ "DDIMScheduler", "DDPMScheduler", "PNDMScheduler", "HeunDiscreteScheduler", "EulerAncestralDiscreteScheduler", "KDPM2DiscreteScheduler", "KDPM2AncestralDiscreteScheduler", "DPMSolverSDEScheduler", ] __lowercase = self.get_dummy_components() __lowercase = self.pipeline_class(**UpperCamelCase_ ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=UpperCamelCase_ ) pipe.to(UpperCamelCase_ ) pipe.set_progress_bar_config(disable=UpperCamelCase_ ) __lowercase = self.get_dummy_inputs(UpperCamelCase_ ) __lowercase = 2 __lowercase = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue __lowercase = getattr(UpperCamelCase_ , scheduler_enum.name ) __lowercase = scheduler_cls.from_config(pipe.scheduler.config ) __lowercase = pipe(**UpperCamelCase_ )[0] outputs.append(UpperCamelCase_ ) assert check_same_shape(UpperCamelCase_ ) @require_torch_gpu @slow class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Optional[Any] ) -> str: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self : str ) -> Tuple: """simple docstring""" __lowercase = torch.manual_seed(33 ) __lowercase = StableDiffusionPipeline.from_pretrained("""CompVis/stable-diffusion-v1-4""" , torch_dtype=torch.floataa ) pipe.to("""cuda""" ) __lowercase = StableDiffusionLatentUpscalePipeline.from_pretrained( """stabilityai/sd-x2-latent-upscaler""" , torch_dtype=torch.floataa ) upscaler.to("""cuda""" ) __lowercase = "a photo of an astronaut high resolution, unreal engine, ultra realistic" __lowercase = pipe(UpperCamelCase_ , generator=UpperCamelCase_ , output_type="""latent""" ).images __lowercase = upscaler( prompt=UpperCamelCase_ , image=UpperCamelCase_ , num_inference_steps=20 , guidance_scale=0 , generator=UpperCamelCase_ , output_type="""np""" , ).images[0] __lowercase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy""" ) assert np.abs((expected_image - image).mean() ) < 5E-2 def snake_case__ ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.manual_seed(33 ) __lowercase = StableDiffusionLatentUpscalePipeline.from_pretrained( """stabilityai/sd-x2-latent-upscaler""" , torch_dtype=torch.floataa ) upscaler.to("""cuda""" ) __lowercase = "the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas" __lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png""" ) __lowercase = upscaler( prompt=UpperCamelCase_ , image=UpperCamelCase_ , num_inference_steps=20 , guidance_scale=0 , generator=UpperCamelCase_ , output_type="""np""" , ).images[0] __lowercase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy""" ) assert np.abs((expected_image - image).max() ) < 5E-2
710
import random def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ = False ) -> dict: __lowercase = {i: [] for i in range(lowercase__ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(lowercase__ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(lowercase__ ): for j in range(i + 1 , lowercase__ ): if random.random() < probability: graph[i].append(lowercase__ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(lowercase__ ) return graph def UpperCAmelCase__ ( lowercase__ ) -> dict: return { i: [j for j in range(lowercase__ ) if i != j] for i in range(lowercase__ ) } if __name__ == "__main__": import doctest doctest.testmod()
634
0
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : Union[str, Any] , lowercase : Union[str, Any] , lowercase : Optional[int]=13 , lowercase : Optional[Any]=7 , lowercase : int=True , lowercase : Any=True , lowercase : Dict=True , lowercase : List[Any]=True , lowercase : Optional[Any]=99 , lowercase : List[str]=32 , lowercase : int=5 , lowercase : str=4 , lowercase : int=37 , lowercase : Optional[Any]="gelu" , lowercase : Union[str, Any]=0.1 , lowercase : str=0.1 , lowercase : Any=512 , lowercase : Any=16 , lowercase : Optional[int]=2 , lowercase : int=0.02 , lowercase : Optional[int]=4 , ) -> Dict: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_attention_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_choices def snake_case__ ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_attention_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowercase = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__A , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self : Dict ) -> int: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase = config_and_inputs __lowercase = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Union[str, Any] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def snake_case__ ( self : Any ) -> str: """simple docstring""" __lowercase = FlaxAlbertModelTester(self ) @slow def snake_case__ ( self : Union[str, Any] ) -> str: """simple docstring""" for model_class_name in self.all_model_classes: __lowercase = model_class_name.from_pretrained("""albert-base-v2""" ) __lowercase = model(np.ones((1, 1) ) ) self.assertIsNotNone(__A ) @require_flax class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def snake_case__ ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase = FlaxAlbertModel.from_pretrained("""albert-base-v2""" ) __lowercase = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __lowercase = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) __lowercase = model(__A , attention_mask=__A )[0] __lowercase = (1, 11, 768) self.assertEqual(output.shape , __A ) __lowercase = np.array( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , __A , atol=1E-4 ) )
711
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor UpperCamelCase__ = random.Random() def UpperCAmelCase__ ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ) -> str: if rng is None: __lowercase = global_rng __lowercase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : List[Any] , lowercase : Tuple , lowercase : Union[str, Any]=7 , lowercase : List[Any]=400 , lowercase : Any=2_000 , lowercase : Optional[int]=24 , lowercase : Any=24 , lowercase : List[str]=0.0 , lowercase : Dict=16_000 , lowercase : Union[str, Any]=True , lowercase : Dict=True , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = min_seq_length __lowercase = max_seq_length __lowercase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowercase = feature_size __lowercase = num_mel_bins __lowercase = padding_value __lowercase = sampling_rate __lowercase = return_attention_mask __lowercase = do_normalize def snake_case__ ( self : Optional[int] ) -> Any: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self : List[str] , lowercase : Tuple=False , lowercase : int=False ) -> Optional[Any]: """simple docstring""" def _flatten(lowercase : Optional[Any] ): return list(itertools.chain(*lowercase ) ) if equal_length: __lowercase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowercase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowercase = [np.asarray(lowercase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = SpeechaTextFeatureExtractor if is_speech_available() else None def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase = SpeechaTextFeatureExtractionTester(self ) def snake_case__ ( self : Tuple , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" self.assertTrue(np.all(np.mean(lowercase , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowercase , axis=0 ) - 1 ) < 1E-3 ) ) def snake_case__ ( self : List[Any] ) -> str: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = [np.asarray(lowercase ) for speech_input in speech_inputs] # Test feature size __lowercase = feature_extractor(lowercase , padding=lowercase , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input __lowercase = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features __lowercase = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) # Test batched __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(lowercase , lowercase ): self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __lowercase = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowercase = np.asarray(lowercase ) __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features __lowercase = feature_extractor(lowercase , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(lowercase , lowercase ): self.assertTrue(np.allclose(lowercase , lowercase , atol=1E-3 ) ) def snake_case__ ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(lowercase , lowercase ): __lowercase = feature_extractor( lowercase , padding=lowercase , max_length=lowercase , return_attention_mask=lowercase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(lowercase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def snake_case__ ( self : Any ) -> Dict: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = ["""longest""", """max_length""", """do_not_pad"""] __lowercase = [None, 16, None] for max_length, padding in zip(lowercase , lowercase ): __lowercase = feature_extractor( lowercase , max_length=lowercase , padding=lowercase , return_tensors="""np""" , return_attention_mask=lowercase ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = [np.sum(lowercase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def snake_case__ ( self : str ) -> List[Any]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""max_length""" , max_length=4 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""longest""" , max_length=4 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) __lowercase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowercase = feature_extractor( lowercase , padding="""longest""" , max_length=16 , truncation=lowercase , return_tensors="""np""" , return_attention_mask=lowercase , ) __lowercase = inputs.input_features __lowercase = inputs.attention_mask __lowercase = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" import torch __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = np.random.rand(100 , 32 ).astype(np.floataa ) __lowercase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowercase = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self : Optional[int] , lowercase : Union[str, Any] ) -> int: """simple docstring""" from datasets import load_dataset __lowercase = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __lowercase = ds.sort("""id""" ).select(range(lowercase ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on __lowercase = self._load_datasamples(1 ) __lowercase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowercase = feature_extractor(lowercase , return_tensors="""pt""" ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , lowercase , atol=1E-4 ) )
634
0
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class _lowerCAmelCase : """simple docstring""" def __init__( self : str , lowercase : List[str] , lowercase : Dict=99 , lowercase : List[Any]=13 , lowercase : str=7 , lowercase : Optional[Any]=9 , lowercase : Optional[int]=True , lowercase : str=True , lowercase : Optional[Any]=False , lowercase : Any=32 , lowercase : Optional[int]=5 , lowercase : List[str]=4 , lowercase : Optional[int]=37 , lowercase : Tuple=8 , lowercase : Optional[Any]=0.1 , lowercase : Tuple=0.002 , lowercase : Tuple=1 , lowercase : Any=0 , lowercase : Any=0 , lowercase : Dict=None , lowercase : Any=None , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = encoder_seq_length __lowercase = decoder_seq_length # For common tests __lowercase = self.decoder_seq_length __lowercase = is_training __lowercase = use_attention_mask __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = d_ff __lowercase = relative_attention_num_buckets __lowercase = dropout_rate __lowercase = initializer_factor __lowercase = eos_token_id __lowercase = pad_token_id __lowercase = decoder_start_token_id __lowercase = None __lowercase = decoder_layers def snake_case__ ( self : str ) -> Tuple: """simple docstring""" return TaConfig.from_pretrained("""google/umt5-base""" ) def snake_case__ ( self : Any , lowercase : Any , lowercase : int , lowercase : str , lowercase : int=None , lowercase : List[Any]=None , lowercase : Optional[int]=None , lowercase : Tuple=None , lowercase : Union[str, Any]=None , ) -> Any: """simple docstring""" if attention_mask is None: __lowercase = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: __lowercase = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: __lowercase = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=_a ) if decoder_head_mask is None: __lowercase = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=_a ) if cross_attn_head_mask is None: __lowercase = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=_a ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def snake_case__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) __lowercase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input __lowercase = input_ids.clamp(self.pad_token_id + 1 ) __lowercase = decoder_input_ids.clamp(self.pad_token_id + 1 ) __lowercase = self.get_config() __lowercase = config.num_attention_heads __lowercase = self.prepare_inputs_dict(_a , _a , _a ) return config, input_dict def snake_case__ ( self : Tuple ) -> Any: """simple docstring""" __lowercase = self.prepare_config_and_inputs() return config, inputs_dict def snake_case__ ( self : int ) -> List[str]: """simple docstring""" return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def snake_case__ ( self : Tuple ) -> List[Any]: """simple docstring""" return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def snake_case__ ( self : Dict , lowercase : List[Any] , lowercase : List[str] , lowercase : Tuple , lowercase : Any , lowercase : Tuple , lowercase : Union[str, Any] , ) -> Optional[Any]: """simple docstring""" __lowercase = UMTaModel(config=_a ) model.to(_a ) model.eval() __lowercase = model( input_ids=_a , decoder_input_ids=_a , attention_mask=_a , decoder_attention_mask=_a , ) __lowercase = model(input_ids=_a , decoder_input_ids=_a ) __lowercase = result.last_hidden_state __lowercase = result.past_key_values __lowercase = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(_a ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def snake_case__ ( self : Dict , lowercase : Tuple , lowercase : List[Any] , lowercase : int , lowercase : Any , lowercase : List[Any] , lowercase : Any , ) -> Union[str, Any]: """simple docstring""" __lowercase = UMTaModel(config=_a ).get_decoder().to(_a ).eval() # first forward pass __lowercase = model(_a , use_cache=_a ) __lowercase = model(_a ) __lowercase = model(_a , use_cache=_a ) self.parent.assertTrue(len(_a ) == len(_a ) ) self.parent.assertTrue(len(_a ) == len(_a ) + 1 ) __lowercase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowercase = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and __lowercase = torch.cat([input_ids, next_tokens] , dim=-1 ) __lowercase = model(_a )["""last_hidden_state"""] __lowercase = model(_a , past_key_values=_a )["""last_hidden_state"""] # select random slice __lowercase = ids_tensor((1,) , output_from_past.shape[-1] ).item() __lowercase = output_from_no_past[:, -1, random_slice_idx].detach() __lowercase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_a , _a , atol=1E-3 ) ) def snake_case__ ( self : Optional[Any] , lowercase : int , lowercase : Union[str, Any] , ) -> List[Any]: """simple docstring""" __lowercase = UMTaModel(config=_a ).to(_a ).half().eval() __lowercase = model(**_a )["""last_hidden_state"""] self.parent.assertFalse(torch.isnan(_a ).any().item() ) @require_torch class _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): """simple docstring""" lowercase__ : Union[str, Any] = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) lowercase__ : Optional[int] = (UMTaForConditionalGeneration,) if is_torch_available() else () lowercase__ : List[str] = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) lowercase__ : Union[str, Any] = True lowercase__ : Tuple = False lowercase__ : Optional[Any] = False lowercase__ : List[Any] = True lowercase__ : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests lowercase__ : Union[str, Any] = [0.8, 0.9] def snake_case__ ( self : int ) -> Optional[int]: """simple docstring""" __lowercase = UMTaModelTester(self ) @unittest.skip("""Test has a segmentation fault on torch 1.8.0""" ) def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() __lowercase = UMTaModel(config_and_inputs[0] ).to(_a ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( _a , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=_a , opset_version=9 , input_names=["""input_ids""", """decoder_input_ids"""] , ) @unittest.skipIf(torch_device == """cpu""" , """Cant do half precision""" ) def snake_case__ ( self : Optional[Any] ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*_a ) def snake_case__ ( self : Any ) -> List[str]: """simple docstring""" __lowercase = ["""encoder_attentions""", """decoder_attentions""", """cross_attentions"""] __lowercase = self.model_tester.prepare_config_and_inputs() __lowercase = config_and_inputs[0] __lowercase = UMTaForConditionalGeneration(_a ).eval() model.to(_a ) __lowercase = { """head_mask""": torch.zeros(config.num_layers , config.num_heads , device=_a ), """decoder_head_mask""": torch.zeros(config.num_decoder_layers , config.num_heads , device=_a ), """cross_attn_head_mask""": torch.zeros(config.num_decoder_layers , config.num_heads , device=_a ), } for attn_name, (name, mask) in zip(_a , head_masking.items() ): __lowercase = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": __lowercase = torch.ones( config.num_decoder_layers , config.num_heads , device=_a ) __lowercase = model.generate( config_and_inputs[1]["""input_ids"""] , num_beams=1 , max_length=3 , output_attentions=_a , return_dict_in_generate=_a , **_a , ) # We check the state of decoder_attentions and cross_attentions just from the last step __lowercase = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip("""Does not work on the tiny model as we keep hitting edge cases.""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" pass @require_torch @require_sentencepiece @require_tokenizers class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow @unittest.skip( """Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged""" ) def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = UMTaForConditionalGeneration.from_pretrained("""google/umt5-small""" , return_dict=_a ).to(_a ) __lowercase = AutoTokenizer.from_pretrained("""google/umt5-small""" , use_fast=_a , legacy=_a ) __lowercase = [ """Bonjour monsieur <extra_id_0> bien <extra_id_1>.""", """No se como puedo <extra_id_0>.""", """This is the reason why we <extra_id_0> them.""", """The <extra_id_0> walks in <extra_id_1>, seats""", """A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.""", ] __lowercase = tokenizer(_a , return_tensors="""pt""" , padding=_a ).input_ids # fmt: off __lowercase = torch.tensor( [ [ 38_530, 210_703, 256_299, 1_410, 256_298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25_922, 256_299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1_460, 339, 312, 19_014, 10_620, 758, 256_299, 2_355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256_299, 14_869, 281, 301, 256_298, 275, 119_983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256_299, 14_869, 281, 2_234, 289, 2_275, 333,61_391, 289, 256_298, 543, 256_297, 168_714, 329, 256_296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(_a , _a ) __lowercase = model.generate(input_ids.to(_a ) ) __lowercase = [ """<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>""", """<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>""", """<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>""", """<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>""", """<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>""", ] __lowercase = tokenizer.batch_decode(_a ) self.assertEqual(_a , _a )
712
def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: __lowercase = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def UpperCAmelCase__ ( ) -> List[str]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
634
0
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING UpperCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(_snake_case ) class _lowerCAmelCase ( _snake_case ): """simple docstring""" def __init__( self : Any , **lowercase : Any ) -> Tuple: """simple docstring""" super().__init__(**lowercase ) if self.framework == "tf": raise ValueError(F"The {self.__class__} is only available in PyTorch." ) requires_backends(self , """vision""" ) self.check_model_type(lowercase ) def __call__( self : List[str] , lowercase : Union[str, "Image.Image", List[Dict[str, Any]]] , lowercase : Union[str, List[str]] = None , **lowercase : Optional[Any] , ) -> Any: """simple docstring""" if "text_queries" in kwargs: __lowercase = kwargs.pop("""text_queries""" ) if isinstance(lowercase , (str, Image.Image) ): __lowercase = {"""image""": image, """candidate_labels""": candidate_labels} else: __lowercase = image __lowercase = super().__call__(lowercase , **lowercase ) return results def snake_case__ ( self : Any , **lowercase : Any ) -> List[Any]: """simple docstring""" __lowercase = {} if "threshold" in kwargs: __lowercase = kwargs["""threshold"""] if "top_k" in kwargs: __lowercase = kwargs["""top_k"""] return {}, {}, postprocess_params def snake_case__ ( self : Tuple , lowercase : List[Any] ) -> List[Any]: """simple docstring""" __lowercase = load_image(inputs["""image"""] ) __lowercase = inputs["""candidate_labels"""] if isinstance(lowercase , lowercase ): __lowercase = candidate_labels.split(""",""" ) __lowercase = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(lowercase ): __lowercase = self.tokenizer(lowercase , return_tensors=self.framework ) __lowercase = self.image_processor(lowercase , return_tensors=self.framework ) yield { "is_last": i == len(lowercase ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def snake_case__ ( self : Optional[Any] , lowercase : Union[str, Any] ) -> Tuple: """simple docstring""" __lowercase = model_inputs.pop("""target_size""" ) __lowercase = model_inputs.pop("""candidate_label""" ) __lowercase = model_inputs.pop("""is_last""" ) __lowercase = self.model(**lowercase ) __lowercase = {"""target_size""": target_size, """candidate_label""": candidate_label, """is_last""": is_last, **outputs} return model_outputs def snake_case__ ( self : Any , lowercase : Tuple , lowercase : Any=0.1 , lowercase : Optional[int]=None ) -> List[Any]: """simple docstring""" __lowercase = [] for model_output in model_outputs: __lowercase = model_output["""candidate_label"""] __lowercase = BaseModelOutput(lowercase ) __lowercase = self.image_processor.post_process_object_detection( outputs=lowercase , threshold=lowercase , target_sizes=model_output["""target_size"""] )[0] for index in outputs["scores"].nonzero(): __lowercase = outputs["""scores"""][index].item() __lowercase = self._get_bounding_box(outputs["""boxes"""][index][0] ) __lowercase = {"""score""": score, """label""": label, """box""": box} results.append(lowercase ) __lowercase = sorted(lowercase , key=lambda lowercase : x["score"] , reverse=lowercase ) if top_k: __lowercase = results[:top_k] return results def snake_case__ ( self : Dict , lowercase : "torch.Tensor" ) -> str: """simple docstring""" if self.framework != "pt": raise ValueError("""The ZeroShotObjectDetectionPipeline is only available in PyTorch.""" ) __lowercase , __lowercase , __lowercase , __lowercase = box.int().tolist() __lowercase = { """xmin""": xmin, """ymin""": ymin, """xmax""": xmax, """ymax""": ymax, } return bbox
713
def UpperCAmelCase__ ( lowercase__ = 100 ) -> int: __lowercase = n * (n + 1) * (2 * n + 1) / 6 __lowercase = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F"""{solution() = }""")
634
0
from scipy.stats import spearmanr import datasets UpperCamelCase__ = "\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no correlation.\nPositive correlations imply that as data in dataset x increases, so\ndoes data in dataset y. Negative correlations imply that as x increases,\ny decreases. Correlations of -1 or +1 imply an exact monotonic relationship.\n\nUnlike the Pearson correlation, the Spearman correlation does not\nassume that both datasets are normally distributed.\n\nThe p-value roughly indicates the probability of an uncorrelated system\nproducing datasets that have a Spearman correlation at least as extreme\nas the one computed from these datasets. The p-values are not entirely\nreliable but are probably reasonable for datasets larger than 500 or so.\n" UpperCamelCase__ = "\nArgs:\n predictions (`List[float]`): Predicted labels, as returned by a model.\n references (`List[float]`): Ground truth labels.\n return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns\n only the spearmanr score. Defaults to `False`.\nReturns:\n spearmanr (`float`): Spearman correlation coefficient.\n p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input.\nExamples:\n Example 1:\n >>> spearmanr_metric = datasets.load_metric(\"spearmanr\")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4])\n >>> print(results)\n {'spearmanr': -0.7}\n\n Example 2:\n >>> spearmanr_metric = datasets.load_metric(\"spearmanr\")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5],\n ... predictions=[10, 9, 2.5, 6, 4],\n ... return_pvalue=True)\n >>> print(results['spearmanr'])\n -0.7\n >>> print(round(results['spearmanr_pvalue'], 2))\n 0.19\n" UpperCamelCase__ = R"\\n@book{kokoska2000crc,\n title={CRC standard probability and statistics tables and formulae},\n author={Kokoska, Stephen and Zwillinger, Daniel},\n year={2000},\n publisher={Crc Press}\n}\n@article{2020SciPy-NMeth,\n author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\n title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\n journal = {Nature Methods},\n year = {2020},\n volume = {17},\n pages = {261--272},\n adsurl = {https://rdcu.be/b08Wh},\n doi = {10.1038/s41592-019-0686-2},\n}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def snake_case__ ( self : Tuple ) -> Dict: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"""] , ) def snake_case__ ( self : str , lowercase : Tuple , lowercase : Tuple , lowercase : Optional[int]=False ) -> List[str]: """simple docstring""" __lowercase = spearmanr(_UpperCAmelCase , _UpperCAmelCase ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
714
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets UpperCamelCase__ = datasets.logging.get_logger(__name__) UpperCamelCase__ = "\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = \"Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric\",\n author = \"Moosavi, Nafise Sadat and\n Strube, Michael\",\n booktitle = \"Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)\",\n month = aug,\n year = \"2016\",\n address = \"Berlin, Germany\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/P16-1060\",\n doi = \"10.18653/v1/P16-1060\",\n pages = \"632--642\",\n}\n\n" UpperCamelCase__ = "\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a \"-\"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section \"*_conll File Format\"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n" UpperCamelCase__ = "\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting 'keep_singletons=False', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the 'NP_only' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting 'min_span', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n 'mentions': mentions\n 'muc': MUC metric [Vilain et al, 1995]\n 'bcub': B-cubed [Bagga and Baldwin, 1998]\n 'ceafe': CEAFe [Luo et al., 2005]\n 'lea': LEA [Moosavi and Strube, 2016]\n 'conll_score': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric('coval')\n >>> words = ['bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -',\n ... 'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)',\n ... 'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)',\n ... 'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -',\n ... 'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -',\n ... 'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {'mentions/recall': 1.0,[...] 'conll_score': 100.0}\n" def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__=False , lowercase__=False , lowercase__=True , lowercase__=False , lowercase__="dummy_doc" ) -> str: __lowercase = {doc: key_lines} __lowercase = {doc: sys_lines} __lowercase = {} __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase , __lowercase = reader.get_doc_mentions(lowercase__ , key_doc_lines[doc] , lowercase__ ) key_singletons_num += singletons_num if NP_only or min_span: __lowercase = reader.set_annotated_parse_trees(lowercase__ , key_doc_lines[doc] , lowercase__ , lowercase__ ) __lowercase , __lowercase = reader.get_doc_mentions(lowercase__ , sys_doc_lines[doc] , lowercase__ ) sys_singletons_num += singletons_num if NP_only or min_span: __lowercase = reader.set_annotated_parse_trees(lowercase__ , key_doc_lines[doc] , lowercase__ , lowercase__ ) if remove_nested: __lowercase , __lowercase = reader.remove_nested_coref_mentions(lowercase__ , lowercase__ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters __lowercase , __lowercase = reader.remove_nested_coref_mentions(lowercase__ , lowercase__ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters __lowercase = reader.get_mention_assignments(lowercase__ , lowercase__ ) __lowercase = reader.get_mention_assignments(lowercase__ , lowercase__ ) __lowercase = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F"annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}" ) logger.info( """Number of resulting singleton clusters in the key """ F"annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}" ) if not keep_singletons: logger.info( F"{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system " """files, respectively""" ) return doc_coref_infos def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) -> Optional[Any]: __lowercase = get_coref_infos(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) __lowercase = {} __lowercase = 0 __lowercase = 0 for name, metric in metrics: __lowercase , __lowercase , __lowercase = evaluator.evaluate_documents(lowercase__ , lowercase__ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F"{name}/recall": recall, F"{name}/precision": precision, F"{name}/f1": fa} ) logger.info( name.ljust(10 ) , F"Recall: {recall * 100:.2f}" , F" Precision: {precision * 100:.2f}" , F" F1: {fa * 100:.2f}" , ) if conll_subparts_num == 3: __lowercase = (conll / 3) * 100 logger.info(F"CoNLL score: {conll:.2f}" ) output_scores.update({"""conll_score""": conll} ) return output_scores def UpperCAmelCase__ ( lowercase__ ) -> List[Any]: __lowercase = False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: __lowercase = line.split()[5] if not parse_col == "-": __lowercase = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def snake_case__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def snake_case__ ( self : Tuple , lowercase : Dict , lowercase : Optional[int] , lowercase : Dict=True , lowercase : List[str]=False , lowercase : int=False , lowercase : Dict=False ) -> str: """simple docstring""" __lowercase = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: __lowercase = util.check_gold_parse_annotation(lowercase ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" __lowercase = evaluate( key_lines=lowercase , sys_lines=lowercase , metrics=lowercase , NP_only=lowercase , remove_nested=lowercase , keep_singletons=lowercase , min_span=lowercase , ) return score
634
0
def UpperCAmelCase__ ( lowercase__ ) -> Any: __lowercase = 1 for i in range(1 , num + 1 ): fact *= i return fact def UpperCAmelCase__ ( lowercase__ ) -> Optional[int]: __lowercase = 0 while number > 0: __lowercase = number % 10 sum_of_digits += last_digit __lowercase = number // 10 # Removing the last_digit from the given number return sum_of_digits def UpperCAmelCase__ ( lowercase__ = 100 ) -> Optional[int]: __lowercase = factorial(__A ) __lowercase = split_and_add(__A ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
715
UpperCamelCase__ = { "joule": 1.0, "kilojoule": 10_00, "megajoule": 1_00_00_00, "gigajoule": 10_00_00_00_00, "wattsecond": 1.0, "watthour": 36_00, "kilowatthour": 3_60_00_00, "newtonmeter": 1.0, "calorie_nutr": 41_86.8, "kilocalorie_nutr": 4_18_68_00.00, "electronvolt": 1.6_0217_6634e-19, "britishthermalunit_it": 10_55.0_55_85, "footpound": 1.355_818, } def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> float: if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: __lowercase = ( F"Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n" F"Valid values are: {', '.join(lowercase__ )}" ) raise ValueError(lowercase__ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
634
0
from collections.abc import Callable def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ ) -> int: __lowercase = a __lowercase = b if function(_lowerCamelCase ) == 0: # one of the a or b is a root for the function return a elif function(_lowerCamelCase ) == 0: return b elif ( function(_lowerCamelCase ) * function(_lowerCamelCase ) > 0 ): # if none of these are root and they are both positive or negative, # then this algorithm can't find the root raise ValueError("""could not find root in given interval.""" ) else: __lowercase = start + (end - start) / 2.0 while abs(start - mid ) > 10**-7: # until precisely equals to 10^-7 if function(_lowerCamelCase ) == 0: return mid elif function(_lowerCamelCase ) * function(_lowerCamelCase ) < 0: __lowercase = mid else: __lowercase = mid __lowercase = start + (end - start) / 2.0 return mid def UpperCAmelCase__ ( lowercase__ ) -> int: return x**3 - 2 * x - 5 if __name__ == "__main__": print(bisection(f, 1, 10_00)) import doctest doctest.testmod()
716
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_UpperCAmelCase ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : str = field(default="""summarization""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowercase__ : ClassVar[Features] = Features({"""text""": Value("""string""" )} ) lowercase__ : ClassVar[Features] = Features({"""summary""": Value("""string""" )} ) lowercase__ : str = "text" lowercase__ : str = "summary" @property def snake_case__ ( self : List[Any] ) -> Dict[str, str]: """simple docstring""" return {self.text_column: "text", self.summary_column: "summary"}
634
0
import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = "▁" UpperCamelCase__ = { "vocab_file": "vocab.json", "spm_file": "sentencepiece.bpe.model", "tokenizer_config_file": "tokenizer_config.json", } UpperCamelCase__ = { "vocab_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json", }, "spm_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model", }, "tokenizer_config_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json", }, } UpperCamelCase__ = { "facebook/m2m100_418M": 10_24, } # fmt: off UpperCamelCase__ = { "m2m100": ["af", "am", "ar", "ast", "az", "ba", "be", "bg", "bn", "br", "bs", "ca", "ceb", "cs", "cy", "da", "de", "el", "en", "es", "et", "fa", "ff", "fi", "fr", "fy", "ga", "gd", "gl", "gu", "ha", "he", "hi", "hr", "ht", "hu", "hy", "id", "ig", "ilo", "is", "it", "ja", "jv", "ka", "kk", "km", "kn", "ko", "lb", "lg", "ln", "lo", "lt", "lv", "mg", "mk", "ml", "mn", "mr", "ms", "my", "ne", "nl", "no", "ns", "oc", "or", "pa", "pl", "ps", "pt", "ro", "ru", "sd", "si", "sk", "sl", "so", "sq", "sr", "ss", "su", "sv", "sw", "ta", "th", "tl", "tn", "tr", "uk", "ur", "uz", "vi", "wo", "xh", "yi", "yo", "zh", "zu"], "wmt21": ["en", "ha", "is", "ja", "cs", "ru", "zh", "de"] } class _lowerCAmelCase ( _A ): """simple docstring""" lowercase__ : List[str] = VOCAB_FILES_NAMES lowercase__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : int = PRETRAINED_VOCAB_FILES_MAP lowercase__ : str = ["""input_ids""", """attention_mask"""] lowercase__ : Optional[Any] = [] lowercase__ : Union[str, Any] = [] def __init__( self : Dict , lowercase : Optional[Any] , lowercase : Optional[int] , lowercase : Tuple=None , lowercase : int=None , lowercase : Union[str, Any]="<s>" , lowercase : int="</s>" , lowercase : str="</s>" , lowercase : int="<pad>" , lowercase : List[str]="<unk>" , lowercase : str="m2m100" , lowercase : Optional[Dict[str, Any]] = None , lowercase : int=8 , **lowercase : List[Any] , ) -> None: """simple docstring""" __lowercase = {} if sp_model_kwargs is None else sp_model_kwargs __lowercase = language_codes __lowercase = FAIRSEQ_LANGUAGE_CODES[language_codes] __lowercase = {lang_code: F"__{lang_code}__" for lang_code in fairseq_language_code} __lowercase = kwargs.get("""additional_special_tokens""" , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(__lowerCamelCase ) for lang_code in fairseq_language_code if self.get_lang_token(__lowerCamelCase ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=__lowerCamelCase , tgt_lang=__lowerCamelCase , bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , sep_token=__lowerCamelCase , unk_token=__lowerCamelCase , pad_token=__lowerCamelCase , language_codes=__lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=__lowerCamelCase , **__lowerCamelCase , ) __lowercase = vocab_file __lowercase = load_json(__lowerCamelCase ) __lowercase = {v: k for k, v in self.encoder.items()} __lowercase = spm_file __lowercase = load_spm(__lowerCamelCase , self.sp_model_kwargs ) __lowercase = len(self.encoder ) __lowercase = { self.get_lang_token(__lowerCamelCase ): self.encoder_size + i for i, lang_code in enumerate(__lowerCamelCase ) } __lowercase = {lang_code: self.encoder_size + i for i, lang_code in enumerate(__lowerCamelCase )} __lowercase = {v: k for k, v in self.lang_token_to_id.items()} __lowercase = src_lang if src_lang is not None else "en" __lowercase = tgt_lang __lowercase = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) __lowercase = num_madeup_words @property def snake_case__ ( self : str ) -> int: """simple docstring""" return len(self.encoder ) + len(self.lang_token_to_id ) @property def snake_case__ ( self : Any ) -> str: """simple docstring""" return self._src_lang @src_lang.setter def snake_case__ ( self : List[str] , lowercase : str ) -> None: """simple docstring""" __lowercase = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def snake_case__ ( self : Union[str, Any] , lowercase : str ) -> List[str]: """simple docstring""" return self.sp_model.encode(__lowerCamelCase , out_type=__lowerCamelCase ) def snake_case__ ( self : List[str] , lowercase : str ) -> Tuple: """simple docstring""" if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(__lowerCamelCase , self.encoder[self.unk_token] ) def snake_case__ ( self : Union[str, Any] , lowercase : int ) -> str: """simple docstring""" if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(__lowerCamelCase , self.unk_token ) def snake_case__ ( self : Optional[int] , lowercase : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase = [] __lowercase = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__lowerCamelCase ) + token __lowercase = [] else: current_sub_tokens.append(__lowerCamelCase ) out_string += self.sp_model.decode(__lowerCamelCase ) return out_string.strip() def snake_case__ ( self : int , lowercase : List[int] , lowercase : Optional[List[int]] = None , lowercase : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCamelCase , token_ids_a=__lowerCamelCase , already_has_special_tokens=__lowerCamelCase ) __lowercase = [1] * len(self.prefix_tokens ) __lowercase = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(__lowerCamelCase )) + suffix_ones return prefix_ones + ([0] * len(__lowerCamelCase )) + ([0] * len(__lowerCamelCase )) + suffix_ones def snake_case__ ( self : Union[str, Any] , lowercase : List[int] , lowercase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def snake_case__ ( self : List[str] ) -> Dict: """simple docstring""" __lowercase = {self.convert_ids_to_tokens(__lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[int] ) -> Dict: """simple docstring""" __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str , lowercase : Dict ) -> None: """simple docstring""" __lowercase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __lowercase = {} __lowercase = load_spm(self.spm_file , self.sp_model_kwargs ) def snake_case__ ( self : List[str] , lowercase : str , lowercase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" __lowercase = Path(__lowerCamelCase ) if not save_dir.is_dir(): raise OSError(F"{save_directory} should be a directory" ) __lowercase = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["vocab_file"] ) __lowercase = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["spm_file"] ) save_json(self.encoder , __lowerCamelCase ) if os.path.abspath(self.spm_file ) != os.path.abspath(__lowerCamelCase ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , __lowerCamelCase ) elif not os.path.isfile(self.spm_file ): with open(__lowerCamelCase , """wb""" ) as fi: __lowercase = self.sp_model.serialized_model_proto() fi.write(__lowerCamelCase ) return (str(__lowerCamelCase ), str(__lowerCamelCase )) def snake_case__ ( self : Union[str, Any] , lowercase : List[str] , lowercase : str = "en" , lowercase : Optional[List[str]] = None , lowercase : str = "ro" , **lowercase : Tuple , ) -> BatchEncoding: """simple docstring""" __lowercase = src_lang __lowercase = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(__lowerCamelCase , __lowerCamelCase , **__lowerCamelCase ) def snake_case__ ( self : Union[str, Any] , lowercase : Dict , lowercase : Optional[str] , lowercase : Optional[str] , **lowercase : Dict ) -> Optional[int]: """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) __lowercase = src_lang __lowercase = self(__lowerCamelCase , add_special_tokens=__lowerCamelCase , **__lowerCamelCase ) __lowercase = self.get_lang_id(__lowerCamelCase ) __lowercase = tgt_lang_id return inputs def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" self.set_src_lang_special_tokens(self.src_lang ) def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" self.set_tgt_lang_special_tokens(self.tgt_lang ) def snake_case__ ( self : Tuple , lowercase : str ) -> None: """simple docstring""" __lowercase = self.get_lang_token(__lowerCamelCase ) __lowercase = self.lang_token_to_id[lang_token] __lowercase = [self.cur_lang_id] __lowercase = [self.eos_token_id] def snake_case__ ( self : Tuple , lowercase : str ) -> None: """simple docstring""" __lowercase = self.get_lang_token(__lowerCamelCase ) __lowercase = self.lang_token_to_id[lang_token] __lowercase = [self.cur_lang_id] __lowercase = [self.eos_token_id] def snake_case__ ( self : List[Any] , lowercase : str ) -> str: """simple docstring""" return self.lang_code_to_token[lang] def snake_case__ ( self : List[str] , lowercase : str ) -> int: """simple docstring""" __lowercase = self.get_lang_token(__lowerCamelCase ) return self.lang_token_to_id[lang_token] def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Any: __lowercase = sentencepiece.SentencePieceProcessor(**_lowerCamelCase ) spm.Load(str(_lowerCamelCase ) ) return spm def UpperCAmelCase__ ( lowercase__ ) -> Optional[Any]: with open(_lowerCamelCase , """r""" ) as f: return json.load(_lowerCamelCase ) def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Union[str, Any]: with open(_lowerCamelCase , """w""" ) as f: json.dump(_lowerCamelCase , _lowerCamelCase , indent=2 )
717
def UpperCAmelCase__ ( lowercase__ ) -> Optional[int]: __lowercase = len(lowercase__ ) __lowercase = sum(lowercase__ ) __lowercase = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): __lowercase = True for i in range(1 , s + 1 ): __lowercase = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): __lowercase = dp[i][j - 1] if arr[i - 1] <= j: __lowercase = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: __lowercase = s - 2 * j break return diff
634
0
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinvaConfig, SwinvaForImageClassification def UpperCAmelCase__ ( lowercase__ ) -> Tuple: __lowercase = SwinvaConfig() __lowercase = swinva_name.split("""_""" ) __lowercase = name_split[1] if "to" in name_split[3]: __lowercase = int(name_split[3][-3:] ) else: __lowercase = int(name_split[3] ) if "to" in name_split[2]: __lowercase = int(name_split[2][-2:] ) else: __lowercase = int(name_split[2][6:] ) if model_size == "tiny": __lowercase = 96 __lowercase = (2, 2, 6, 2) __lowercase = (3, 6, 12, 24) elif model_size == "small": __lowercase = 96 __lowercase = (2, 2, 18, 2) __lowercase = (3, 6, 12, 24) elif model_size == "base": __lowercase = 128 __lowercase = (2, 2, 18, 2) __lowercase = (4, 8, 16, 32) else: __lowercase = 192 __lowercase = (2, 2, 18, 2) __lowercase = (6, 12, 24, 48) if "to" in swinva_name: __lowercase = (12, 12, 12, 6) if ("22k" in swinva_name) and ("to" not in swinva_name): __lowercase = 21_841 __lowercase = '''huggingface/label-files''' __lowercase = '''imagenet-22k-id2label.json''' __lowercase = json.load(open(hf_hub_download(a_ , a_ , repo_type="""dataset""" ) , """r""" ) ) __lowercase = {int(a_ ): v for k, v in idalabel.items()} __lowercase = idalabel __lowercase = {v: k for k, v in idalabel.items()} else: __lowercase = 1_000 __lowercase = '''huggingface/label-files''' __lowercase = '''imagenet-1k-id2label.json''' __lowercase = json.load(open(hf_hub_download(a_ , a_ , repo_type="""dataset""" ) , """r""" ) ) __lowercase = {int(a_ ): v for k, v in idalabel.items()} __lowercase = idalabel __lowercase = {v: k for k, v in idalabel.items()} __lowercase = img_size __lowercase = num_classes __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size return config def UpperCAmelCase__ ( lowercase__ ) -> Tuple: if "patch_embed.proj" in name: __lowercase = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: __lowercase = name.replace("""patch_embed.norm""" , """embeddings.norm""" ) if "layers" in name: __lowercase = '''encoder.''' + name if "attn.proj" in name: __lowercase = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: __lowercase = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: __lowercase = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: __lowercase = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: __lowercase = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: __lowercase = name.replace("""mlp.fc2""" , """output.dense""" ) if "q_bias" in name: __lowercase = name.replace("""q_bias""" , """query.bias""" ) if "k_bias" in name: __lowercase = name.replace("""k_bias""" , """key.bias""" ) if "v_bias" in name: __lowercase = name.replace("""v_bias""" , """value.bias""" ) if "cpb_mlp" in name: __lowercase = name.replace("""cpb_mlp""" , """continuous_position_bias_mlp""" ) if name == "norm.weight": __lowercase = '''layernorm.weight''' if name == "norm.bias": __lowercase = '''layernorm.bias''' if "head" in name: __lowercase = name.replace("""head""" , """classifier""" ) else: __lowercase = '''swinv2.''' + name return name def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> str: for key in orig_state_dict.copy().keys(): __lowercase = orig_state_dict.pop(a_ ) if "mask" in key: continue elif "qkv" in key: __lowercase = key.split(""".""" ) __lowercase = int(key_split[1] ) __lowercase = int(key_split[3] ) __lowercase = model.swinva.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: __lowercase = val[:dim, :] __lowercase = val[dim : dim * 2, :] __lowercase = val[-dim:, :] else: __lowercase = val[:dim] __lowercase = val[ dim : dim * 2 ] __lowercase = val[-dim:] else: __lowercase = val return orig_state_dict def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> int: __lowercase = timm.create_model(a_ , pretrained=a_ ) timm_model.eval() __lowercase = get_swinva_config(a_ ) __lowercase = SwinvaForImageClassification(a_ ) model.eval() __lowercase = convert_state_dict(timm_model.state_dict() , a_ ) model.load_state_dict(a_ ) __lowercase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowercase = AutoImageProcessor.from_pretrained("""microsoft/{}""".format(swinva_name.replace("""_""" , """-""" ) ) ) __lowercase = Image.open(requests.get(a_ , stream=a_ ).raw ) __lowercase = image_processor(images=a_ , return_tensors="""pt""" ) __lowercase = timm_model(inputs["""pixel_values"""] ) __lowercase = model(**a_ ).logits assert torch.allclose(a_ , a_ , atol=1E-3 ) print(F"Saving model {swinva_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(a_ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(a_ ) model.push_to_hub( repo_path_or_name=Path(a_ , a_ ) , organization="""nandwalritik""" , commit_message="""Add model""" , ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--swinv2_name", default="swinv2_tiny_patch4_window8_256", type=str, help="Name of the Swinv2 timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) UpperCamelCase__ = parser.parse_args() convert_swinva_checkpoint(args.swinva_name, args.pytorch_dump_folder_path)
718
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class _lowerCAmelCase : """simple docstring""" def __init__( self : List[Any] , lowercase : str , lowercase : Union[str, Any]=13 , lowercase : Tuple=32 , lowercase : Optional[Any]=2 , lowercase : Tuple=3 , lowercase : Tuple=16 , lowercase : Tuple=[1, 2, 1] , lowercase : Optional[Any]=[2, 2, 4] , lowercase : Dict=2 , lowercase : Optional[int]=2.0 , lowercase : List[Any]=True , lowercase : str=0.0 , lowercase : Any=0.0 , lowercase : Optional[int]=0.1 , lowercase : int="gelu" , lowercase : Tuple=False , lowercase : Optional[Any]=True , lowercase : int=0.02 , lowercase : Union[str, Any]=1E-5 , lowercase : Dict=True , lowercase : Any=None , lowercase : str=True , lowercase : str=10 , lowercase : Dict=8 , lowercase : int=["stage1", "stage2", "stage3"] , lowercase : Optional[int]=[1, 2, 3] , ) -> Any: """simple docstring""" __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = patch_norm __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = is_training __lowercase = scope __lowercase = use_labels __lowercase = type_sequence_label_size __lowercase = encoder_stride __lowercase = out_features __lowercase = out_indices def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def snake_case__ ( self : List[str] ) -> int: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def snake_case__ ( self : Any , lowercase : List[Any] , lowercase : Optional[int] , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = MaskFormerSwinModel(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) __lowercase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowercase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def snake_case__ ( self : Any , lowercase : Tuple , lowercase : Any , lowercase : Optional[int] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinBackbone(config=lowercase ) model.to(lowercase ) model.eval() __lowercase = model(lowercase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(lowercase ): __lowercase = ["""stem"""] __lowercase = MaskFormerSwinBackbone(config=lowercase ) def snake_case__ ( self : int ) -> Any: """simple docstring""" __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Optional[int] = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) lowercase__ : List[str] = {"""feature-extraction""": MaskFormerSwinModel} if is_torch_available() else {} lowercase__ : List[str] = False lowercase__ : int = False lowercase__ : int = False lowercase__ : Tuple = False lowercase__ : Optional[Any] = False def snake_case__ ( self : Dict ) -> Optional[Any]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) __lowercase = ConfigTester(self , config_class=lowercase , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( """`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn't work well with""" """ `nn.DataParallel`""" ) ) def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" pass def snake_case__ ( self : Tuple ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" return def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def snake_case__ ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowercase ) @unittest.skip("""Swin does not use inputs_embeds""" ) def snake_case__ ( self : int ) -> Any: """simple docstring""" pass @unittest.skip("""Swin does not support feedforward chunking""" ) def snake_case__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" pass def snake_case__ ( self : Dict ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase ) @unittest.skip(reason="""MaskFormerSwin is only used as backbone and doesn't support output_attentions""" ) def snake_case__ ( self : Any ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip(reason="""MaskFormerSwin is only used as an internal backbone""" ) def snake_case__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" pass def snake_case__ ( self : Tuple , lowercase : Tuple , lowercase : Optional[int] , lowercase : Optional[int] , lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase , lowercase ) ) __lowercase = outputs.hidden_states __lowercase = getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowercase ) , lowercase ) # Swin has a different seq_length __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def snake_case__ ( self : int ) -> Dict: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , lowercase ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowercase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowercase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowercase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True self.check_hidden_states_output(lowercase , lowercase , lowercase , (padded_height, padded_width) ) @unittest.skip(reason="""MaskFormerSwin doesn't have pretrained checkpoints""" ) def snake_case__ ( self : Any ) -> Any: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def snake_case__ ( self : List[str] ) -> str: """simple docstring""" pass @unittest.skip(reason="""This will be fixed once MaskFormerSwin is replaced by native Swin""" ) def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" pass def snake_case__ ( self : Tuple ) -> List[Any]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(lowercase : Optional[int] ): __lowercase = 0 return t def check_equivalence(lowercase : Optional[int] , lowercase : str , lowercase : str , lowercase : Tuple={} ): with torch.no_grad(): __lowercase = model(**lowercase , return_dict=lowercase , **lowercase ) __lowercase = model(**lowercase , return_dict=lowercase , **lowercase ).to_tuple() def recursive_check(lowercase : int , lowercase : Optional[Any] ): if isinstance(lowercase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(lowercase , lowercase ): recursive_check(lowercase , lowercase ) elif isinstance(lowercase , lowercase ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(lowercase , lowercase ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(lowercase ) , set_nan_tensor_to_zero(lowercase ) , atol=1E-5 ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:" F" {torch.isnan(lowercase ).any()} and `inf`: {torch.isinf(lowercase )}. Dict has" F" `nan`: {torch.isnan(lowercase ).any()} and `inf`: {torch.isinf(lowercase )}." ) , ) recursive_check(lowercase , lowercase ) for model_class in self.all_model_classes: __lowercase = model_class(lowercase ) model.to(lowercase ) model.eval() __lowercase = self._prepare_for_class(lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) check_equivalence(lowercase , lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) check_equivalence(lowercase , lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase ) check_equivalence(lowercase , lowercase , lowercase , {"""output_hidden_states""": True} ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) __lowercase = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) check_equivalence(lowercase , lowercase , lowercase , {"""output_hidden_states""": True} ) @require_torch class _lowerCAmelCase ( unittest.TestCase , _UpperCAmelCase ): """simple docstring""" lowercase__ : List[str] = (MaskFormerSwinBackbone,) if is_torch_available() else () lowercase__ : Any = MaskFormerSwinConfig def snake_case__ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = MaskFormerSwinModelTester(self ) def snake_case__ ( self : Any ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = inputs_dict["""pixel_values"""].shape[0] for backbone_class in self.all_model_classes: __lowercase = backbone_class(lowercase ) backbone.to(lowercase ) backbone.eval() __lowercase = backbone(**lowercase ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , lowercase ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True __lowercase = backbone(**lowercase , output_hidden_states=lowercase ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) __lowercase , __lowercase , __lowercase = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: __lowercase = backbone(**lowercase , output_attentions=lowercase ) self.assertIsNotNone(outputs.attentions )
634
0
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel UpperCamelCase__ = HfApi() UpperCamelCase__ = {} # fmt: off UpperCamelCase__ = torch.tensor([ -0.7_515, -1.6_883, 0.2_420, 0.0_300, 0.6_347, 1.3_433, -1.1_743, -3.7_467, 1.2_342, -2.2_485, 0.4_636, 0.8_076, -0.7_991, 0.3_969, 0.8_498, 0.9_189, -1.8_887, -3.3_522, 0.7_639, 0.2_040, 0.6_271, -2.7_148, -1.6_316, 3.0_839, 0.3_186, 0.2_721, -0.9_759, -1.2_461, 2.6_257, 1.3_557 ]) UpperCamelCase__ = torch.tensor([ -2.3_639, -2.5_344, 0.0_054, -0.6_674, 1.5_990, 1.0_158, 0.3_124, -2.1_436, 1.8_795, -2.5_429, -0.1_566, -0.3_973, 1.2_490, 2.6_447, 1.2_283, -0.5_208, -2.8_154, -3.5_119, 2.3_838, 1.2_033, 1.7_201, -2.1_256, -1.4_576, 2.7_948, 2.4_204, -0.9_752, -1.2_546, 0.8_027, 3.2_758, 3.1_365 ]) UpperCamelCase__ = torch.tensor([ -0.6_531, -0.6_891, -0.3_172, -0.5_375, -0.9_140, -0.5_367, -0.1_175, -0.7_869, -0.3_808, -0.4_513, -0.2_098, -0.0_083, 0.3_183, 0.5_140, 0.2_247, -0.1_304, -0.1_302, -0.2_802, -0.2_084, -0.2_025, -0.4_967, -0.4_873, -0.0_861, 0.6_925, 0.0_250, 0.1_290, -0.1_543, 0.6_316, 1.0_460, 1.4_943 ]) UpperCamelCase__ = torch.tensor([ 0.0_911, 0.1_107, 0.0_182, 0.0_435, -0.0_805, -0.0_608, 0.0_381, 0.2_172, -0.0_280, 0.1_327, -0.0_299, -0.0_255, -0.0_050, -0.1_170, -0.1_046, 0.0_309, 0.1_367, 0.1_728, -0.0_533, -0.0_748, -0.0_534, 0.1_624, 0.0_384, -0.1_805, -0.0_707, 0.0_642, 0.0_220, -0.0_134, -0.1_333, -0.1_505 ]) UpperCamelCase__ = torch.tensor([ 0.1_321, 0.1_337, 0.0_440, 0.0_622, -0.0_591, -0.0_370, 0.0_503, 0.2_133, -0.0_177, 0.1_415, -0.0_116, -0.0_112, 0.0_044, -0.0_980, -0.0_789, 0.0_395, 0.1_502, 0.1_785, -0.0_488, -0.0_514, -0.0_404, 0.1_539, 0.0_454, -0.1_559, -0.0_665, 0.0_659, 0.0_383, -0.0_005, -0.1_266, -0.1_386 ]) UpperCamelCase__ = torch.tensor([ 0.1_154, 0.1_218, 0.0_307, 0.0_526, -0.0_711, -0.0_541, 0.0_366, 0.2_078, -0.0_267, 0.1_317, -0.0_226, -0.0_193, -0.0_014, -0.1_055, -0.0_902, 0.0_330, 0.1_391, 0.1_709, -0.0_562, -0.0_693, -0.0_560, 0.1_482, 0.0_381, -0.1_683, -0.0_681, 0.0_661, 0.0_331, -0.0_046, -0.1_268, -0.1_431 ]) UpperCamelCase__ = torch.tensor([ 0.1_192, 0.1_240, 0.0_414, 0.0_606, -0.0_557, -0.0_412, 0.0_430, 0.2_042, -0.0_200, 0.1_385, -0.0_115, -0.0_132, 0.0_017, -0.0_965, -0.0_802, 0.0_398, 0.1_433, 0.1_747, -0.0_458, -0.0_533, -0.0_407, 0.1_545, 0.0_419, -0.1_574, -0.0_645, 0.0_626, 0.0_341, -0.0_010, -0.1_199, -0.1_390 ]) UpperCamelCase__ = torch.tensor([ 0.1_075, 0.1_074, 0.0_205, 0.0_431, -0.0_774, -0.0_607, 0.0_298, 0.2_042, -0.0_320, 0.1_267, -0.0_281, -0.0_250, -0.0_064, -0.1_091, -0.0_946, 0.0_290, 0.1_328, 0.1_650, -0.0_580, -0.0_738, -0.0_586, 0.1_440, 0.0_337, -0.1_746, -0.0_712, 0.0_605, 0.0_250, -0.0_099, -0.1_316, -0.1_473 ]) UpperCamelCase__ = torch.tensor([ -1.4_572, -2.0_481, -0.0_414, -0.6_005, 1.4_136, 0.5_848, 0.4_028, -2.7_330, 1.2_212, -2.1_228, 0.2_155, 0.4_039, 0.7_662, 2.0_535, 0.7_477, -0.3_243, -2.1_758, -2.7_648, 1.6_947, 0.7_026, 1.2_338, -1.6_078, -0.8_682, 2.2_810, 1.8_574, -0.5_718, -0.5_586, -0.0_186, 2.3_415, 2.1_251]) UpperCamelCase__ = torch.tensor([ -1.3_690, -1.9_720, -0.4_090, -0.6_966, 1.4_660, 0.9_938, -0.1_385, -2.7_324, 0.7_736, -1.8_917, 0.2_923, 0.4_293, 0.1_693, 1.4_112, 1.1_887, -0.3_181, -2.2_160, -2.6_381, 1.3_170, 0.8_163, 0.9_240, -1.6_544, -0.6_099, 2.5_259, 1.6_430, -0.9_090, -0.9_392, -0.0_126, 2.4_268, 2.3_266 ]) UpperCamelCase__ = torch.tensor([ -1.3_525, -1.9_628, -0.3_956, -0.6_860, 1.4_664, 1.0_014, -0.1_259, -2.7_212, 0.7_772, -1.8_811, 0.2_996, 0.4_388, 0.1_704, 1.4_029, 1.1_701, -0.3_027, -2.2_053, -2.6_287, 1.3_350, 0.8_131, 0.9_274, -1.6_292, -0.6_098, 2.5_131, 1.6_505, -0.8_958, -0.9_298, -0.0_151, 2.4_257, 2.3_355 ]) UpperCamelCase__ = torch.tensor([ -2.0_585, -2.7_897, -0.2_850, -0.8_940, 1.9_052, 0.5_702, 0.6_345, -3.8_959, 1.5_932, -3.2_319, 0.1_974, 0.0_287, 1.7_566, 2.6_543, 0.8_387, -0.5_351, -3.2_736, -4.3_375, 2.9_029, 1.6_390, 1.4_640, -2.1_701, -1.9_013, 2.9_341, 3.4_981, -0.6_255, -1.1_644, -0.1_591, 3.7_097, 3.2_066 ]) UpperCamelCase__ = torch.tensor([ -2.3_139, -2.5_594, -0.0_197, -0.6_785, 1.7_001, 1.1_606, 0.3_075, -2.1_740, 1.8_071, -2.5_630, -0.0_926, -0.3_811, 1.2_116, 2.6_246, 1.2_731, -0.5_398, -2.8_153, -3.6_140, 2.3_893, 1.3_262, 1.6_258, -2.1_856, -1.3_267, 2.8_395, 2.3_779, -1.0_623, -1.2_468, 0.8_959, 3.3_367, 3.2_243 ]) UpperCamelCase__ = torch.tensor([ -2.0_628, -2.7_667, -0.2_089, -0.8_263, 2.0_539, 0.5_992, 0.6_495, -3.8_336, 1.6_025, -3.2_817, 0.1_721, -0.0_633, 1.7_516, 2.7_039, 0.8_100, -0.5_908, -3.2_113, -4.4_343, 2.9_257, 1.3_632, 1.5_562, -2.1_489, -1.9_894, 3.0_560, 3.3_396, -0.7_328, -1.0_417, 0.0_383, 3.7_093, 3.2_343 ]) UpperCamelCase__ = torch.tensor([ -1.4_574, -2.0_569, -0.0_473, -0.6_117, 1.4_018, 0.5_769, 0.4_129, -2.7_344, 1.2_241, -2.1_397, 0.2_000, 0.3_937, 0.7_616, 2.0_453, 0.7_324, -0.3_391, -2.1_746, -2.7_744, 1.6_963, 0.6_921, 1.2_187, -1.6_172, -0.8_877, 2.2_439, 1.8_471, -0.5_839, -0.5_605, -0.0_464, 2.3_250, 2.1_219 ]) # fmt: on UpperCamelCase__ = api.list_models(filter="diffusers") for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": UpperCamelCase__ = '''/home/patrick/google_checkpoints/''' + mod.modelId.split("/")[-1] print(F"""Started running {mod.modelId}!!!""") if mod.modelId.startswith("CompVis"): UpperCamelCase__ = UNetaDModel.from_pretrained(local_checkpoint, subfolder="unet") else: UpperCamelCase__ = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) UpperCamelCase__ = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) UpperCamelCase__ = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): UpperCamelCase__ = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results["_".join("_".join(mod.modelId.split("/")).split("-"))], atol=1e-3 ) print(F"""{mod.modelId} has passed successfully!!!""")
719
import importlib import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Union import torch from ..utils import BaseOutput UpperCamelCase__ = "scheduler_config.json" class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : List[Any] = 1 lowercase__ : Tuple = 2 lowercase__ : Union[str, Any] = 3 lowercase__ : Union[str, Any] = 4 lowercase__ : str = 5 lowercase__ : Any = 6 lowercase__ : Any = 7 lowercase__ : List[str] = 8 lowercase__ : Union[str, Any] = 9 lowercase__ : int = 10 lowercase__ : List[str] = 11 lowercase__ : List[Any] = 12 lowercase__ : str = 13 lowercase__ : Optional[int] = 14 @dataclass class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : torch.FloatTensor class _lowerCAmelCase : """simple docstring""" lowercase__ : Optional[int] = SCHEDULER_CONFIG_NAME lowercase__ : int = [] lowercase__ : Dict = True @classmethod def snake_case__ ( cls : str , lowercase : Dict[str, Any] = None , lowercase : Optional[str] = None , lowercase : Any=False , **lowercase : List[str] , ) -> Union[str, Any]: """simple docstring""" __lowercase , __lowercase , __lowercase = cls.load_config( pretrained_model_name_or_path=lowercase , subfolder=lowercase , return_unused_kwargs=lowercase , return_commit_hash=lowercase , **lowercase , ) return cls.from_config(lowercase , return_unused_kwargs=lowercase , **lowercase ) def snake_case__ ( self : Dict , lowercase : Union[str, os.PathLike] , lowercase : bool = False , **lowercase : List[str] ) -> Optional[Any]: """simple docstring""" self.save_config(save_directory=lowercase , push_to_hub=lowercase , **lowercase ) @property def snake_case__ ( self : Optional[Any] ) -> Dict: """simple docstring""" return self._get_compatibles() @classmethod def snake_case__ ( cls : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = list(set([cls.__name__] + cls._compatibles ) ) __lowercase = importlib.import_module(__name__.split(""".""" )[0] ) __lowercase = [ getattr(lowercase , lowercase ) for c in compatible_classes_str if hasattr(lowercase , lowercase ) ] return compatible_classes
634
0
import requests UpperCamelCase__ = """""" # <-- Put your OpenWeatherMap appid here! UpperCamelCase__ = """https://api.openweathermap.org/data/2.5/""" def UpperCAmelCase__ ( lowercase__ = "Chicago" , lowercase__ = APPID ) -> dict: return requests.get(URL_BASE + """weather""" , params=locals() ).json() def UpperCAmelCase__ ( lowercase__ = "Kolkata, India" , lowercase__ = APPID ) -> dict: return requests.get(URL_BASE + """forecast""" , params=locals() ).json() def UpperCAmelCase__ ( lowercase__ = 55.68 , lowercase__ = 12.57 , lowercase__ = APPID ) -> dict: return requests.get(URL_BASE + """onecall""" , params=locals() ).json() if __name__ == "__main__": from pprint import pprint while True: UpperCamelCase__ = input("Enter a location:").strip() if location: pprint(current_weather(location)) else: break
720
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "asapp/sew-tiny-100k": "https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json", # See all SEW models at https://huggingface.co/models?filter=sew } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """sew""" def __init__( self : List[Any] , lowercase : int=32 , lowercase : List[str]=768 , lowercase : Dict=12 , lowercase : str=12 , lowercase : str=3_072 , lowercase : Optional[int]=2 , lowercase : List[str]="gelu" , lowercase : List[str]=0.1 , lowercase : Tuple=0.1 , lowercase : Dict=0.1 , lowercase : Any=0.0 , lowercase : Dict=0.1 , lowercase : Optional[int]=0.1 , lowercase : List[str]=0.02 , lowercase : Dict=1E-5 , lowercase : Tuple="group" , lowercase : int="gelu" , lowercase : Any=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , lowercase : Optional[int]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowercase : Any=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowercase : List[str]=False , lowercase : Tuple=128 , lowercase : int=16 , lowercase : Union[str, Any]=True , lowercase : List[str]=0.05 , lowercase : Optional[int]=10 , lowercase : Any=2 , lowercase : Optional[Any]=0.0 , lowercase : Optional[Any]=10 , lowercase : int=0 , lowercase : Optional[int]="mean" , lowercase : List[Any]=False , lowercase : str=False , lowercase : int=256 , lowercase : str=0 , lowercase : List[Any]=1 , lowercase : List[Any]=2 , **lowercase : List[Any] , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase , pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase ) __lowercase = hidden_size __lowercase = feat_extract_norm __lowercase = feat_extract_activation __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = list(lowercase ) __lowercase = conv_bias __lowercase = num_conv_pos_embeddings __lowercase = num_conv_pos_embedding_groups __lowercase = len(self.conv_dim ) __lowercase = num_hidden_layers __lowercase = intermediate_size __lowercase = squeeze_factor __lowercase = hidden_act __lowercase = num_attention_heads __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = feat_proj_dropout __lowercase = final_dropout __lowercase = layerdrop __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" F"but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)" F"= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowercase = apply_spec_augment __lowercase = mask_time_prob __lowercase = mask_time_length __lowercase = mask_time_min_masks __lowercase = mask_feature_prob __lowercase = mask_feature_length __lowercase = mask_feature_min_masks # ctc loss __lowercase = ctc_loss_reduction __lowercase = ctc_zero_infinity # sequence classification __lowercase = use_weighted_layer_sum __lowercase = classifier_proj_size @property def snake_case__ ( self : Dict ) -> str: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
634
0
import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowercase__ : Optional[int] = ["""input_ids""", """attention_mask"""] def __init__( self : Union[str, Any] , lowercase : Tuple="</s>" , lowercase : str="<unk>" , lowercase : Tuple="<pad>" , lowercase : List[Any]=125 , lowercase : List[Any]=None , **lowercase : Any , ) -> None: """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: __lowercase = [F"<extra_id_{i}>" for i in range(_lowercase )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens __lowercase = len(set(filter(lambda lowercase : bool("""extra_id""" in str(_lowercase ) ) , _lowercase ) ) ) if extra_tokens != extra_ids: raise ValueError( F"Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are" """ provided to ByT5Tokenizer. In this case the additional_special_tokens must include the""" """ extra_ids tokens""" ) __lowercase = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else pad_token __lowercase = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else eos_token __lowercase = AddedToken(_lowercase , lstrip=_lowercase , rstrip=_lowercase ) if isinstance(_lowercase , _lowercase ) else unk_token super().__init__( eos_token=_lowercase , unk_token=_lowercase , pad_token=_lowercase , extra_ids=_lowercase , additional_special_tokens=_lowercase , **_lowercase , ) __lowercase = extra_ids __lowercase = 2**8 # utf is 8 bits # define special tokens dict __lowercase = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } __lowercase = len(self.special_tokens_encoder ) __lowercase = len(_lowercase ) for i, token in enumerate(_lowercase ): __lowercase = self.vocab_size + i - n __lowercase = {v: k for k, v in self.special_tokens_encoder.items()} @property def snake_case__ ( self : List[str] ) -> int: """simple docstring""" return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def snake_case__ ( self : Optional[int] , lowercase : str , lowercase : int = None , lowercase : List[Any] = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowercase , token_ids_a=_lowercase , already_has_special_tokens=_lowercase ) # normal case: some special tokens if token_ids_a is None: return ([0] * len(_lowercase )) + [1] return ([0] * len(_lowercase )) + [1] + ([0] * len(_lowercase )) + [1] def snake_case__ ( self : Optional[Any] , lowercase : Optional[Any] ) -> List[int]: """simple docstring""" if len(_lowercase ) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F"This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated" """ eos tokens being added.""" ) return token_ids else: return token_ids + [self.eos_token_id] def snake_case__ ( self : Optional[int] , lowercase : int , lowercase : Dict = None ) -> List[int]: """simple docstring""" __lowercase = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def snake_case__ ( self : List[Any] , lowercase : List[Any] , lowercase : Any = None ) -> List[int]: """simple docstring""" __lowercase = self._add_eos_if_not_present(_lowercase ) if token_ids_a is None: return token_ids_a else: __lowercase = self._add_eos_if_not_present(_lowercase ) return token_ids_a + token_ids_a def snake_case__ ( self : Union[str, Any] , lowercase : Dict ) -> List[str]: """simple docstring""" __lowercase = [chr(_lowercase ) for i in text.encode("""utf-8""" )] return tokens def snake_case__ ( self : List[Any] , lowercase : Optional[Any] ) -> int: """simple docstring""" if token in self.special_tokens_encoder: __lowercase = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: __lowercase = self.added_tokens_encoder[token] elif len(_lowercase ) != 1: __lowercase = self.unk_token_id else: __lowercase = ord(_lowercase ) + self._num_special_tokens return token_id def snake_case__ ( self : Union[str, Any] , lowercase : str ) -> Optional[Any]: """simple docstring""" if index in self.special_tokens_decoder: __lowercase = self.special_tokens_decoder[index] else: __lowercase = chr(index - self._num_special_tokens ) return token def snake_case__ ( self : str , lowercase : Union[str, Any] ) -> List[Any]: """simple docstring""" __lowercase = b"""""" for token in tokens: if token in self.special_tokens_decoder: __lowercase = self.special_tokens_decoder[token].encode("""utf-8""" ) elif token in self.added_tokens_decoder: __lowercase = self.special_tokens_decoder[token].encode("""utf-8""" ) elif token in self.special_tokens_encoder: __lowercase = token.encode("""utf-8""" ) elif token in self.added_tokens_encoder: __lowercase = token.encode("""utf-8""" ) else: __lowercase = bytes([ord(_lowercase )] ) bstring += tok_string __lowercase = bstring.decode("""utf-8""" , errors="""ignore""" ) return string def snake_case__ ( self : Dict , lowercase : int , lowercase : str = None ) -> Tuple[str]: """simple docstring""" return ()
721
import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizerOutput from transformers.testing_utils import require_phonemizer from ...test_tokenization_common import TokenizerTesterMixin @require_phonemizer class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = WavaVecaPhonemeCTCTokenizer lowercase__ : Optional[int] = False def snake_case__ ( self : str ) -> int: """simple docstring""" super().setUp() __lowercase = ( """<s> <pad> </s> <unk> n s t ə l a i k d m ɛ ɾ e ɪ p o ɐ z ð f j v b ɹ ʁ ʊ iː r w ʌ u ɡ æ aɪ ʃ h ɔ ɑː """ """ŋ ɚ eɪ β uː y ɑ̃ oʊ ᵻ eː θ aʊ ts oː ɔ̃ ɣ ɜ ɑ dʒ əl x ɜː ç ʒ tʃ ɔː ɑːɹ ɛ̃ ʎ ɔːɹ ʋ aː ɕ œ ø oːɹ ɲ yː """ """ʔ iə i5 s. tɕ ?? nʲ ɛː œ̃ ɭ ɔø ʑ tʲ ɨ ɛɹ ts. rʲ ɪɹ ɭʲ i.5 ɔɪ q sʲ u5 ʊɹ iɜ a5 iɛ5 øː ʕ ja əɜ th ɑ5 """ """oɪ dʲ ə5 tɕh ts.h mʲ ɯ dʑ vʲ e̞ tʃʲ ei5 o5 onɡ5 ɑu5 iɑ5 ai5 aɪɚ kh ə1 ʐ i2 ʉ ħ t[ aɪə ʲ ju ə2 u2 oɜ """ """pː iɛɜ ou5 y5 uɜ tː uo5 d[ uoɜ tsh ɑɜ ɵ i̪5 uei5 ɟ aɜ ɑɨ i.ɜ eʊ o2 ɐ̃ ä pʲ kʲ n̩ ɒ ph ɑu2 uɨ əɪ ɫ ɬ """ """yɜ bʲ ɑ2 s̪ aiɜ χ ɐ̃ʊ̃ 1 ə4 yæɜ a2 ɨː t̪ iouɜ ũ onɡɜ aɨ iɛ2 ɔɨ ɑuɜ o̞ ei2 iou2 c kː y2 ɖ oe dˤ yɛɜ """ """əʊ S ɡʲ onɡ2 u\" eiɜ ʈ ɯᵝ iou5 dZ r̝̊ i.2 tS s^ ʝ yə5 iɑɜ uə5 pf ɨu iɑ2 ou2 ər2 fʲ ai2 r̝ uəɜ ɳ əɨ """ """ua5 uɪ ɽ bː yu5 uo2 yɛ5 l̩ ɻ ərɜ ʂ i̪2 ouɜ uaɜ a. a.ː yæ5 dː r̩ ee ɪu ər5 i̪ ɜ æi u: i.ː t^ o1 ɪ^ """ """ai ueiɜ æː ɛɪ eə i. ɴ ie ua2 ɑ1 o4 tʃː o: ɑ: u1 N i̪1 au yæ2 u. qː yəɜ y: kʰ tʃʰ iʊ sx õ uo tʰ """ """uai5 bʰ u.ː uə2 ʊə d^ s̪ː yiɜ dʰ r. oe: i1 ɟː yu2 nʲʲ i̪4 uei2 tsʲ ɸ ĩ ɑ4 t̪ː eɑ u4 e: tsː ʈʰ ɡʰ """ """ɯɯ dʒʲ ʂʲ X ɵː uaiɜ tɕʲ ã t^ː ẽː yɛ2 cː i.1 ɛʊ dˤdˤ dʒː i4 ɡː yi ɕʲ ɟʰ pʰ dʑʲ yuɜ ua1 ua4 æiː ɐɐ """ """ui iou1 ʊː a1 iou4 cʰ iɛ1 yə2 ɖʰ ẽ ʒʲ ää ər4 iːː ɪː iɑ1 ər1 œː øi ɪuː cʰcʰ əː1 iː1 ũ kʰː o̞o̞ xʲ """ """ou1 iɛ4 e̞e̞ y1 dzː dʲʲ dʰː ɯᵝɯᵝ lː uo1 i.4 i: yɛ5ʲ a4""" ).split(""" """ ) __lowercase = dict(zip(lowercase , range(len(lowercase ) ) ) ) __lowercase = {"""pad_token""": """<pad>""", """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>"""} __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowercase ) + """\n""" ) def snake_case__ ( self : List[Any] , lowercase : Optional[Any] , lowercase : List[str]=False , lowercase : List[str]=20 , lowercase : str=5 ) -> Tuple[str, list]: """simple docstring""" __lowercase = [(i, tokenizer.decode([i] , clean_up_tokenization_spaces=lowercase )) for i in range(len(lowercase ) )] __lowercase = list(filter(lambda lowercase : [t[0]] == tokenizer.encode(t[1] , do_phonemize=lowercase ) , lowercase ) ) if max_length is not None and len(lowercase ) > max_length: __lowercase = toks[:max_length] if min_length is not None and len(lowercase ) < min_length and len(lowercase ) > 0: while len(lowercase ) < min_length: __lowercase = toks + toks # toks_str = [t[1] for t in toks] __lowercase = [t[0] for t in toks] # Ensure consistency __lowercase = tokenizer.decode(lowercase , clean_up_tokenization_spaces=lowercase ) if " " not in output_txt and len(lowercase ) > 1: __lowercase = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=lowercase ) + """ """ + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=lowercase ) ) if with_prefix_space: __lowercase = """ """ + output_txt __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) return output_txt, output_ids def snake_case__ ( self : Tuple , **lowercase : int ) -> Union[str, Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return WavaVecaPhonemeCTCTokenizer.from_pretrained(self.tmpdirname , **lowercase ) def snake_case__ ( self : Any ) -> List[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) # check adding a single token tokenizer.add_tokens("""xxx""" ) __lowercase = tokenizer("""m xxx ɪ""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 392, 17] ) # xxx should be last token tokenizer.add_tokens(["""aaa""", """bbb""", """ccc"""] ) __lowercase = tokenizer("""m aaa ɪ ccc""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [13, 393, 17, 395] ) # aaa and ccc should be after xxx and 2 after aaa __lowercase = tokenizer("""maɪ c""" , do_phonemize=lowercase ).input_ids self.assertEqual(lowercase , [3, 200] ) # mai should be <unk> (=3) def snake_case__ ( self : int ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) def snake_case__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98], [24, 22, 5, 24, 22, 5, 77], ] __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> Optional[Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(lowercase , """h ə l oʊ | h aʊ | ɑːɹ | j uː |""" ) def snake_case__ ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) self.assertEqual(tokenizer(lowercase ).input_ids , tokenizer(lowercase , do_phonemize=lowercase ).input_ids ) def snake_case__ ( self : List[Any] ) -> Tuple: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, tokenizer.word_delimiter_token_id, 15, 8, tokenizer.word_delimiter_token_id, 98], [tokenizer.word_delimiter_token_id, 24, 22, tokenizer.word_delimiter_token_id, 5, 24, 22, 5, 77], ] # fmt: on # decode with word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] ) __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ""", """j ð s j ð s oːɹ"""] ) # decode with no word_del_token filter __lowercase = tokenizer.decode(sample_ids[0] , filter_word_delimiter_token=lowercase ) __lowercase = tokenizer.batch_decode(lowercase , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , batch_tokens[0] ) self.assertEqual(lowercase , ["""k s ɾ | ɾ l | ɭʲ""", """| j ð | s j ð s oːɹ"""] ) def snake_case__ ( self : int ) -> str: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : Optional[int] ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) __lowercase = """Hello how are you""" __lowercase = tokenizer.phonemize(lowercase , phonemizer_lang="""en-us""" ) __lowercase = tokenizer.decode(tokenizer(lowercase ).input_ids , filter_word_delimiter_token=lowercase ) self.assertEqual(""" """.join([p.strip() for p in phonemes.split(""" |""" )] ).strip() , lowercase ) def snake_case__ ( self : str ) -> List[str]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained( """facebook/wav2vec2-lv-60-espeak-cv-ft""" , word_delimiter_token=lowercase ) __lowercase = """Hello how are you""" __lowercase = tokenizer(lowercase , phonemizer_lang="""en-us""" ).input_ids __lowercase = tokenizer(lowercase , phonemizer_lang="""fr-fr""" ).input_ids self.assertNotEqual(lowercase , lowercase ) __lowercase = tokenizer.decode(lowercase ) __lowercase = tokenizer.decode(lowercase ) self.assertEqual(lowercase , """h ə l oʊ h aʊ ɑːɹ j uː""" ) self.assertEqual(lowercase , """ɛ l o h aʊ a ʁ j u""" ) def snake_case__ ( self : int ) -> int: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) __lowercase = """Hello how Are you""" __lowercase = """hello how are you""" __lowercase = tokenizer(lowercase ).input_ids __lowercase = tokenizer(lowercase ).input_ids self.assertEqual(lowercase , lowercase ) def snake_case__ ( self : int ) -> Union[str, Any]: """simple docstring""" __lowercase = self.tokenizer_class.from_pretrained("""facebook/wav2vec2-lv-60-espeak-cv-ft""" ) tokenizer.add_tokens(["""!""", """?"""] ) tokenizer.add_special_tokens({"""cls_token""": """$$$"""} ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98, 392, 392, 393, 392, 392, 393, 394, 394], [24, 22, 5, 24, 22, 5, 77, tokenizer.pad_token_id, 394, 394], ] # fmt: on __lowercase = tokenizer.batch_decode(lowercase ) self.assertEqual(lowercase , ["""k s ɾ ɾ l ɭʲ!?!? $$$""", """j ð s j ð s oːɹ $$$"""] ) @staticmethod def snake_case__ ( lowercase : List[str] , lowercase : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase = [d[key] for d in offsets] return retrieved_list def snake_case__ ( self : str ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) tokenizer.add_tokens("""|""" ) # fmt: off # ksssɾɾ|ɾɾ<pad>ɾɾ|<pad>ɾlll|ɭʲ -> k s ɾ ɾ | ɾ l | ɭʲ" __lowercase = [11, 5, 5, 5, 15, 15, tokenizer.pad_token_id, 15, 15, tokenizer.word_delimiter_token_id, tokenizer.pad_token_id, 15, 8, 8, 8, tokenizer.word_delimiter_token_id, 98] # fmt: on __lowercase = tokenizer.decode(lowercase , output_char_offsets=lowercase , filter_word_delimiter_token=lowercase ) # check Wav2Vec2CTCTokenizerOutput keys for char self.assertEqual(len(outputs.keys() ) , 2 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""char_offsets""" in outputs ) self.assertTrue(isinstance(lowercase , lowercase ) ) # check that order of chars is correct and identical for both outputs self.assertEqual(""" """.join(self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) ) , outputs.text ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """char""" ) , ["""k""", """s""", """ɾ""", """ɾ""", """|""", """ɾ""", """l""", """|""", """ɭʲ"""] ) # check that offsets are actually correct for char # 0-1 is 11, 1-4 is 5, 4-6 is first 15, 6-7 is <pad> (thus not shown), 7-9 is second 15, 9-10 is word_delimiter_token, # 10-11 is <pad> (thus not shown), 11-12 is third 15, 12-15 is 8, 15-16 is word_delimiter_token, 16-17 is 98 self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """start_offset""" ) , [0, 1, 4, 7, 9, 11, 12, 15, 16] ) self.assertListEqual( self.get_from_offsets(outputs["""char_offsets"""] , """end_offset""" ) , [1, 4, 6, 9, 10, 12, 15, 16, 17] ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase = self.get_tokenizer(word_delimiter_token="""|""" ) def check_list_tuples_equal(lowercase : List[str] , lowercase : Dict ): self.assertTrue(isinstance(lowercase , lowercase ) ) self.assertTrue(isinstance(outputs_list[0] , lowercase ) ) # transform list to ModelOutput __lowercase = WavaVecaPhonemeCTCTokenizerOutput( {k: [d[k] for d in outputs_list] for k in outputs_list[0]} ) self.assertListEqual(outputs_batch["""text"""] , outputs_batch_a["""text"""] ) def recursive_check(lowercase : List[Any] , lowercase : Optional[int] ): if isinstance(lowercase , lowercase ): [recursive_check(lowercase , lowercase ) for la, la in zip(lowercase , lowercase )] self.assertEqual(lowercase , lowercase ) if "char_offsets" in outputs_batch: recursive_check(outputs_batch["""char_offsets"""] , outputs_batch_a["""char_offsets"""] ) # fmt: off __lowercase = [ [11, 5, 15, tokenizer.pad_token_id, 15, 4, 8, 98, 32, 32, 32, 32, 4, 33, tokenizer.word_delimiter_token_id, 32, 32, 33, 34, 34], [24, 22, 5, tokenizer.word_delimiter_token_id, tokenizer.word_delimiter_token_id, 24, 22, 22, 22, 4, 5, 77, tokenizer.pad_token_id, 22, 22, 4, 34, 34, 34, 34], ] # fmt: on # We assume that `decode` works as expected. All we will check now is # the output type is correct and the output is identical to `decode` # char __lowercase = tokenizer.batch_decode(lowercase , output_char_offsets=lowercase ) __lowercase = [tokenizer.decode(lowercase , output_char_offsets=lowercase ) for ids in sample_ids] check_list_tuples_equal(lowercase , lowercase ) @unittest.skip("""Wav2Vec2PhonemeTokenizer always lower cases letters to correctly map to phonemes""" ) def snake_case__ ( self : List[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeTokenizer always puts spaces between phonemes""" ) def snake_case__ ( self : List[str] ) -> List[str]: """simple docstring""" pass @unittest.skip("""encodes to text to ids, but decodes ids to phonemes -> not possible to have internal consistency""" ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" pass @unittest.skip("""Wav2Vec2PhonemeModel has no max model length => no testing""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass def snake_case__ ( self : Any ) -> int: """simple docstring""" __lowercase = self.get_tokenizers(do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) __lowercase = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] __lowercase = tokenizer.add_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size + len(lowercase ) ) __lowercase = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) __lowercase = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} __lowercase = tokenizer.add_special_tokens(lowercase ) __lowercase = tokenizer.vocab_size __lowercase = len(lowercase ) self.assertNotEqual(lowercase , 0 ) self.assertEqual(lowercase , lowercase ) self.assertEqual(lowercase , len(lowercase ) ) self.assertEqual(lowercase , all_size_a + len(lowercase ) ) __lowercase = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=lowercase ) self.assertGreaterEqual(len(lowercase ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : str ) -> int: """simple docstring""" pass @unittest.skip("""The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode.""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" pass def snake_case__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = self.get_tokenizers(fast=lowercase , do_lower_case=lowercase ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = ["""ð""", """ɪ""", """s""", """ɪ""", """z""", """ɐ""", """t""", """ɛ""", """k""", """s""", """t"""] __lowercase = tokenizer.convert_tokens_to_string(lowercase ) self.assertIsInstance(output["""text"""] , lowercase )
634
0
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available UpperCamelCase__ = logging.getLogger(__name__) @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : str lowercase__ : List[str] lowercase__ : Optional[List[str]] @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : List[int] lowercase__ : List[int] lowercase__ : Optional[List[int]] = None lowercase__ : Optional[List[int]] = None class _lowerCAmelCase ( _lowercase ): """simple docstring""" lowercase__ : Optional[int] = '''train''' lowercase__ : Tuple = '''dev''' lowercase__ : Optional[Any] = '''test''' class _lowerCAmelCase : """simple docstring""" @staticmethod def snake_case__ ( lowercase : List[str] , lowercase : Union[Split, str] ) -> List[InputExample]: """simple docstring""" raise NotImplementedError @staticmethod def snake_case__ ( lowercase : str ) -> List[str]: """simple docstring""" raise NotImplementedError @staticmethod def snake_case__ ( lowercase : List[InputExample] , lowercase : List[str] , lowercase : int , lowercase : PreTrainedTokenizer , lowercase : int=False , lowercase : Optional[int]="[CLS]" , lowercase : Union[str, Any]=1 , lowercase : Union[str, Any]="[SEP]" , lowercase : int=False , lowercase : List[Any]=False , lowercase : Any=0 , lowercase : List[str]=0 , lowercase : int=-100 , lowercase : List[str]=0 , lowercase : Tuple=True , ) -> List[InputFeatures]: """simple docstring""" __lowercase = {label: i for i, label in enumerate(A_ )} __lowercase = [] for ex_index, example in enumerate(A_ ): if ex_index % 10_000 == 0: logger.info("""Writing example %d of %d""" , A_ , len(A_ ) ) __lowercase = [] __lowercase = [] for word, label in zip(example.words , example.labels ): __lowercase = tokenizer.tokenize(A_ ) # bert-base-multilingual-cased sometimes output "nothing ([]) when calling tokenize with just a space. if len(A_ ) > 0: tokens.extend(A_ ) # Use the real label id for the first token of the word, and padding ids for the remaining tokens label_ids.extend([label_map[label]] + [pad_token_label_id] * (len(A_ ) - 1) ) # Account for [CLS] and [SEP] with "- 2" and with "- 3" for RoBERTa. __lowercase = tokenizer.num_special_tokens_to_add() if len(A_ ) > max_seq_length - special_tokens_count: __lowercase = tokens[: (max_seq_length - special_tokens_count)] __lowercase = label_ids[: (max_seq_length - special_tokens_count)] # The convention in BERT is: # (a) For sequence pairs: # tokens: [CLS] is this jack ##son ##ville ? [SEP] no it is not . [SEP] # type_ids: 0 0 0 0 0 0 0 0 1 1 1 1 1 1 # (b) For single sequences: # tokens: [CLS] the dog is hairy . [SEP] # type_ids: 0 0 0 0 0 0 0 # # Where "type_ids" are used to indicate whether this is the first # sequence or the second sequence. The embedding vectors for `type=0` and # `type=1` were learned during pre-training and are added to the wordpiece # embedding vector (and position vector). This is not *strictly* necessary # since the [SEP] token unambiguously separates the sequences, but it makes # it easier for the model to learn the concept of sequences. # # For classification tasks, the first vector (corresponding to [CLS]) is # used as the "sentence vector". Note that this only makes sense because # the entire model is fine-tuned. tokens += [sep_token] label_ids += [pad_token_label_id] if sep_token_extra: # roberta uses an extra separator b/w pairs of sentences tokens += [sep_token] label_ids += [pad_token_label_id] __lowercase = [sequence_a_segment_id] * len(A_ ) if cls_token_at_end: tokens += [cls_token] label_ids += [pad_token_label_id] segment_ids += [cls_token_segment_id] else: __lowercase = [cls_token] + tokens __lowercase = [pad_token_label_id] + label_ids __lowercase = [cls_token_segment_id] + segment_ids __lowercase = tokenizer.convert_tokens_to_ids(A_ ) # The mask has 1 for real tokens and 0 for padding tokens. Only real # tokens are attended to. __lowercase = [1 if mask_padding_with_zero else 0] * len(A_ ) # Zero-pad up to the sequence length. __lowercase = max_seq_length - len(A_ ) if pad_on_left: __lowercase = ([pad_token] * padding_length) + input_ids __lowercase = ([0 if mask_padding_with_zero else 1] * padding_length) + input_mask __lowercase = ([pad_token_segment_id] * padding_length) + segment_ids __lowercase = ([pad_token_label_id] * padding_length) + label_ids else: input_ids += [pad_token] * padding_length input_mask += [0 if mask_padding_with_zero else 1] * padding_length segment_ids += [pad_token_segment_id] * padding_length label_ids += [pad_token_label_id] * padding_length assert len(A_ ) == max_seq_length assert len(A_ ) == max_seq_length assert len(A_ ) == max_seq_length assert len(A_ ) == max_seq_length if ex_index < 5: logger.info("""*** Example ***""" ) logger.info("""guid: %s""" , example.guid ) logger.info("""tokens: %s""" , """ """.join([str(A_ ) for x in tokens] ) ) logger.info("""input_ids: %s""" , """ """.join([str(A_ ) for x in input_ids] ) ) logger.info("""input_mask: %s""" , """ """.join([str(A_ ) for x in input_mask] ) ) logger.info("""segment_ids: %s""" , """ """.join([str(A_ ) for x in segment_ids] ) ) logger.info("""label_ids: %s""" , """ """.join([str(A_ ) for x in label_ids] ) ) if "token_type_ids" not in tokenizer.model_input_names: __lowercase = None features.append( InputFeatures( input_ids=A_ , attention_mask=A_ , token_type_ids=A_ , label_ids=A_ ) ) return features if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset class _lowerCAmelCase ( _lowercase ): """simple docstring""" lowercase__ : List[InputFeatures] lowercase__ : int = nn.CrossEntropyLoss().ignore_index def __init__( self : Tuple , lowercase : TokenClassificationTask , lowercase : str , lowercase : PreTrainedTokenizer , lowercase : List[str] , lowercase : str , lowercase : Optional[int] = None , lowercase : Optional[int]=False , lowercase : Split = Split.train , ) -> Dict: """simple docstring""" __lowercase = os.path.join( A_ , """cached_{}_{}_{}""".format(mode.value , tokenizer.__class__.__name__ , str(A_ ) ) , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __lowercase = cached_features_file + """.lock""" with FileLock(A_ ): if os.path.exists(A_ ) and not overwrite_cache: logger.info(F"Loading features from cached file {cached_features_file}" ) __lowercase = torch.load(A_ ) else: logger.info(F"Creating features from dataset file at {data_dir}" ) __lowercase = token_classification_task.read_examples_from_file(A_ , A_ ) # TODO clean up all this to leverage built-in features of tokenizers __lowercase = token_classification_task.convert_examples_to_features( A_ , A_ , A_ , A_ , cls_token_at_end=bool(model_type in ["""xlnet"""] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ["""xlnet"""] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=A_ , pad_on_left=bool(tokenizer.padding_side == """left""" ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info(F"Saving features into cached file {cached_features_file}" ) torch.save(self.features , A_ ) def __len__( self : Optional[int] ) -> List[str]: """simple docstring""" return len(self.features ) def __getitem__( self : List[Any] , lowercase : Optional[int] ) -> InputFeatures: """simple docstring""" return self.features[i] if is_tf_available(): import tensorflow as tf class _lowerCAmelCase : """simple docstring""" lowercase__ : List[InputFeatures] lowercase__ : int = -100 def __init__( self : Any , lowercase : TokenClassificationTask , lowercase : str , lowercase : PreTrainedTokenizer , lowercase : List[str] , lowercase : str , lowercase : Optional[int] = None , lowercase : str=False , lowercase : Split = Split.train , ) -> Tuple: """simple docstring""" __lowercase = token_classification_task.read_examples_from_file(A_ , A_ ) # TODO clean up all this to leverage built-in features of tokenizers __lowercase = token_classification_task.convert_examples_to_features( A_ , A_ , A_ , A_ , cls_token_at_end=bool(model_type in ["""xlnet"""] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ["""xlnet"""] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=A_ , pad_on_left=bool(tokenizer.padding_side == """left""" ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) def gen(): for ex in self.features: if ex.token_type_ids is None: yield ( {"input_ids": ex.input_ids, "attention_mask": ex.attention_mask}, ex.label_ids, ) else: yield ( { "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label_ids, ) if "token_type_ids" not in tokenizer.model_input_names: __lowercase = tf.data.Dataset.from_generator( A_ , ({"""input_ids""": tf.intaa, """attention_mask""": tf.intaa}, tf.intaa) , ( {"""input_ids""": tf.TensorShape([None] ), """attention_mask""": tf.TensorShape([None] )}, tf.TensorShape([None] ), ) , ) else: __lowercase = tf.data.Dataset.from_generator( A_ , ({"""input_ids""": tf.intaa, """attention_mask""": tf.intaa, """token_type_ids""": tf.intaa}, tf.intaa) , ( { """input_ids""": tf.TensorShape([None] ), """attention_mask""": tf.TensorShape([None] ), """token_type_ids""": tf.TensorShape([None] ), }, tf.TensorShape([None] ), ) , ) def snake_case__ ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase = self.dataset.apply(tf.data.experimental.assert_cardinality(len(self.features ) ) ) return self.dataset def __len__( self : Optional[Any] ) -> str: """simple docstring""" return len(self.features ) def __getitem__( self : Union[str, Any] , lowercase : Optional[Any] ) -> InputFeatures: """simple docstring""" return self.features[i]
700
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) # TODO Update this UpperCamelCase__ = { "facebook/esm-1b": "https://huggingface.co/facebook/esm-1b/resolve/main/config.json", # See all ESM models at https://huggingface.co/models?filter=esm } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """esm""" def __init__( self : Any , lowercase : Optional[Any]=None , lowercase : Optional[int]=None , lowercase : List[Any]=None , lowercase : Optional[int]=768 , lowercase : str=12 , lowercase : Union[str, Any]=12 , lowercase : Dict=3_072 , lowercase : Optional[int]=0.1 , lowercase : str=0.1 , lowercase : Dict=1_026 , lowercase : Tuple=0.02 , lowercase : str=1E-1_2 , lowercase : Dict="absolute" , lowercase : Optional[Any]=True , lowercase : int=None , lowercase : int=False , lowercase : List[str]=False , lowercase : Tuple=None , lowercase : Tuple=None , **lowercase : Union[str, Any] , ) -> int: """simple docstring""" super().__init__(pad_token_id=lowercase , mask_token_id=lowercase , **lowercase ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = emb_layer_norm_before __lowercase = token_dropout __lowercase = is_folding_model if is_folding_model: if esmfold_config is None: logger.info("""No esmfold_config supplied for folding model, using default values.""" ) __lowercase = EsmFoldConfig() elif isinstance(lowercase , lowercase ): __lowercase = EsmFoldConfig(**lowercase ) __lowercase = esmfold_config if vocab_list is None: logger.warning("""No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!""" ) __lowercase = get_default_vocab_list() else: __lowercase = vocab_list else: __lowercase = None __lowercase = None if self.esmfold_config is not None and getattr(self.esmfold_config , """use_esm_attn_map""" , lowercase ): raise ValueError("""The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!""" ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase = super().to_dict() if isinstance(self.esmfold_config , lowercase ): __lowercase = self.esmfold_config.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : str = None lowercase__ : bool = True lowercase__ : bool = False lowercase__ : bool = False lowercase__ : bool = False lowercase__ : float = 0 lowercase__ : bool = True lowercase__ : bool = False lowercase__ : int = 128 lowercase__ : "TrunkConfig" = None def snake_case__ ( self : List[str] ) -> Any: """simple docstring""" if self.trunk is None: __lowercase = TrunkConfig() elif isinstance(self.trunk , lowercase ): __lowercase = TrunkConfig(**self.trunk ) def snake_case__ ( self : Dict ) -> Any: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.trunk.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 48 lowercase__ : int = 1_024 lowercase__ : int = 128 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : int = 32 lowercase__ : float = 0 lowercase__ : float = 0 lowercase__ : bool = False lowercase__ : int = 4 lowercase__ : Optional[int] = 128 lowercase__ : "StructureModuleConfig" = None def snake_case__ ( self : Tuple ) -> str: """simple docstring""" if self.structure_module is None: __lowercase = StructureModuleConfig() elif isinstance(self.structure_module , lowercase ): __lowercase = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(F"`max_recycles` should be positive, got {self.max_recycles}." ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( """`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got""" F" {self.sequence_state_dim} and {self.sequence_state_dim}." ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( """`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got""" F" {self.pairwise_state_dim} and {self.pairwise_state_dim}." ) __lowercase = self.sequence_state_dim // self.sequence_head_width __lowercase = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( """`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got""" F" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}." ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( """`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got""" F" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}." ) if self.pairwise_state_dim % 2 != 0: raise ValueError(F"`pairwise_state_dim` should be even, got {self.pairwise_state_dim}." ) if self.dropout >= 0.4: raise ValueError(F"`dropout` should not be greater than 0.4, got {self.dropout}." ) def snake_case__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = asdict(self ) __lowercase = self.structure_module.to_dict() return output @dataclass class _lowerCAmelCase : """simple docstring""" lowercase__ : int = 384 lowercase__ : int = 128 lowercase__ : int = 16 lowercase__ : int = 128 lowercase__ : int = 12 lowercase__ : int = 4 lowercase__ : int = 8 lowercase__ : float = 0.1 lowercase__ : int = 8 lowercase__ : int = 1 lowercase__ : int = 2 lowercase__ : int = 7 lowercase__ : int = 10 lowercase__ : float = 1E-8 lowercase__ : float = 1E5 def snake_case__ ( self : Tuple ) -> List[str]: """simple docstring""" return asdict(self ) def UpperCAmelCase__ ( ) -> List[Any]: return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
634
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available from .timesteps import ( fastaa_timesteps, smartaa_timesteps, smartaa_timesteps, smartaaa_timesteps, smartaaa_timesteps, superaa_timesteps, superaa_timesteps, superaaa_timesteps, ) @dataclass class _lowerCAmelCase ( lowercase__ ): """simple docstring""" lowercase__ : Union[List[PIL.Image.Image], np.ndarray] lowercase__ : Optional[List[bool]] lowercase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_if import IFPipeline from .pipeline_if_imgaimg import IFImgaImgPipeline from .pipeline_if_imgaimg_superresolution import IFImgaImgSuperResolutionPipeline from .pipeline_if_inpainting import IFInpaintingPipeline from .pipeline_if_inpainting_superresolution import IFInpaintingSuperResolutionPipeline from .pipeline_if_superresolution import IFSuperResolutionPipeline from .safety_checker import IFSafetyChecker from .watermark import IFWatermarker
701
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCAmelCase ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : Tuple = LxmertTokenizer lowercase__ : List[str] = LxmertTokenizerFast lowercase__ : Optional[Any] = True lowercase__ : List[Any] = True def snake_case__ ( self : Tuple ) -> Tuple: """simple docstring""" super().setUp() __lowercase = [ """[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def snake_case__ ( self : Optional[int] , lowercase : int ) -> List[Any]: """simple docstring""" __lowercase = """UNwant\u00E9d,running""" __lowercase = """unwanted, running""" return input_text, output_text def snake_case__ ( self : str ) -> Any: """simple docstring""" __lowercase = self.tokenizer_class(self.vocab_file ) __lowercase = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(lowercase , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , [7, 4, 5, 10, 8, 9] ) def snake_case__ ( self : Union[str, Any] ) -> int: """simple docstring""" if not self.test_rust_tokenizer: return __lowercase = self.get_tokenizer() __lowercase = self.get_rust_tokenizer() __lowercase = """I was born in 92000, and this is falsé.""" __lowercase = tokenizer.tokenize(lowercase ) __lowercase = rust_tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) __lowercase = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase ) self.assertListEqual(lowercase , lowercase ) __lowercase = self.get_rust_tokenizer() __lowercase = tokenizer.encode(lowercase ) __lowercase = rust_tokenizer.encode(lowercase ) self.assertListEqual(lowercase , lowercase )
634
0
UpperCamelCase__ = { 0: "0", 1: "1", 2: "2", 3: "3", 4: "4", 5: "5", 6: "6", 7: "7", 8: "8", 9: "9", 10: "a", 11: "b", 12: "c", 13: "d", 14: "e", 15: "f", } def UpperCAmelCase__ ( lowercase__ ): assert type(lowercase__ ) in (int, float) and decimal == int(lowercase__ ) __lowercase = int(lowercase__ ) __lowercase = """""" __lowercase = False if decimal < 0: __lowercase = True decimal *= -1 while decimal > 0: __lowercase , __lowercase = divmod(lowercase__ , 16 ) __lowercase = values[remainder] + hexadecimal __lowercase = """0x""" + hexadecimal if negative: __lowercase = """-""" + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
702
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> bool: __lowercase = len(lowercase__ ) __lowercase = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): __lowercase = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): __lowercase = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: __lowercase = subset[i - 1][j] if arr[i - 1] <= j: __lowercase = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
634
0
import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/config.json""", # See all BART models at https://huggingface.co/models?filter=bart } class _lowerCAmelCase ( UpperCamelCase_ ): """simple docstring""" lowercase__ : Optional[int] = """bart""" lowercase__ : Any = ["""past_key_values"""] lowercase__ : Any = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : int , lowercase : Dict=50_265 , lowercase : Dict=1_024 , lowercase : Tuple=12 , lowercase : Union[str, Any]=4_096 , lowercase : Tuple=16 , lowercase : Tuple=12 , lowercase : str=4_096 , lowercase : Union[str, Any]=16 , lowercase : Optional[int]=0.0 , lowercase : Optional[Any]=0.0 , lowercase : List[str]="gelu" , lowercase : Optional[Any]=1_024 , lowercase : Any=0.1 , lowercase : Optional[int]=0.0 , lowercase : int=0.0 , lowercase : Tuple=0.02 , lowercase : str=0.0 , lowercase : Optional[Any]=False , lowercase : Optional[int]=True , lowercase : List[Any]=3 , lowercase : Union[str, Any]=1 , lowercase : Dict=0 , lowercase : int=2 , lowercase : int=True , lowercase : Optional[Any]=2 , lowercase : Tuple=2 , **lowercase : Any , ) -> str: """simple docstring""" __lowercase = vocab_size __lowercase = max_position_embeddings __lowercase = d_model __lowercase = encoder_ffn_dim __lowercase = encoder_layers __lowercase = encoder_attention_heads __lowercase = decoder_ffn_dim __lowercase = decoder_layers __lowercase = decoder_attention_heads __lowercase = dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = activation_function __lowercase = init_std __lowercase = encoder_layerdrop __lowercase = decoder_layerdrop __lowercase = classifier_dropout __lowercase = use_cache __lowercase = encoder_layers __lowercase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=__A , pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , is_encoder_decoder=__A , decoder_start_token_id=__A , forced_eos_token_id=__A , **__A , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , __A ): __lowercase = self.bos_token_id warnings.warn( F"Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. " """The config can simply be saved and uploaded again to be fixed.""" ) class _lowerCAmelCase ( UpperCamelCase_ ): """simple docstring""" @property def snake_case__ ( self : Optional[int] ) -> Tuple: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: __lowercase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: __lowercase = {0: "batch"} __lowercase = {0: "batch", 1: "past_decoder_sequence + sequence"} else: __lowercase = {0: "batch", 1: "decoder_sequence"} __lowercase = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(__A , direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. __lowercase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: __lowercase = self.num_layers for i in range(__A ): __lowercase = {0: "batch", 2: "past_sequence + sequence"} __lowercase = {0: "batch", 2: "past_sequence + sequence"} else: __lowercase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property def snake_case__ ( self : int ) -> Tuple: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: __lowercase = super().outputs else: __lowercase = super(__A , self ).outputs if self.use_past: __lowercase = self.num_layers for i in range(__A ): __lowercase = {0: "batch", 2: "past_sequence + sequence"} __lowercase = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def snake_case__ ( self : Any , lowercase : PreTrainedTokenizer , lowercase : int = -1 , lowercase : int = -1 , lowercase : bool = False , lowercase : Optional[TensorType] = None , ) -> Optional[Any]: """simple docstring""" __lowercase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __A , __A , __A , __A , __A ) # Generate decoder inputs __lowercase = seq_length if not self.use_past else 1 __lowercase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __A , __A , __A , __A , __A ) __lowercase = {F"decoder_{name}": tensor for name, tensor in decoder_inputs.items()} __lowercase = dict(**__A , **__A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch __lowercase = common_inputs["input_ids"].shape __lowercase = common_inputs["decoder_input_ids"].shape[1] __lowercase = self.num_attention_heads __lowercase = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __lowercase = decoder_seq_length + 3 __lowercase = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __lowercase = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(__A , __A )] , dim=1 ) __lowercase = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __lowercase = self.num_layers __lowercase = min(__A , __A ) __lowercase = max(__A , __A ) - min_num_layers __lowercase = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(__A ): common_inputs["past_key_values"].append( ( torch.zeros(__A ), torch.zeros(__A ), torch.zeros(__A ), torch.zeros(__A ), ) ) # TODO: test this. __lowercase = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(__A , __A ): common_inputs["past_key_values"].append((torch.zeros(__A ), torch.zeros(__A )) ) return common_inputs def snake_case__ ( self : List[Any] , lowercase : PreTrainedTokenizer , lowercase : int = -1 , lowercase : int = -1 , lowercase : bool = False , lowercase : Optional[TensorType] = None , ) -> Union[str, Any]: """simple docstring""" __lowercase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __A , __A , __A , __A , __A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch __lowercase = common_inputs["input_ids"].shape # Not using the same length for past_key_values __lowercase = seqlen + 2 __lowercase = self.num_layers __lowercase = self.num_attention_heads __lowercase = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __lowercase = common_inputs["attention_mask"].dtype __lowercase = torch.cat( [common_inputs["""attention_mask"""], torch.ones(__A , __A , dtype=__A )] , dim=1 ) __lowercase = [ (torch.zeros(__A ), torch.zeros(__A )) for _ in range(__A ) ] return common_inputs def snake_case__ ( self : str , lowercase : PreTrainedTokenizer , lowercase : int = -1 , lowercase : int = -1 , lowercase : bool = False , lowercase : Optional[TensorType] = None , ) -> str: """simple docstring""" __lowercase = compute_effective_axis_dimension( __A , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __lowercase = tokenizer.num_special_tokens_to_add(__A ) __lowercase = compute_effective_axis_dimension( __A , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__A ) # Generate dummy inputs according to compute batch and sequence __lowercase = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size __lowercase = dict(tokenizer(__A , return_tensors=__A ) ) return common_inputs def snake_case__ ( self : Tuple , lowercase : PreTrainedTokenizer , lowercase : int = -1 , lowercase : int = -1 , lowercase : bool = False , lowercase : Optional[TensorType] = None , ) -> int: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: __lowercase = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __A , batch_size=__A , seq_length=__A , is_pair=__A , framework=__A ) elif self.task == "causal-lm": __lowercase = self._generate_dummy_inputs_for_causal_lm( __A , batch_size=__A , seq_length=__A , is_pair=__A , framework=__A ) else: __lowercase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __A , batch_size=__A , seq_length=__A , is_pair=__A , framework=__A ) return common_inputs def snake_case__ ( self : int , lowercase : Dict , lowercase : Dict , lowercase : Any , lowercase : Optional[int] ) -> Optional[Any]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: __lowercase = super()._flatten_past_key_values_(__A , __A , __A , __A ) else: __lowercase = super(__A , self )._flatten_past_key_values_( __A , __A , __A , __A )
703
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "hustvl/yolos-small": "https://huggingface.co/hustvl/yolos-small/resolve/main/config.json", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Union[str, Any] = """yolos""" def __init__( self : Optional[int] , lowercase : Any=768 , lowercase : Tuple=12 , lowercase : Tuple=12 , lowercase : str=3_072 , lowercase : Optional[Any]="gelu" , lowercase : Union[str, Any]=0.0 , lowercase : Dict=0.0 , lowercase : Optional[int]=0.02 , lowercase : Optional[Any]=1E-1_2 , lowercase : Tuple=[512, 864] , lowercase : Optional[int]=16 , lowercase : Dict=3 , lowercase : Optional[Any]=True , lowercase : Optional[int]=100 , lowercase : Optional[int]=True , lowercase : Any=False , lowercase : Any=1 , lowercase : Any=5 , lowercase : List[str]=2 , lowercase : Union[str, Any]=5 , lowercase : str=2 , lowercase : Tuple=0.1 , **lowercase : str , ) -> Optional[Any]: """simple docstring""" super().__init__(**lowercase ) __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = qkv_bias __lowercase = num_detection_tokens __lowercase = use_mid_position_embeddings __lowercase = auxiliary_loss # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = eos_coefficient class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Dict = version.parse("""1.11""" ) @property def snake_case__ ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def snake_case__ ( self : int ) -> float: """simple docstring""" return 1E-4 @property def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" return 12
634
0
import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def UpperCAmelCase__ ( lowercase__="" ) -> Tuple: __lowercase = tempfile.mkdtemp() return os.path.join(_lowerCamelCase , str(uuid.uuida() ) + suffix ) @require_soundfile @require_torch class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = torch.rand(12 , dtype=torch.floataa ) - 0.5 __lowercase = AgentAudio(lowercase ) __lowercase = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowercase , agent_type.to_raw() , atol=1E-4 ) ) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(lowercase ) ) # Ensure that the file contains the same value as the original tensor __lowercase = sf.read(lowercase ) self.assertTrue(torch.allclose(lowercase , torch.tensor(lowercase ) , atol=1E-4 ) ) def snake_case__ ( self : List[str] ) -> str: """simple docstring""" __lowercase = torch.rand(12 , dtype=torch.floataa ) - 0.5 __lowercase = get_new_path(suffix=""".wav""" ) sf.write(lowercase , lowercase , 16_000 ) __lowercase = AgentAudio(lowercase ) self.assertTrue(torch.allclose(lowercase , agent_type.to_raw() , atol=1E-4 ) ) self.assertEqual(agent_type.to_string() , lowercase ) @require_vision @require_torch class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Any ) -> Tuple: """simple docstring""" __lowercase = torch.randint(0 , 256 , (64, 64, 3) ) __lowercase = AgentImage(lowercase ) __lowercase = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowercase , agent_type._tensor , atol=1E-4 ) ) self.assertIsInstance(agent_type.to_raw() , Image.Image ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowercase ) ) def snake_case__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase = Path(get_tests_dir("""fixtures/tests_samples/COCO""" ) ) / "000000039769.png" __lowercase = Image.open(lowercase ) __lowercase = AgentImage(lowercase ) self.assertTrue(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowercase ) ) def snake_case__ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" __lowercase = Path(get_tests_dir("""fixtures/tests_samples/COCO""" ) ) / "000000039769.png" __lowercase = Image.open(lowercase ) __lowercase = AgentImage(lowercase ) self.assertFalse(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowercase ) ) class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : List[str] ) -> int: """simple docstring""" __lowercase = "Hey!" __lowercase = AgentText(lowercase ) self.assertEqual(lowercase , agent_type.to_string() ) self.assertEqual(lowercase , agent_type.to_raw() ) self.assertEqual(lowercase , lowercase )
704
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowerCAmelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase__ : int = IFImgaImgSuperResolutionPipeline lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} lowercase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) lowercase__ : Tuple = PipelineTesterMixin.required_optional_params - {"""latents"""} def snake_case__ ( self : Tuple ) -> Any: """simple docstring""" return self._get_superresolution_dummy_components() def snake_case__ ( self : List[str] , lowercase : Optional[int] , lowercase : Optional[Any]=0 ) -> Union[str, Any]: """simple docstring""" if str(lowercase ).startswith("""mps""" ): __lowercase = torch.manual_seed(lowercase ) else: __lowercase = torch.Generator(device=lowercase ).manual_seed(lowercase ) __lowercase = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase ) ).to(lowercase ) __lowercase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """original_image""": original_image, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def snake_case__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def snake_case__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def snake_case__ ( self : List[Any] ) -> Dict: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1E-1 ) def snake_case__ ( self : Dict ) -> int: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def snake_case__ ( self : Optional[Any] ) -> int: """simple docstring""" self._test_save_load_local() def snake_case__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
634
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "openai/imagegpt-small": "", "openai/imagegpt-medium": "", "openai/imagegpt-large": "", } class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ : Optional[Any] = """imagegpt""" lowercase__ : int = ["""past_key_values"""] lowercase__ : int = { """hidden_size""": """n_embd""", """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : str , lowercase : int=512 + 1 , lowercase : List[str]=32 * 32 , lowercase : int=512 , lowercase : Optional[int]=24 , lowercase : List[str]=8 , lowercase : Dict=None , lowercase : Optional[Any]="quick_gelu" , lowercase : str=0.1 , lowercase : Optional[Any]=0.1 , lowercase : List[str]=0.1 , lowercase : str=1E-5 , lowercase : Union[str, Any]=0.02 , lowercase : int=True , lowercase : Optional[int]=True , lowercase : Optional[Any]=False , lowercase : List[Any]=False , lowercase : str=False , **lowercase : Tuple , ) -> str: """simple docstring""" __lowercase = vocab_size __lowercase = n_positions __lowercase = n_embd __lowercase = n_layer __lowercase = n_head __lowercase = n_inner __lowercase = activation_function __lowercase = resid_pdrop __lowercase = embd_pdrop __lowercase = attn_pdrop __lowercase = layer_norm_epsilon __lowercase = initializer_range __lowercase = scale_attn_weights __lowercase = use_cache __lowercase = scale_attn_by_inverse_layer_idx __lowercase = reorder_and_upcast_attn __lowercase = tie_word_embeddings super().__init__(tie_word_embeddings=_a , **_a ) class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): """simple docstring""" @property def snake_case__ ( self : List[str] ) -> str: """simple docstring""" return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ] ) def snake_case__ ( self : int , lowercase : "FeatureExtractionMixin" , lowercase : int = 1 , lowercase : int = -1 , lowercase : bool = False , lowercase : Optional["TensorType"] = None , lowercase : int = 3 , lowercase : int = 32 , lowercase : int = 32 , ) -> Any: """simple docstring""" __lowercase = self._generate_dummy_images(_a , _a , _a , _a ) __lowercase = dict(preprocessor(images=_a , return_tensors=_a ) ) return inputs
705
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ = { "configuration_luke": ["LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP", "LukeConfig"], "tokenization_luke": ["LukeTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "LUKE_PRETRAINED_MODEL_ARCHIVE_LIST", "LukeForEntityClassification", "LukeForEntityPairClassification", "LukeForEntitySpanClassification", "LukeForMultipleChoice", "LukeForQuestionAnswering", "LukeForSequenceClassification", "LukeForTokenClassification", "LukeForMaskedLM", "LukeModel", "LukePreTrainedModel", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
0
import operator as op UpperCamelCase__ = "scaler.pt" UpperCamelCase__ = "pytorch_model" UpperCamelCase__ = "random_states" UpperCamelCase__ = "optimizer" UpperCamelCase__ = "scheduler" UpperCamelCase__ = "pytorch_model.bin" UpperCamelCase__ = "pytorch_model.bin.index.json" UpperCamelCase__ = "model.safetensors" UpperCamelCase__ = "model.safetensors.index.json" UpperCamelCase__ = "1.10.2" UpperCamelCase__ = "py38" UpperCamelCase__ = "4.17.0" UpperCamelCase__ = ["ml.p3.16xlarge", "ml.p3dn.24xlarge", "ml.p4dn.24xlarge"] UpperCamelCase__ = ["FULL_SHARD", "SHARD_GRAD_OP", "NO_SHARD", "HYBRID_SHARD", "HYBRID_SHARD_ZERO2"] UpperCamelCase__ = ["TRANSFORMER_BASED_WRAP", "SIZE_BASED_WRAP", "NO_WRAP"] UpperCamelCase__ = ["BACKWARD_PRE", "BACKWARD_POST", "NO_PREFETCH"] UpperCamelCase__ = ["FULL_STATE_DICT", "LOCAL_STATE_DICT", "SHARDED_STATE_DICT"] UpperCamelCase__ = "2.0.1" UpperCamelCase__ = ["pdsh", "standard", "openmpi", "mvapich"] UpperCamelCase__ = ["default", "reduce-overhead", "max-autotune"] UpperCamelCase__ = {">": op.gt, ">=": op.ge, "==": op.eq, "!=": op.ne, "<=": op.le, "<": op.lt} # These are the args for `torch.distributed.launch` for pytorch < 1.9 UpperCamelCase__ = [ "nnodes", "nproc_per_node", "rdzv_backend", "rdzv_endpoint", "rdzv_id", "rdzv_conf", "standalone", "max_restarts", "monitor_interval", "start_method", "role", "module", "m", "no_python", "run_path", "log_dir", "r", "redirects", "t", "tee", "node_rank", "master_addr", "master_port", ] UpperCamelCase__ = ["DEEPSPEED", "MULTI_GPU", "FSDP", "MEGATRON_LM"] UpperCamelCase__ = ["DEEPSPEED", "MULTI_XPU", "FSDP"]
706
def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> int: if len(lowercase__ ) != len(lowercase__ ): raise ValueError("""String lengths must match!""" ) __lowercase = 0 for chara, chara in zip(lowercase__ , lowercase__ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
634
0
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { "edbeeching/decision-transformer-gym-hopper-medium": ( "https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json" ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" lowercase__ : Tuple = """decision_transformer""" lowercase__ : Any = ["""past_key_values"""] lowercase__ : Tuple = { """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Union[str, Any] , lowercase : str=17 , lowercase : List[str]=4 , lowercase : Optional[int]=128 , lowercase : Union[str, Any]=4_096 , lowercase : List[str]=True , lowercase : List[str]=1 , lowercase : List[Any]=1_024 , lowercase : Optional[int]=3 , lowercase : str=1 , lowercase : List[str]=None , lowercase : Optional[int]="relu" , lowercase : Dict=0.1 , lowercase : List[str]=0.1 , lowercase : Optional[Any]=0.1 , lowercase : Tuple=1E-5 , lowercase : List[str]=0.02 , lowercase : List[str]=True , lowercase : List[str]=True , lowercase : Tuple=50_256 , lowercase : Tuple=50_256 , lowercase : Dict=False , lowercase : Tuple=False , **lowercase : Optional[Any] , ) -> Dict: """simple docstring""" __lowercase = state_dim __lowercase = act_dim __lowercase = hidden_size __lowercase = max_ep_len __lowercase = action_tanh __lowercase = vocab_size __lowercase = n_positions __lowercase = n_layer __lowercase = n_head __lowercase = n_inner __lowercase = activation_function __lowercase = resid_pdrop __lowercase = embd_pdrop __lowercase = attn_pdrop __lowercase = layer_norm_epsilon __lowercase = initializer_range __lowercase = scale_attn_weights __lowercase = use_cache __lowercase = scale_attn_by_inverse_layer_idx __lowercase = reorder_and_upcast_attn __lowercase = bos_token_id __lowercase = eos_token_id super().__init__(bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase )
707
from __future__ import annotations from collections.abc import Callable UpperCamelCase__ = list[list[float | int]] def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Matrix: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(size + 1 )] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for row in range(lowercase__ ): for col in range(lowercase__ ): __lowercase = matrix[row][col] __lowercase = vector[row][0] __lowercase = 0 __lowercase = 0 while row < size and col < size: # pivoting __lowercase = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase__ , lowercase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: __lowercase , __lowercase = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase__ ): __lowercase = augmented[rowa][col] / augmented[row][col] __lowercase = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase__ ): for row in range(lowercase__ ): __lowercase = augmented[row][col] / augmented[col][col] for cola in range(lowercase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase__ ) ] def UpperCAmelCase__ ( lowercase__ ) -> Callable[[int], int]: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(lowercase__ )] for _ in range(lowercase__ )] __lowercase = [[0] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for x_val, y_val in enumerate(lowercase__ ): for col in range(lowercase__ ): __lowercase = (x_val + 1) ** (size - col - 1) __lowercase = y_val __lowercase = solve(lowercase__ , lowercase__ ) def interpolated_func(lowercase__ ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase__ ) ) return interpolated_func def UpperCAmelCase__ ( lowercase__ ) -> int: return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCAmelCase__ ( lowercase__ = question_function , lowercase__ = 10 ) -> int: __lowercase = [func(lowercase__ ) for x_val in range(1 , order + 1 )] __lowercase = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] __lowercase = 0 __lowercase = 42 __lowercase = 42 for poly in polynomials: __lowercase = 1 while func(lowercase__ ) == poly(lowercase__ ): x_val += 1 ret += poly(lowercase__ ) return ret if __name__ == "__main__": print(F"""{solution() = }""")
634
0
from __future__ import annotations from fractions import Fraction def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> bool: return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def UpperCAmelCase__ ( lowercase__ ) -> list[str]: __lowercase = [] __lowercase = 11 __lowercase = int("""1""" + """0""" * digit_len ) for num in range(UpperCAmelCase__ , UpperCAmelCase__ ): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(UpperCAmelCase__ , UpperCAmelCase__ ): solutions.append(F"{num}/{den}" ) den += 1 num += 1 __lowercase = 10 return solutions def UpperCAmelCase__ ( lowercase__ = 2 ) -> int: __lowercase = 1.0 for fraction in fraction_list(UpperCAmelCase__ ): __lowercase = Fraction(UpperCAmelCase__ ) result *= frac.denominator / frac.numerator return int(UpperCAmelCase__ ) if __name__ == "__main__": print(solution())
708
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ = { "configuration_ctrl": ["CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP", "CTRLConfig"], "tokenization_ctrl": ["CTRLTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "CTRLForSequenceClassification", "CTRLLMHeadModel", "CTRLModel", "CTRLPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ "TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCTRLForSequenceClassification", "TFCTRLLMHeadModel", "TFCTRLModel", "TFCTRLPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ctrl import CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRLConfig from .tokenization_ctrl import CTRLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ctrl import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, CTRLPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_ctrl import ( TF_CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, TFCTRLForSequenceClassification, TFCTRLLMHeadModel, TFCTRLModel, TFCTRLPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
634
0
from functools import reduce UpperCamelCase__ = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCAmelCase__ ( lowercase__ = N ) -> int: return max( # mypy cannot properly interpret reduce int(reduce(lambda lowercase__ , lowercase__ : str(int(_lowercase ) * int(_lowercase ) ) , n[i : i + 13] ) ) for i in range(len(_lowercase ) - 12 ) ) if __name__ == "__main__": print(F"""{solution() = }""")
709
import unittest import numpy as np def UpperCAmelCase__ ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = None , ) -> np.ndarray: __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) __lowercase = np.shape(lowercase__ ) if shape_a[0] != shape_b[0]: __lowercase = ( """Expected the same number of rows for A and B. """ F"Instead found A of size {shape_a} and B of size {shape_b}" ) raise ValueError(lowercase__ ) if shape_b[1] != shape_c[1]: __lowercase = ( """Expected the same number of columns for B and C. """ F"Instead found B of size {shape_b} and C of size {shape_c}" ) raise ValueError(lowercase__ ) __lowercase = pseudo_inv if a_inv is None: try: __lowercase = np.linalg.inv(lowercase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Dict ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) __lowercase = schur_complement(lowercase , lowercase , lowercase ) __lowercase = np.block([[a, b], [b.T, c]] ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) __lowercase = np.linalg.det(lowercase ) self.assertAlmostEqual(lowercase , det_a * det_s ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1], [6, 3]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) def snake_case__ ( self : Tuple ) -> None: """simple docstring""" __lowercase = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) __lowercase = np.array([[0, 3], [3, 0], [2, 3]] ) __lowercase = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(lowercase ): schur_complement(lowercase , lowercase , lowercase ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
634
0