code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import collections import os import re from pathlib import Path _UpperCamelCase : Union[str, Any] = "src/transformers" # Matches is_xxx_available() _UpperCamelCase : int = re.compile(r"is\_([a-z_]*)_available()") # Catches a one-line _import_struct = {xxx} _UpperCamelCase : List[str] = re.compile(r"^_import_structure\s+=\s+\{([^\}]+)\}") # Catches a line with a key-values pattern: "bla": ["foo", "bar"] _UpperCamelCase : List[Any] = re.compile(r"\s+\"\S*\":\s+\[([^\]]*)\]") # Catches a line if not is_foo_available _UpperCamelCase : Optional[Any] = re.compile(r"^\s*if\s+not\s+is\_[a-z_]*\_available\(\)") # Catches a line _import_struct["bla"].append("foo") _UpperCamelCase : Any = re.compile(r"^\s*_import_structure\[\"\S*\"\]\.append\(\"(\S*)\"\)") # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] _UpperCamelCase : Tuple = re.compile(r"^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]") # Catches a line with an object between quotes and a comma: "MyModel", _UpperCamelCase : Tuple = re.compile(r"^\s+\"([^\"]+)\",") # Catches a line with objects between brackets only: ["foo", "bar"], _UpperCamelCase : str = re.compile(r"^\s+\[([^\]]+)\]") # Catches a line with from foo import bar, bla, boo _UpperCamelCase : Dict = re.compile(r"\s+from\s+\S*\s+import\s+([^\(\s].*)\n") # Catches a line with try: _UpperCamelCase : Tuple = re.compile(r"^\s*try:") # Catches a line with else: _UpperCamelCase : Optional[Any] = re.compile(r"^\s*else:") def a_ ( _lowerCAmelCase : Optional[int] ): '''simple docstring''' if _re_test_backend.search(_lowerCAmelCase ) is None: return None lowercase__ : Union[str, Any] = [b[0] for b in _re_backend.findall(_lowerCAmelCase )] backends.sort() return "_and_".join(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' with open(_lowerCAmelCase , 'r' , encoding='utf-8' , newline='\n' ) as f: lowercase__ : List[str] = f.readlines() lowercase__ : List[Any] = 0 while line_index < len(_lowerCAmelCase ) and not lines[line_index].startswith('_import_structure = {' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(_lowerCAmelCase ): return None # First grab the objects without a specific backend in _import_structure lowercase__ : List[Any] = [] while not lines[line_index].startswith('if TYPE_CHECKING' ) and find_backend(lines[line_index] ) is None: lowercase__ : Any = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(_lowerCAmelCase ): lowercase__ : int = _re_one_line_import_struct.search(_lowerCAmelCase ).groups()[0] lowercase__ : Union[str, Any] = re.findall(R'\[([^\]]+)\]' , _lowerCAmelCase ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(', ' )] ) line_index += 1 continue lowercase__ : Tuple = _re_import_struct_key_value.search(_lowerCAmelCase ) if single_line_import_search is not None: lowercase__ : List[str] = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(', ' ) if len(_lowerCAmelCase ) > 0] objects.extend(_lowerCAmelCase ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) line_index += 1 lowercase__ : str = {'none': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('if TYPE_CHECKING' ): # If the line is an if not is_backend_available, we grab all objects associated. lowercase__ : Optional[int] = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowercase__ : Tuple = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowercase__ : int = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 4 ): lowercase__ : int = lines[line_index] if _re_import_struct_add_one.search(_lowerCAmelCase ) is not None: objects.append(_re_import_struct_add_one.search(_lowerCAmelCase ).groups()[0] ) elif _re_import_struct_add_many.search(_lowerCAmelCase ) is not None: lowercase__ : List[Any] = _re_import_struct_add_many.search(_lowerCAmelCase ).groups()[0].split(', ' ) lowercase__ : int = [obj[1:-1] for obj in imports if len(_lowerCAmelCase ) > 0] objects.extend(_lowerCAmelCase ) elif _re_between_brackets.search(_lowerCAmelCase ) is not None: lowercase__ : Dict = _re_between_brackets.search(_lowerCAmelCase ).groups()[0].split(', ' ) lowercase__ : Any = [obj[1:-1] for obj in imports if len(_lowerCAmelCase ) > 0] objects.extend(_lowerCAmelCase ) elif _re_quote_object.search(_lowerCAmelCase ) is not None: objects.append(_re_quote_object.search(_lowerCAmelCase ).groups()[0] ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) elif line.startswith(' ' * 12 + '"' ): objects.append(line[13:-3] ) line_index += 1 lowercase__ : Optional[int] = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend lowercase__ : List[Any] = [] while ( line_index < len(_lowerCAmelCase ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('else' ) ): lowercase__ : Any = lines[line_index] lowercase__ : List[str] = _re_import.search(_lowerCAmelCase ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 8 ): objects.append(line[8:-2] ) line_index += 1 lowercase__ : Union[str, Any] = {'none': objects} # Let's continue with backend-specific objects while line_index < len(_lowerCAmelCase ): # If the line is an if is_backend_available, we grab all objects associated. lowercase__ : Tuple = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowercase__ : Tuple = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowercase__ : Any = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 8 ): lowercase__ : Dict = lines[line_index] lowercase__ : Optional[Any] = _re_import.search(_lowerCAmelCase ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 12 ): objects.append(line[12:-2] ) line_index += 1 lowercase__ : str = objects else: line_index += 1 return import_dict_objects, type_hint_objects def a_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict ): '''simple docstring''' def find_duplicates(_lowerCAmelCase : Union[str, Any] ): return [k for k, v in collections.Counter(_lowerCAmelCase ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] lowercase__ : Optional[Any] = [] for key in import_dict_objects.keys(): lowercase__ : Optional[int] = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(f"""Duplicate _import_structure definitions for: {duplicate_imports}""" ) lowercase__ : Union[str, Any] = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(f"""Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}""" ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): lowercase__ : int = 'base imports' if key == 'none' else f"""{key} backend""" errors.append(f"""Differences for {name}:""" ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(f""" {a} in TYPE_HINT but not in _import_structure.""" ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(f""" {a} in _import_structure but not in TYPE_HINT.""" ) return errors def a_ ( ): '''simple docstring''' lowercase__ : List[str] = [] for root, _, files in os.walk(_lowerCAmelCase ): if "__init__.py" in files: lowercase__ : List[Any] = os.path.join(_lowerCAmelCase , '__init__.py' ) lowercase__ : Dict = parse_init(_lowerCAmelCase ) if objects is not None: lowercase__ : List[Any] = analyze_results(*_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: lowercase__ : int = f"""Problem in {fname}, both halves do not define the same objects.\n{errors[0]}""" failures.append('\n'.join(_lowerCAmelCase ) ) if len(_lowerCAmelCase ) > 0: raise ValueError('\n\n'.join(_lowerCAmelCase ) ) def a_ ( ): '''simple docstring''' lowercase__ : Optional[int] = [] for path, directories, files in os.walk(_lowerCAmelCase ): for folder in directories: # Ignore private modules if folder.startswith('_' ): directories.remove(_lowerCAmelCase ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(_lowerCAmelCase ) / folder).glob('*.py' ) ) ) == 0: continue lowercase__ : Dict = str((Path(_lowerCAmelCase ) / folder).relative_to(_lowerCAmelCase ) ) lowercase__ : Dict = short_path.replace(os.path.sep , '.' ) submodules.append(_lowerCAmelCase ) for fname in files: if fname == "__init__.py": continue lowercase__ : List[str] = str((Path(_lowerCAmelCase ) / fname).relative_to(_lowerCAmelCase ) ) lowercase__ : Optional[Any] = short_path.replace('.py' , '' ).replace(os.path.sep , '.' ) if len(submodule.split('.' ) ) == 1: submodules.append(_lowerCAmelCase ) return submodules _UpperCamelCase : Union[str, Any] = [ "convert_pytorch_checkpoint_to_tf2", "modeling_flax_pytorch_utils", "models.esm.openfold_utils", ] def a_ ( ): '''simple docstring''' from transformers.utils import direct_transformers_import lowercase__ : Tuple = direct_transformers_import(_lowerCAmelCase ) lowercase__ : Optional[Any] = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(_lowerCAmelCase , '__init__.py' ) , 'r' ) as f: lowercase__ : Any = f.read() import_structure_keys.update(set(re.findall(R'import_structure\[\"([^\"]*)\"\]' , _lowerCAmelCase ) ) ) lowercase__ : Union[str, Any] = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(_lowerCAmelCase ) > 0: lowercase__ : int = '\n'.join(f"""- {module}""" for module in module_not_registered ) raise ValueError( 'The following submodules are not properly registed in the main init of Transformers:\n' f"""{list_of_modules}\n""" 'Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.' ) if __name__ == "__main__": check_all_inits() check_submodules()
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase : Tuple = { "configuration_whisper": ["WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP", "WhisperConfig", "WhisperOnnxConfig"], "feature_extraction_whisper": ["WhisperFeatureExtractor"], "processing_whisper": ["WhisperProcessor"], "tokenization_whisper": ["WhisperTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = ["WhisperTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : str = [ "WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "WhisperForConditionalGeneration", "WhisperModel", "WhisperPreTrainedModel", "WhisperForAudioClassification", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Any = [ "TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFWhisperForConditionalGeneration", "TFWhisperModel", "TFWhisperPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Tuple = [ "FlaxWhisperForConditionalGeneration", "FlaxWhisperModel", "FlaxWhisperPreTrainedModel", "FlaxWhisperForAudioClassification", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _UpperCamelCase : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging _UpperCamelCase : List[str] = logging.get_logger(__name__) _UpperCamelCase : Dict = "▁" _UpperCamelCase : Any = {"vocab_file": "sentencepiece.bpe.model"} _UpperCamelCase : List[str] = { "vocab_file": { "facebook/mbart-large-50-one-to-many-mmt": ( "https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt/resolve/main/sentencepiece.bpe.model" ), } } _UpperCamelCase : str = { "facebook/mbart-large-50-one-to-many-mmt": 10_24, } # fmt: off _UpperCamelCase : Optional[Any] = ["ar_AR", "cs_CZ", "de_DE", "en_XX", "es_XX", "et_EE", "fi_FI", "fr_XX", "gu_IN", "hi_IN", "it_IT", "ja_XX", "kk_KZ", "ko_KR", "lt_LT", "lv_LV", "my_MM", "ne_NP", "nl_XX", "ro_RO", "ru_RU", "si_LK", "tr_TR", "vi_VN", "zh_CN", "af_ZA", "az_AZ", "bn_IN", "fa_IR", "he_IL", "hr_HR", "id_ID", "ka_GE", "km_KH", "mk_MK", "ml_IN", "mn_MN", "mr_IN", "pl_PL", "ps_AF", "pt_XX", "sv_SE", "sw_KE", "ta_IN", "te_IN", "th_TH", "tl_XX", "uk_UA", "ur_PK", "xh_ZA", "gl_ES", "sl_SI"] class UpperCAmelCase_ ( _a): lowerCamelCase__ : Tuple = VOCAB_FILES_NAMES lowerCamelCase__ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Any = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : Optional[Any] = ["input_ids", "attention_mask"] lowerCamelCase__ : List[int] = [] lowerCamelCase__ : List[int] = [] def __init__( self , a , a=None , a=None , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a = None , **a , ) -> None: # Mask token behave like a normal word, i.e. include the space before it lowercase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token lowercase__ : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs lowercase__ : Any = kwargs.get('additional_special_tokens' , [] ) kwargs["additional_special_tokens"] += [ code for code in FAIRSEQ_LANGUAGE_CODES if code not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=a , tgt_lang=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , sp_model_kwargs=self.sp_model_kwargs , **a , ) lowercase__ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(a ) ) lowercase__ : Optional[Any] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token lowercase__ : List[str] = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab lowercase__ : Optional[int] = 1 lowercase__ : str = len(self.sp_model ) lowercase__ : int = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(a ) } lowercase__ : List[str] = {v: k for k, v in self.lang_code_to_id.items()} lowercase__ : Dict = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) lowercase__ : Optional[Any] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} lowercase__ : List[Any] = src_lang if src_lang is not None else 'en_XX' lowercase__ : Optional[int] = self.lang_code_to_id[self._src_lang] lowercase__ : List[str] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def _UpperCAmelCase ( self ) -> int: return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def _UpperCAmelCase ( self ) -> str: return self._src_lang @src_lang.setter def _UpperCAmelCase ( self , a ) -> None: lowercase__ : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self ) -> Dict: lowercase__ : int = self.__dict__.copy() lowercase__ : Any = None return state def __setstate__( self , a ) -> None: lowercase__ : Any = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowercase__ : Tuple = {} lowercase__ : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCAmelCase ( self ) -> Dict: lowercase__ : Dict = {self.convert_ids_to_tokens(a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _UpperCAmelCase ( self , a ) -> List[str]: return self.sp_model.encode(a , out_type=a ) def _UpperCAmelCase ( self , a ) -> int: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] lowercase__ : Optional[int] = self.sp_model.PieceToId(a ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _UpperCAmelCase ( self , a ) -> str: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : str = [] lowercase__ : Optional[int] = '' lowercase__ : str = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(a ) + token lowercase__ : List[Any] = True lowercase__ : Any = [] else: current_sub_tokens.append(a ) lowercase__ : Any = False out_string += self.sp_model.decode(a ) return out_string.strip() def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Dict = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , a ) elif not os.path.isfile(self.vocab_file ): with open(a , 'wb' ) as fi: lowercase__ : int = self.sp_model.serialized_model_proto() fi.write(a ) return (out_vocab_file,) def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) lowercase__ : Dict = [1] * len(self.prefix_tokens ) lowercase__ : Tuple = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(a )) + suffix_ones return prefix_ones + ([0] * len(a )) + ([0] * len(a )) + suffix_ones def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _UpperCAmelCase ( self , a , a , a , a , **a ) -> Optional[Any]: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowercase__ : Dict = src_lang lowercase__ : Optional[Any] = self(a , add_special_tokens=a , return_tensors=a , **a ) lowercase__ : Dict = self.convert_tokens_to_ids(a ) lowercase__ : Dict = tgt_lang_id return inputs def _UpperCAmelCase ( self , a , a = "en_XX" , a = None , a = "ro_RO" , **a , ) -> BatchEncoding: lowercase__ : Optional[Any] = src_lang lowercase__ : Any = tgt_lang return super().prepare_seqaseq_batch(a , a , **a ) def _UpperCAmelCase ( self ) -> int: return self.set_src_lang_special_tokens(self.src_lang ) def _UpperCAmelCase ( self ) -> Tuple: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def _UpperCAmelCase ( self , a ) -> None: lowercase__ : List[str] = self.lang_code_to_id[src_lang] lowercase__ : Optional[Any] = [self.cur_lang_code_id] lowercase__ : List[str] = [self.eos_token_id] def _UpperCAmelCase ( self , a ) -> None: lowercase__ : str = self.lang_code_to_id[tgt_lang] lowercase__ : List[str] = [self.cur_lang_code_id] lowercase__ : List[Any] = [self.eos_token_id]
77
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a_ ( _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : int = args.pruning_method lowercase__ : Tuple = args.threshold lowercase__ : str = args.model_name_or_path.rstrip('/' ) lowercase__ : List[Any] = args.target_model_path print(f"""Load fine-pruned model from {model_name_or_path}""" ) lowercase__ : Optional[Any] = torch.load(os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) lowercase__ : List[str] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowercase__ : Tuple = tensor print(f"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowercase__ : List[str] = tensor print(f"""Copied layer {name}""" ) elif "bias" in name: lowercase__ : Optional[Any] = tensor print(f"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowercase__ : Optional[Any] = MagnitudeBinarizer.apply(inputs=_lowerCAmelCase , threshold=_lowerCAmelCase ) lowercase__ : Optional[int] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowercase__ : Optional[Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ : Any = TopKBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[Any] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowercase__ : Any = name[:-6] lowercase__ : Optional[Any] = model[f"""{prefix_}mask_scores"""] lowercase__ : Tuple = ThresholdBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[str] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowercase__ : Union[str, Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ , lowercase__ : Tuple = -0.1, 1.1 lowercase__ : Optional[Any] = torch.sigmoid(_lowerCAmelCase ) lowercase__ : Optional[Any] = s * (r - l) + l lowercase__ : Optional[Any] = s_bar.clamp(min=0.0 , max=1.0 ) lowercase__ : Union[str, Any] = tensor * mask print(f"""Pruned layer {name}""" ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: lowercase__ : Union[str, Any] = os.path.join( os.path.dirname(_lowerCAmelCase ) , f"""bertarized_{os.path.basename(_lowerCAmelCase )}""" ) if not os.path.isdir(_lowerCAmelCase ): shutil.copytree(_lowerCAmelCase , _lowerCAmelCase ) print(f"""\nCreated folder {target_model_path}""" ) torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": _UpperCamelCase : int = argparse.ArgumentParser() parser.add_argument( "--pruning_method", choices=["l0", "magnitude", "topK", "sigmoied_threshold"], type=str, required=True, help=( "Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning," " sigmoied_threshold = Soft movement pruning)" ), ) parser.add_argument( "--threshold", type=float, required=False, help=( "For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model." "For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared." "Not needed for `l0`" ), ) parser.add_argument( "--model_name_or_path", type=str, required=True, help="Folder containing the model that was previously fine-pruned", ) parser.add_argument( "--target_model_path", default=None, type=str, required=False, help="Folder containing the model that was previously fine-pruned", ) _UpperCamelCase : Dict = parser.parse_args() main(args)
77
1
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase : int = logging.get_logger(__name__) _UpperCamelCase : str = [ ("bert.bert", "visual_bert"), ("bert.cls", "cls"), ("bert.classifier", "cls"), ("token_type_embeddings_visual", "visual_token_type_embeddings"), ("position_embeddings_visual", "visual_position_embeddings"), ("projection", "visual_projection"), ] _UpperCamelCase : Dict = [ "nlvr2_coco_pre_trained.th", "nlvr2_fine_tuned.th", "nlvr2_pre_trained.th", "vcr_coco_pre_train.th", "vcr_fine_tune.th", "vcr_pre_train.th", "vqa_coco_pre_trained.th", "vqa_fine_tuned.th", "vqa_pre_trained.th", ] def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Dict = torch.load(_lowerCAmelCase , map_location='cpu' ) return sd def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str=rename_keys_prefix ): '''simple docstring''' lowercase__ : Union[str, Any] = OrderedDict() lowercase__ : List[str] = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue lowercase__ : List[str] = key for name_pair in rename_keys_prefix: lowercase__ : List[Any] = new_key.replace(name_pair[0] , name_pair[1] ) lowercase__ : List[str] = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately lowercase__ : str = new_d['cls.predictions.bias'] return new_d @torch.no_grad() def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : str ): '''simple docstring''' assert ( checkpoint_path.split('/' )[-1] in ACCEPTABLE_CHECKPOINTS ), f"""The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.""" # Get Config if "pre" in checkpoint_path: lowercase__ : List[Any] = 'pretraining' if "vcr" in checkpoint_path: lowercase__ : int = {'visual_embedding_dim': 512} elif "vqa_advanced" in checkpoint_path: lowercase__ : Tuple = {'visual_embedding_dim': 2048} elif "vqa" in checkpoint_path: lowercase__ : Union[str, Any] = {'visual_embedding_dim': 2048} elif "nlvr" in checkpoint_path: lowercase__ : List[str] = {'visual_embedding_dim': 1024} else: raise NotImplementedError(f"""No implementation found for `{checkpoint_path}`.""" ) else: if "vcr" in checkpoint_path: lowercase__ : Optional[Any] = {'visual_embedding_dim': 512} lowercase__ : int = 'multichoice' elif "vqa_advanced" in checkpoint_path: lowercase__ : int = {'visual_embedding_dim': 2048} lowercase__ : Tuple = 'vqa_advanced' elif "vqa" in checkpoint_path: lowercase__ : Union[str, Any] = {'visual_embedding_dim': 2048, 'num_labels': 3129} lowercase__ : int = 'vqa' elif "nlvr" in checkpoint_path: lowercase__ : Tuple = { 'visual_embedding_dim': 1024, 'num_labels': 2, } lowercase__ : List[str] = 'nlvr' lowercase__ : Optional[Any] = VisualBertConfig(**_lowerCAmelCase ) # Load State Dict lowercase__ : List[Any] = load_state_dict(_lowerCAmelCase ) lowercase__ : str = get_new_dict(_lowerCAmelCase , _lowerCAmelCase ) if model_type == "pretraining": lowercase__ : Optional[Any] = VisualBertForPreTraining(_lowerCAmelCase ) elif model_type == "vqa": lowercase__ : Optional[Any] = VisualBertForQuestionAnswering(_lowerCAmelCase ) elif model_type == "nlvr": lowercase__ : Dict = VisualBertForVisualReasoning(_lowerCAmelCase ) elif model_type == "multichoice": lowercase__ : List[Any] = VisualBertForMultipleChoice(_lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) # Save Checkpoints Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) model.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": _UpperCamelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument("orig_checkpoint_path", type=str, help="A path to .th on local filesystem.") parser.add_argument("pytorch_dump_folder_path", type=str, help="Path to the output PyTorch model.") _UpperCamelCase : int = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
77
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[List[PIL.Image.Image], np.ndarray] lowerCamelCase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : np.ndarray lowerCamelCase__ : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
77
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _UpperCamelCase : List[Any] = { "configuration_bridgetower": [ "BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP", "BridgeTowerConfig", "BridgeTowerTextConfig", "BridgeTowerVisionConfig", ], "processing_bridgetower": ["BridgeTowerProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : List[str] = ["BridgeTowerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : List[str] = [ "BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST", "BridgeTowerForContrastiveLearning", "BridgeTowerForImageAndTextRetrieval", "BridgeTowerForMaskedLM", "BridgeTowerModel", "BridgeTowerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys _UpperCamelCase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure)
77
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class UpperCAmelCase_ : def __init__( self , a , a , a , a , a , a=0.2 , a=0.2 ) -> Dict: lowercase__ : Any = bp_numa lowercase__ : Optional[int] = bp_numa lowercase__ : Tuple = bp_numa lowercase__ : Optional[Any] = conva_get[:2] lowercase__ : Optional[int] = conva_get[2] lowercase__ : Optional[Any] = size_pa lowercase__ : Union[str, Any] = rate_w lowercase__ : Union[str, Any] = rate_t lowercase__ : List[Any] = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ : Optional[Any] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : int = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : Any = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 def _UpperCAmelCase ( self , a ) -> Union[str, Any]: # save model dict with pickle lowercase__ : Optional[Any] = { 'num_bp1': self.num_bpa, 'num_bp2': self.num_bpa, 'num_bp3': self.num_bpa, 'conv1': self.conva, 'step_conv1': self.step_conva, 'size_pooling1': self.size_poolinga, 'rate_weight': self.rate_weight, 'rate_thre': self.rate_thre, 'w_conv1': self.w_conva, 'wkj': self.wkj, 'vji': self.vji, 'thre_conv1': self.thre_conva, 'thre_bp2': self.thre_bpa, 'thre_bp3': self.thre_bpa, } with open(a , 'wb' ) as f: pickle.dump(a , a ) print(f"""Model saved: {save_path}""" ) @classmethod def _UpperCAmelCase ( cls , a ) -> Any: # read saved model with open(a , 'rb' ) as f: lowercase__ : Optional[int] = pickle.load(a ) # noqa: S301 lowercase__ : Optional[int] = model_dic.get('conv1' ) conv_get.append(model_dic.get('step_conv1' ) ) lowercase__ : List[Any] = model_dic.get('size_pooling1' ) lowercase__ : Tuple = model_dic.get('num_bp1' ) lowercase__ : int = model_dic.get('num_bp2' ) lowercase__ : int = model_dic.get('num_bp3' ) lowercase__ : Union[str, Any] = model_dic.get('rate_weight' ) lowercase__ : Tuple = model_dic.get('rate_thre' ) # create model instance lowercase__ : Tuple = CNN(a , a , a , a , a , a , a ) # modify model parameter lowercase__ : str = model_dic.get('w_conv1' ) lowercase__ : Optional[int] = model_dic.get('wkj' ) lowercase__ : Tuple = model_dic.get('vji' ) lowercase__ : str = model_dic.get('thre_conv1' ) lowercase__ : Union[str, Any] = model_dic.get('thre_bp2' ) lowercase__ : List[str] = model_dic.get('thre_bp3' ) return conv_ins def _UpperCAmelCase ( self , a ) -> str: return 1 / (1 + np.exp(-1 * x )) def _UpperCAmelCase ( self , a ) -> Any: return round(a , 3 ) def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[str]: # convolution process lowercase__ : int = convs[0] lowercase__ : Optional[Any] = convs[1] lowercase__ : int = np.shape(a )[0] # get the data slice of original image data, data_focus lowercase__ : Optional[Any] = [] for i_focus in range(0 , size_data - size_conv + 1 , a ): for j_focus in range(0 , size_data - size_conv + 1 , a ): lowercase__ : Optional[int] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(a ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ : Union[str, Any] = [] lowercase__ : Dict = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(a ): lowercase__ : Any = [] for i_focus in range(len(a ) ): lowercase__ : Tuple = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(a ) ) lowercase__ : Optional[Any] = np.asmatrix(a ).reshape( a , a ) data_featuremap.append(a ) # expanding the data slice to One dimenssion lowercase__ : str = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(a ) ) lowercase__ : int = np.asarray(a ) return focus_list, data_featuremap def _UpperCAmelCase ( self , a , a , a="average_pool" ) -> str: # pooling process lowercase__ : List[str] = len(featuremaps[0] ) lowercase__ : List[str] = int(size_map / size_pooling ) lowercase__ : str = [] for i_map in range(len(a ) ): lowercase__ : List[str] = featuremaps[i_map] lowercase__ : Optional[int] = [] for i_focus in range(0 , a , a ): for j_focus in range(0 , a , a ): lowercase__ : List[Any] = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(a ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(a ) ) lowercase__ : List[Any] = np.asmatrix(a ).reshape(a , a ) featuremap_pooled.append(a ) return featuremap_pooled def _UpperCAmelCase ( self , a ) -> List[str]: # expanding three dimension data to one dimension list lowercase__ : Any = [] for i in range(len(a ) ): lowercase__ : Optional[int] = np.shape(data[i] ) lowercase__ : int = data[i].reshape(1 , shapes[0] * shapes[1] ) lowercase__ : str = data_listed.getA().tolist()[0] data_expanded.extend(a ) lowercase__ : int = np.asarray(a ) return data_expanded def _UpperCAmelCase ( self , a ) -> Dict: # expanding matrix to one dimension list lowercase__ : Dict = np.asarray(a ) lowercase__ : Union[str, Any] = np.shape(a ) lowercase__ : Optional[Any] = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[Any]: lowercase__ : Dict = [] lowercase__ : int = 0 for i_map in range(a ): lowercase__ : str = np.ones((size_map, size_map) ) for i in range(0 , a , a ): for j in range(0 , a , a ): lowercase__ : Optional[Any] = pd_pool[ i_pool ] lowercase__ : Union[str, Any] = i_pool + 1 lowercase__ : List[Any] = np.multiply( a , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(a ) return pd_all def _UpperCAmelCase ( self , a , a , a , a , a , a=bool ) -> str: # model traning print('----------------------Start Training-------------------------' ) print((' - - Shape: Train_Data ', np.shape(a )) ) print((' - - Shape: Teach_Data ', np.shape(a )) ) lowercase__ : int = 0 lowercase__ : List[Any] = [] lowercase__ : Union[str, Any] = 1_0_0_0_0 while rp < n_repeat and mse >= error_accuracy: lowercase__ : List[Any] = 0 print(f"""-------------Learning Time {rp}--------------""" ) for p in range(len(a ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ : Optional[int] = np.asmatrix(datas_train[p] ) lowercase__ : int = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ : Union[str, Any] = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Optional[Any] = self.pooling(a , self.size_poolinga ) lowercase__ : Tuple = np.shape(a ) lowercase__ : List[str] = self._expand(a ) lowercase__ : Optional[int] = data_bp_input lowercase__ : Optional[Any] = np.dot(a , self.vji.T ) - self.thre_bpa lowercase__ : str = self.sig(a ) lowercase__ : Tuple = np.dot(a , self.wkj.T ) - self.thre_bpa lowercase__ : Any = self.sig(a ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ : int = np.multiply( (data_teach - bp_outa) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Any = np.multiply( np.dot(a , self.wkj ) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Optional[int] = np.dot(a , self.vji ) lowercase__ : Union[str, Any] = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ : Any = pd_conva_pooled.T.getA().tolist() lowercase__ : List[str] = self._calculate_gradient_from_pool( a , a , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ : Optional[int] = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ : Tuple = self.rate_weight * np.dot(a , a ) lowercase__ : Union[str, Any] = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ : Any = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ : Tuple = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ : Optional[Any] = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ : Dict = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ : str = rp + 1 lowercase__ : List[str] = error_count / patterns all_mse.append(a ) def draw_error(): lowercase__ : Any = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(a , '+-' ) plt.plot(a , 'r--' ) plt.xlabel('Learning Times' ) plt.ylabel('All_mse' ) plt.grid(a , alpha=0.5 ) plt.show() print('------------------Training Complished---------------------' ) print((' - - Training epoch: ', rp, f""" - - Mse: {mse:.6f}""") ) if draw_e: draw_error() return mse def _UpperCAmelCase ( self , a ) -> List[Any]: # model predict lowercase__ : Optional[int] = [] print('-------------------Start Testing-------------------------' ) print((' - - Shape: Test_Data ', np.shape(a )) ) for p in range(len(a ) ): lowercase__ : List[str] = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ : Tuple = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Any = self.pooling(a , self.size_poolinga ) lowercase__ : Union[str, Any] = self._expand(a ) lowercase__ : Optional[Any] = data_bp_input lowercase__ : str = bp_outa * self.vji.T - self.thre_bpa lowercase__ : Optional[Any] = self.sig(a ) lowercase__ : Dict = bp_outa * self.wkj.T - self.thre_bpa lowercase__ : List[str] = self.sig(a ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ : Optional[int] = [list(map(self.do_round , a ) ) for each in produce_out] return np.asarray(a ) def _UpperCAmelCase ( self , a ) -> List[str]: # return the data of image after convoluting process so we can check it out lowercase__ : Any = np.asmatrix(a ) lowercase__ , lowercase__ : str = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Tuple = self.pooling(a , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
77
1
"""simple docstring""" import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _UpperCamelCase : Optional[Any] = logging.get_logger(__name__) _UpperCamelCase : Union[str, Any] = {"vocab_file": "vocab.txt", "emoji_file": "emoji.json"} _UpperCamelCase : List[str] = { "vocab_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt", }, "emoji_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json", }, } _UpperCamelCase : Union[str, Any] = { "abeja/gpt-neox-japanese-2.7b": 20_48, } def a_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict ): '''simple docstring''' with open(_lowerCAmelCase , 'r' , encoding='utf-8' ) as f: lowercase__ : Tuple = json.loads(f.read() ) lowercase__ : Union[str, Any] = collections.OrderedDict() lowercase__ : List[Any] = collections.OrderedDict() lowercase__ : Dict = collections.OrderedDict() with open(_lowerCAmelCase , 'r' , encoding='utf-8' ) as f: lowercase__ : List[Any] = f.readlines() lowercase__ : Tuple = [[t.rstrip('\n' )] if (t == ',' or ',' not in t) else t.rstrip('\n' ).split(',' ) for t in token] for idx, b in enumerate(_lowerCAmelCase ): lowercase__ : List[str] = b lowercase__ : Any = idx for wd in b: lowercase__ : Union[str, Any] = idx return vocab, raw_vocab, ids_to_tokens, emoji class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = VOCAB_FILES_NAMES lowerCamelCase__ : Any = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Optional[int] = ["input_ids", "attention_mask"] def __init__( self , a , a , a="<|endoftext|>" , a="<|endoftext|>" , a="<|startoftext|>" , a="<|endoftext|>" , a=False , **a , ) -> List[str]: super().__init__( unk_token=a , pad_token=a , bos_token=a , eos_token=a , do_clean_text=a , **a , ) if not os.path.isfile(a ): raise ValueError( f"""Can't find a vocabulary file at path '{vocab_file}'. To load the vocabulary from a Google pretrained""" ' model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) if not os.path.isfile(a ): raise ValueError( f"""Can't find a emoji file at path '{emoji_file}'. To load the emoji information from a Google""" ' pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) lowercase__ : str = do_clean_text lowercase__ , lowercase__ , lowercase__ , lowercase__ : List[Any] = load_vocab_and_emoji(a , a ) lowercase__ : List[Any] = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def _UpperCAmelCase ( self ) -> Optional[int]: # self.vocab contains support for character fluctuation unique to Japanese, and has a large number of vocab return len(self.raw_vocab ) def _UpperCAmelCase ( self ) -> List[str]: return dict(self.raw_vocab , **self.added_tokens_encoder ) def _UpperCAmelCase ( self , a ) -> Union[str, Any]: return self.subword_tokenizer.tokenize(a , clean=self.do_clean_text ) def _UpperCAmelCase ( self , a ) -> List[Any]: return self.vocab.get(a , self.vocab.get(self.unk_token ) ) def _UpperCAmelCase ( self , a ) -> Tuple: return self.subword_tokenizer.convert_id_to_token(a ) def _UpperCAmelCase ( self , a ) -> Tuple: lowercase__ : Any = ''.join(a ).strip() return out_string def _UpperCAmelCase ( self , a ) -> List[int]: lowercase__ : List[Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(a , add_special_tokens=a ) + [self.eos_token_id] ) if len(a ) > self.model_max_length: lowercase__ : str = input_ids[-self.model_max_length :] return input_ids def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: lowercase__ : Union[str, Any] = 0 if os.path.isdir(a ): lowercase__ : Optional[Any] = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : Dict = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['emoji_file'] ) else: lowercase__ : int = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : List[str] = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['emoji_file'] ) with open(a , 'w' , encoding='utf-8' ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" ' Please check that the vocabulary is not corrupted!' ) lowercase__ : List[str] = token_index writer.write(','.join(a ) + '\n' ) index += 1 with open(a , 'w' , encoding='utf-8' ) as writer: json.dump(self.emoji , a ) return vocab_file, emoji_file class UpperCAmelCase_ ( _a): def __init__( self , a , a , a ) -> List[Any]: lowercase__ : Tuple = vocab # same as swe lowercase__ : Optional[int] = ids_to_tokens # same as bpe lowercase__ : Tuple = emoji lowercase__ : Dict = np.max([len(a ) for w in self.vocab.keys()] ) lowercase__ : List[Any] = re.compile(R'(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)' ) lowercase__ : List[str] = re.compile(R'[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*' ) lowercase__ : int = re.compile(R'[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}' ) lowercase__ : Any = re.compile( R'([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) lowercase__ : Tuple = re.compile( R'(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) lowercase__ : Any = re.compile( R'((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*' ) lowercase__ : Dict = '─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿' lowercase__ : Union[str, Any] = '▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟' lowercase__ : Dict = str.maketrans({k: '<BLOCK>' for k in keisen + blocks} ) def __len__( self ) -> Optional[int]: return len(self.ids_to_tokens ) def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Optional[int] = self.content_repattera.sub('<URL>' , a ) lowercase__ : str = self.content_repattera.sub('<EMAIL>' , a ) lowercase__ : Tuple = self.content_repattera.sub('<TEL>' , a ) lowercase__ : int = self.content_repattera.sub('<DATE>' , a ) lowercase__ : Optional[int] = self.content_repattera.sub('<DATE>' , a ) lowercase__ : List[str] = self.content_repattera.sub('<PRICE>' , a ) lowercase__ : List[str] = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: lowercase__ : Any = content.replace('<BLOCK><BLOCK>' , '<BLOCK>' ) return content def _UpperCAmelCase ( self , a , a=False ) -> Any: lowercase__ : Dict = text.replace(' ' , '<SP>' ) lowercase__ : Tuple = text.replace(' ' , '<SP>' ) lowercase__ : List[Any] = text.replace('\r\n' , '<BR>' ) lowercase__ : str = text.replace('\n' , '<BR>' ) lowercase__ : Any = text.replace('\r' , '<BR>' ) lowercase__ : int = text.replace('\t' , '<TAB>' ) lowercase__ : Tuple = text.replace('—' , 'ー' ) lowercase__ : Dict = text.replace('−' , 'ー' ) for k, v in self.emoji["emoji"].items(): if k in text: lowercase__ : Any = text.replace(a , a ) if clean: lowercase__ : int = self.clean_text(a ) def check_simbol(a ): lowercase__ : Dict = x.encode() if len(a ) == 1 and len(a ) == 2: lowercase__ : int = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0XC_2A1 and c <= 0XC_2BF) or (c >= 0XC_780 and c <= 0XC_783) or (c >= 0XC_AB9 and c <= 0XC_BBF) or (c >= 0XC_C80 and c <= 0XC_DA2) ): return True return False def checkuae(a ): lowercase__ : Any = x.encode() if len(a ) == 1 and len(a ) == 3: lowercase__ : Optional[Any] = (int(e[0] ) << 1_6) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0XE28_080 and c <= 0XE2B_07F: return True return False lowercase__ : List[Any] = 0 lowercase__ : Tuple = [] while pos < len(a ): lowercase__ : List[Any] = min(len(a ) , pos + self.maxlen + 1 ) if text[pos] == '<' else pos + 3 lowercase__ : str = [] # (token_id, token, pos) for e in range(a , a , -1 ): lowercase__ : Tuple = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(a ) > 2: lowercase__ : Optional[int] = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(a ) > 0: # the smallest token_id is adopted lowercase__ , lowercase__ , lowercase__ : Union[str, Any] = sorted(a , key=lambda a : x[0] )[0] result.append(a ) lowercase__ : str = e else: lowercase__ : List[Any] = pos + 1 lowercase__ : Tuple = text[pos:end] if check_simbol(a ): result.append('<KIGOU>' ) elif checkuae(a ): result.append('<U2000U2BFF>' ) else: for i in wd.encode('utf-8' ): result.append('<|byte%d|>' % i ) lowercase__ : Optional[Any] = end return result def _UpperCAmelCase ( self , a , a="\n" ) -> Tuple: lowercase__ : List[Any] = [] lowercase__ : str = [] lowercase__ : Optional[Any] = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(a ) > 0: words.append(bytearray(a ).decode('utf-8' , errors='replace' ) ) lowercase__ : Any = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji['emoji_inv'][word] ) elif word == "<SP>": words.append(' ' ) elif word == "<BR>": words.append(a ) elif word == "<TAB>": words.append('\t' ) elif word == "<BLOCK>": words.append('▀' ) elif word == "<KIGOU>": words.append('ǀ' ) elif word == "<U2000U2BFF>": words.append('‖' ) else: words.append(a ) if len(a ) > 0: words.append(bytearray(a ).decode('utf-8' , errors='replace' ) ) lowercase__ : Optional[Any] = ''.join(a ) return text
77
"""simple docstring""" from collections.abc import Generator def a_ ( ): '''simple docstring''' lowercase__ , lowercase__ : List[str] = 0, 1 while True: lowercase__ , lowercase__ : Optional[int] = b, a + b yield b def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' lowercase__ : List[Any] = 1 lowercase__ : Any = fibonacci_generator() while len(str(next(_lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
77
1
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np _UpperCamelCase : Any = re.compile(r"\b(a|an|the)\b", re.UNICODE) _UpperCamelCase : Union[str, Any] = None def a_ ( ): '''simple docstring''' lowercase__ : Optional[int] = argparse.ArgumentParser('Official evaluation script for SQuAD version 2.0.' ) parser.add_argument('data_file' , metavar='data.json' , help='Input data JSON file.' ) parser.add_argument('pred_file' , metavar='pred.json' , help='Model predictions.' ) parser.add_argument( '--out-file' , '-o' , metavar='eval.json' , help='Write accuracy metrics to file (default is stdout).' ) parser.add_argument( '--na-prob-file' , '-n' , metavar='na_prob.json' , help='Model estimates of probability of no answer.' ) parser.add_argument( '--na-prob-thresh' , '-t' , type=_lowerCAmelCase , default=1.0 , help='Predict "" if no-answer probability exceeds this (default = 1.0).' , ) parser.add_argument( '--out-image-dir' , '-p' , metavar='out_images' , default=_lowerCAmelCase , help='Save precision-recall curves to directory.' ) parser.add_argument('--verbose' , '-v' , action='store_true' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Tuple = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Optional[int] = bool(qa['answers']['text'] ) return qid_to_has_ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' def remove_articles(_lowerCAmelCase : int ): return ARTICLES_REGEX.sub(' ' , _lowerCAmelCase ) def white_space_fix(_lowerCAmelCase : str ): return " ".join(text.split() ) def remove_punc(_lowerCAmelCase : List[Any] ): lowercase__ : int = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_lowerCAmelCase : List[str] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_lowerCAmelCase ) ) ) ) def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' if not s: return [] return normalize_answer(_lowerCAmelCase ).split() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ): '''simple docstring''' return int(normalize_answer(_lowerCAmelCase ) == normalize_answer(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ): '''simple docstring''' lowercase__ : Dict = get_tokens(_lowerCAmelCase ) lowercase__ : List[str] = get_tokens(_lowerCAmelCase ) lowercase__ : List[Any] = collections.Counter(_lowerCAmelCase ) & collections.Counter(_lowerCAmelCase ) lowercase__ : int = sum(common.values() ) if len(_lowerCAmelCase ) == 0 or len(_lowerCAmelCase ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 lowercase__ : Any = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Dict = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Any = (2 * precision * recall) / (precision + recall) return fa def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Optional[int] = {} lowercase__ : Union[str, Any] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Any = qa['id'] lowercase__ : Union[str, Any] = [t for t in qa['answers']['text'] if normalize_answer(_lowerCAmelCase )] if not gold_answers: # For unanswerable questions, only correct answer is empty string lowercase__ : Dict = [''] if qid not in preds: print(f"""Missing prediction for {qid}""" ) continue lowercase__ : Optional[int] = preds[qid] # Take max over all gold answers lowercase__ : int = max(compute_exact(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) lowercase__ : Optional[Any] = max(compute_fa(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) return exact_scores, fa_scores def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : str = {} for qid, s in scores.items(): lowercase__ : int = na_probs[qid] > na_prob_thresh if pred_na: lowercase__ : Optional[Any] = float(not qid_to_has_ans[qid] ) else: lowercase__ : Optional[Any] = s return new_scores def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str=None ): '''simple docstring''' if not qid_list: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores.values() ) / total), ('f1', 1_0_0.0 * sum(fa_scores.values() ) / total), ('total', total), ] ) else: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores[k] for k in qid_list ) / total), ('f1', 1_0_0.0 * sum(fa_scores[k] for k in qid_list ) / total), ('total', total), ] ) def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' for k in new_eval: lowercase__ : int = new_eval[k] def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple ): '''simple docstring''' plt.step(_lowerCAmelCase , _lowerCAmelCase , color='b' , alpha=0.2 , where='post' ) plt.fill_between(_lowerCAmelCase , _lowerCAmelCase , step='post' , alpha=0.2 , color='b' ) plt.xlabel('Recall' ) plt.ylabel('Precision' ) plt.xlim([0.0, 1.0_5] ) plt.ylim([0.0, 1.0_5] ) plt.title(_lowerCAmelCase ) plt.savefig(_lowerCAmelCase ) plt.clf() def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : str , _lowerCAmelCase : Any=None , _lowerCAmelCase : List[str]=None ): '''simple docstring''' lowercase__ : Optional[int] = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) lowercase__ : Tuple = 0.0 lowercase__ : List[str] = 1.0 lowercase__ : List[str] = 0.0 lowercase__ : Union[str, Any] = [1.0] lowercase__ : List[Any] = [0.0] lowercase__ : Optional[int] = 0.0 for i, qid in enumerate(_lowerCAmelCase ): if qid_to_has_ans[qid]: true_pos += scores[qid] lowercase__ : Tuple = true_pos / float(i + 1 ) lowercase__ : Union[str, Any] = true_pos / float(_lowerCAmelCase ) if i == len(_lowerCAmelCase ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_lowerCAmelCase ) recalls.append(_lowerCAmelCase ) if out_image: plot_pr_curve(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return {"ap": 1_0_0.0 * avg_prec} def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple ): '''simple docstring''' if out_image_dir and not os.path.exists(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) lowercase__ : List[str] = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return lowercase__ : Dict = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_exact.png' ) , title='Precision-Recall curve for Exact Match score' , ) lowercase__ : Tuple = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_f1.png' ) , title='Precision-Recall curve for F1 score' , ) lowercase__ : List[Any] = {k: float(_lowerCAmelCase ) for k, v in qid_to_has_ans.items()} lowercase__ : Any = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_oracle.png' ) , title='Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)' , ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_exact' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_f1' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_oracle' ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' if not qid_list: return lowercase__ : List[str] = [na_probs[k] for k in qid_list] lowercase__ : Tuple = np.ones_like(_lowerCAmelCase ) / float(len(_lowerCAmelCase ) ) plt.hist(_lowerCAmelCase , weights=_lowerCAmelCase , bins=20 , range=(0.0, 1.0) ) plt.xlabel('Model probability of no-answer' ) plt.ylabel('Proportion of dataset' ) plt.title(f"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(_lowerCAmelCase , f"""na_prob_hist_{name}.png""" ) ) plt.clf() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : Tuple = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) lowercase__ : int = num_no_ans lowercase__ : Optional[int] = cur_score lowercase__ : Tuple = 0.0 lowercase__ : Dict = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) for i, qid in enumerate(_lowerCAmelCase ): if qid not in scores: continue if qid_to_has_ans[qid]: lowercase__ : Optional[int] = scores[qid] else: if preds[qid]: lowercase__ : List[Any] = -1 else: lowercase__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: lowercase__ : Dict = cur_score lowercase__ : Optional[int] = na_probs[qid] return 1_0_0.0 * best_score / len(_lowerCAmelCase ), best_thresh def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str ): '''simple docstring''' lowercase__ , lowercase__ : List[Any] = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ , lowercase__ : Dict = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Any = best_exact lowercase__ : Tuple = exact_thresh lowercase__ : Optional[Any] = best_fa lowercase__ : Any = fa_thresh def a_ ( ): '''simple docstring''' with open(OPTS.data_file ) as f: lowercase__ : List[Any] = json.load(_lowerCAmelCase ) lowercase__ : Union[str, Any] = dataset_json['data'] with open(OPTS.pred_file ) as f: lowercase__ : str = json.load(_lowerCAmelCase ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: lowercase__ : Union[str, Any] = json.load(_lowerCAmelCase ) else: lowercase__ : str = {k: 0.0 for k in preds} lowercase__ : int = make_qid_to_has_ans(_lowerCAmelCase ) # maps qid to True/False lowercase__ : List[str] = [k for k, v in qid_to_has_ans.items() if v] lowercase__ : Any = [k for k, v in qid_to_has_ans.items() if not v] lowercase__ , lowercase__ : Any = get_raw_scores(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Optional[Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Union[str, Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Tuple = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase ) if has_ans_qids: lowercase__ : int = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'HasAns' ) if no_ans_qids: lowercase__ : Optional[Any] = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'NoAns' ) if OPTS.na_prob_file: find_all_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'hasAns' ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'noAns' ) if OPTS.out_file: with open(OPTS.out_file , 'w' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) else: print(json.dumps(_lowerCAmelCase , indent=2 ) ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
77
"""simple docstring""" import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class UpperCAmelCase_ : def __init__( self , a ) -> List[str]: if isinstance(a , a ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden lowercase__ : Optional[Any] = deepcopy(a ) elif os.path.exists(a ): with io.open(a , 'r' , encoding='utf-8' ) as f: lowercase__ : List[Any] = json.load(a ) else: try: lowercase__ : Optional[int] = baseaa.urlsafe_baadecode(a ).decode('utf-8' ) lowercase__ : List[str] = json.loads(a ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""" ) lowercase__ : Any = config self.set_stage_and_offload() def _UpperCAmelCase ( self ) -> Dict: # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. lowercase__ : Tuple = self.get_value('zero_optimization.stage' , -1 ) # offload lowercase__ : int = False if self.is_zeroa() or self.is_zeroa(): lowercase__ : str = set(['cpu', 'nvme'] ) lowercase__ : Optional[Any] = set( [ self.get_value('zero_optimization.offload_optimizer.device' ), self.get_value('zero_optimization.offload_param.device' ), ] ) if len(offload_devices & offload_devices_valid ) > 0: lowercase__ : Optional[Any] = True def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Dict = self.config # find the config node of interest if it exists lowercase__ : int = ds_key_long.split('.' ) lowercase__ : Dict = nodes.pop() for node in nodes: lowercase__ : Optional[Any] = config.get(a ) if config is None: return None, ds_key return config, ds_key def _UpperCAmelCase ( self , a , a=None ) -> Union[str, Any]: lowercase__ , lowercase__ : Tuple = self.find_config_node(a ) if config is None: return default return config.get(a , a ) def _UpperCAmelCase ( self , a , a=False ) -> Any: lowercase__ : str = self.config # find the config node of interest if it exists lowercase__ : List[Any] = ds_key_long.split('.' ) for node in nodes: lowercase__ : str = config lowercase__ : str = config.get(a ) if config is None: if must_exist: raise ValueError(f"""Can't find {ds_key_long} entry in the config: {self.config}""" ) else: return # if found remove it if parent_config is not None: parent_config.pop(a ) def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : Union[str, Any] = self.get_value(a ) return False if value is None else bool(a ) def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Any = self.get_value(a ) return False if value is None else not bool(a ) def _UpperCAmelCase ( self ) -> Tuple: return self._stage == 2 def _UpperCAmelCase ( self ) -> List[Any]: return self._stage == 3 def _UpperCAmelCase ( self ) -> str: return self._offload class UpperCAmelCase_ : def __init__( self , a ) -> str: lowercase__ : Tuple = engine def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: # runs backpropagation and handles mixed precision self.engine.backward(a , **a ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class UpperCAmelCase_ ( _a): def __init__( self , a ) -> Dict: super().__init__(a , device_placement=a , scaler=a ) lowercase__ : Union[str, Any] = hasattr(self.optimizer , 'overflow' ) def _UpperCAmelCase ( self , a=None ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def _UpperCAmelCase ( self ) -> Optional[int]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def _UpperCAmelCase ( self ) -> Tuple: if self.__has_overflow__: return self.optimizer.overflow return False class UpperCAmelCase_ ( _a): def __init__( self , a , a ) -> Any: super().__init__(a , a ) def _UpperCAmelCase ( self ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class UpperCAmelCase_ : def __init__( self , a , a=0.001 , a=0 , **a ) -> Tuple: lowercase__ : List[Any] = params lowercase__ : int = lr lowercase__ : int = weight_decay lowercase__ : Union[str, Any] = kwargs class UpperCAmelCase_ : def __init__( self , a , a=None , a=0 , **a ) -> Tuple: lowercase__ : Dict = optimizer lowercase__ : List[str] = total_num_steps lowercase__ : Optional[int] = warmup_num_steps lowercase__ : List[Any] = kwargs
77
1
"""simple docstring""" import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _UpperCamelCase : Any = logging.getLogger(__name__) _UpperCamelCase : Optional[Any] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _UpperCamelCase : Any = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The model checkpoint for weights initialization. Leave None if you want to train a model from" " scratch." ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(_a)} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained config name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "The input training data file (a text file)."}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The input training data files (multiple files in glob format). " "Very often splitting large files to smaller files can prevent tokenizer going out of memory" ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input train ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input eval ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Whether distinct lines of text in the dataset are to be handled as distinct sequences."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Train with masked-language modeling loss instead of language modeling."}) lowerCamelCase__ : bool = field(default=_a , metadata={"help": "Whether ot not to use whole word mask."}) lowerCamelCase__ : float = field( default=0.15 , metadata={"help": "Ratio of tokens to mask for masked language modeling loss"}) lowerCamelCase__ : float = field( default=1 / 6 , metadata={ "help": ( "Ratio of length of a span of masked tokens to surrounding context length for permutation language" " modeling." ) } , ) lowerCamelCase__ : int = field( default=5 , metadata={"help": "Maximum length of a span of masked tokens for permutation language modeling."}) lowerCamelCase__ : int = field( default=-1 , metadata={ "help": ( "Optional input sequence length after tokenization." "The training dataset will be truncated in block of this size for training." "Default to the model max input length for single sentence inputs (take into account special tokens)." ) } , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Overwrite the cached training and evaluation sets"}) def a_ ( _lowerCAmelCase : DataTrainingArguments , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[str] = None , ): '''simple docstring''' def _dataset(_lowerCAmelCase : Any , _lowerCAmelCase : Any=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('You need to set world whole masking and mlm to True for Chinese Whole Word Mask' ) return LineByLineWithRefDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , ref_path=_lowerCAmelCase , ) return LineByLineTextDataset(tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size ) else: return TextDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=_lowerCAmelCase , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(_lowerCAmelCase ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def a_ ( ): '''simple docstring''' lowercase__ : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ : List[Any] = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( 'Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ' 'or remove the --do_eval argument.' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s' , _lowerCAmelCase ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: lowercase__ : List[Any] = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : Any = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: lowercase__ : Union[str, Any] = CONFIG_MAPPING[model_args.model_type]() logger.warning('You are instantiating a new config instance from scratch.' ) if model_args.tokenizer_name: lowercase__ : Dict = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : int = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another' ' script, save it,and load it from here, using --tokenizer_name' ) if model_args.model_name_or_path: lowercase__ : Optional[Any] = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=_lowerCAmelCase , cache_dir=model_args.cache_dir , ) else: logger.info('Training new model from scratch' ) lowercase__ : int = AutoModelWithLMHead.from_config(_lowerCAmelCase ) model.resize_token_embeddings(len(_lowerCAmelCase ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( 'BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the' '--mlm flag (masked language modeling).' ) if data_args.block_size <= 0: lowercase__ : List[str] = tokenizer.max_len # Our input block size will be the max possible for the model else: lowercase__ : int = min(data_args.block_size , tokenizer.max_len ) # Get datasets lowercase__ : Tuple = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) lowercase__ : Optional[Any] = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , evaluate=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": lowercase__ : List[Any] = DataCollatorForPermutationLanguageModeling( tokenizer=_lowerCAmelCase , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: lowercase__ : List[str] = DataCollatorForWholeWordMask( tokenizer=_lowerCAmelCase , mlm_probability=data_args.mlm_probability ) else: lowercase__ : str = DataCollatorForLanguageModeling( tokenizer=_lowerCAmelCase , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer lowercase__ : Optional[int] = Trainer( model=_lowerCAmelCase , args=_lowerCAmelCase , data_collator=_lowerCAmelCase , train_dataset=_lowerCAmelCase , eval_dataset=_lowerCAmelCase , prediction_loss_only=_lowerCAmelCase , ) # Training if training_args.do_train: lowercase__ : List[str] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=_lowerCAmelCase ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase__ : List[Any] = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) lowercase__ : Dict = trainer.evaluate() lowercase__ : List[Any] = math.exp(eval_output['eval_loss'] ) lowercase__ : int = {'perplexity': perplexity} lowercase__ : int = os.path.join(training_args.output_dir , 'eval_results_lm.txt' ) if trainer.is_world_master(): with open(_lowerCAmelCase , 'w' ) as writer: logger.info('***** Eval results *****' ) for key in sorted(result.keys() ): logger.info(' %s = %s' , _lowerCAmelCase , str(result[key] ) ) writer.write('%s = %s\n' % (key, str(result[key] )) ) results.update(_lowerCAmelCase ) return results def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' main() if __name__ == "__main__": main()
77
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _UpperCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , *a , **a ) -> Union[str, Any]: super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _UpperCAmelCase ( self , a=None ) -> Dict: lowercase__ : Any = {} if top_k is not None: lowercase__ : List[str] = top_k return {}, {}, postprocess_params def __call__( self , a , **a ) -> Tuple: return super().__call__(a , **a ) def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : List[Any] = load_image(a ) lowercase__ : Union[str, Any] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _UpperCAmelCase ( self , a ) -> List[str]: lowercase__ : Dict = self.model(**a ) return model_outputs def _UpperCAmelCase ( self , a , a=5 ) -> Dict: if top_k > self.model.config.num_labels: lowercase__ : List[Any] = self.model.config.num_labels if self.framework == "pt": lowercase__ : Tuple = model_outputs.logits.softmax(-1 )[0] lowercase__ , lowercase__ : Optional[Any] = probs.topk(a ) elif self.framework == "tf": lowercase__ : Union[str, Any] = stable_softmax(model_outputs.logits , axis=-1 )[0] lowercase__ : str = tf.math.top_k(a , k=a ) lowercase__ , lowercase__ : Dict = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) lowercase__ : Dict = scores.tolist() lowercase__ : Dict = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
77
1
"""simple docstring""" class UpperCAmelCase_ : def __init__( self ) -> List[str]: lowercase__ : Optional[int] = 0 lowercase__ : int = 0 lowercase__ : List[Any] = {} def _UpperCAmelCase ( self , a ) -> int: if vertex not in self.adjacency: lowercase__ : Any = {} self.num_vertices += 1 def _UpperCAmelCase ( self , a , a , a ) -> List[str]: self.add_vertex(a ) self.add_vertex(a ) if head == tail: return lowercase__ : Tuple = weight lowercase__ : List[Any] = weight def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Optional[int] = self.get_edges() for edge in edges: lowercase__ , lowercase__ , lowercase__ : str = edge edges.remove((tail, head, weight) ) for i in range(len(a ) ): lowercase__ : Optional[Any] = list(edges[i] ) edges.sort(key=lambda a : e[2] ) for i in range(len(a ) - 1 ): if edges[i][2] >= edges[i + 1][2]: lowercase__ : Union[str, Any] = edges[i][2] + 1 for edge in edges: lowercase__ , lowercase__ , lowercase__ : Union[str, Any] = edge lowercase__ : Dict = weight lowercase__ : Optional[int] = weight def __str__( self ) -> Dict: lowercase__ : Any = '' for tail in self.adjacency: for head in self.adjacency[tail]: lowercase__ : List[Any] = self.adjacency[head][tail] string += f"""{head} -> {tail} == {weight}\n""" return string.rstrip('\n' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Tuple = [] for tail in self.adjacency: for head in self.adjacency[tail]: output.append((tail, head, self.adjacency[head][tail]) ) return output def _UpperCAmelCase ( self ) -> Optional[int]: return self.adjacency.keys() @staticmethod def _UpperCAmelCase ( a=None , a=None ) -> Dict: lowercase__ : Dict = Graph() if vertices is None: lowercase__ : int = [] if edges is None: lowercase__ : int = [] for vertex in vertices: g.add_vertex(a ) for edge in edges: g.add_edge(*a ) return g class UpperCAmelCase_ : def __init__( self ) -> List[Any]: lowercase__ : Dict = {} lowercase__ : Optional[Any] = {} def __len__( self ) -> Union[str, Any]: return len(self.parent ) def _UpperCAmelCase ( self , a ) -> List[Any]: if item in self.parent: return self.find(a ) lowercase__ : Tuple = item lowercase__ : List[Any] = 0 return item def _UpperCAmelCase ( self , a ) -> Optional[Any]: if item not in self.parent: return self.make_set(a ) if item != self.parent[item]: lowercase__ : Union[str, Any] = self.find(self.parent[item] ) return self.parent[item] def _UpperCAmelCase ( self , a , a ) -> List[str]: lowercase__ : str = self.find(a ) lowercase__ : List[Any] = self.find(a ) if roota == roota: return roota if self.rank[roota] > self.rank[roota]: lowercase__ : List[Any] = roota return roota if self.rank[roota] < self.rank[roota]: lowercase__ : Optional[Any] = roota return roota if self.rank[roota] == self.rank[roota]: self.rank[roota] += 1 lowercase__ : Optional[int] = roota return roota return None @staticmethod def _UpperCAmelCase ( a ) -> List[Any]: lowercase__ : List[Any] = graph.num_vertices lowercase__ : Dict = Graph.UnionFind() lowercase__ : Optional[Any] = [] while num_components > 1: lowercase__ : Any = {} for vertex in graph.get_vertices(): lowercase__ : str = -1 lowercase__ : Dict = graph.get_edges() for edge in edges: lowercase__ , lowercase__ , lowercase__ : Optional[Any] = edge edges.remove((tail, head, weight) ) for edge in edges: lowercase__ , lowercase__ , lowercase__ : Optional[int] = edge lowercase__ : List[str] = union_find.find(a ) lowercase__ : List[Any] = union_find.find(a ) if seta != seta: if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: lowercase__ : List[Any] = [head, tail, weight] if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: lowercase__ : Optional[Any] = [head, tail, weight] for vertex in cheap_edge: if cheap_edge[vertex] != -1: lowercase__ , lowercase__ , lowercase__ : Tuple = cheap_edge[vertex] if union_find.find(a ) != union_find.find(a ): union_find.union(a , a ) mst_edges.append(cheap_edge[vertex] ) lowercase__ : List[Any] = num_components - 1 lowercase__ : Optional[int] = Graph.build(edges=a ) return mst
77
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : str = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all LED models at https://huggingface.co/models?filter=LED _UpperCamelCase : Optional[Any] = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } _UpperCamelCase : Optional[int] = { "allenai/led-base-16384": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def a_ ( ): '''simple docstring''' lowercase__ : int = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowercase__ : Union[str, Any] = bs[:] lowercase__ : str = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCAmelCase ) cs.append(2**8 + n ) n += 1 lowercase__ : str = [chr(_lowerCAmelCase ) for n in cs] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Dict = set() lowercase__ : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Optional[Any] = char return pairs class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = VOCAB_FILES_NAMES lowerCamelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , a , a , a="replace" , a="<s>" , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a=False , **a , ) -> Any: lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else bos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else eos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else sep_token lowercase__ : Dict = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else cls_token lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else unk_token lowercase__ : Tuple = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( errors=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , add_prefix_space=a , **a , ) with open(a , encoding='utf-8' ) as vocab_handle: lowercase__ : Tuple = json.load(a ) lowercase__ : Dict = {v: k for k, v in self.encoder.items()} lowercase__ : str = errors # how to handle errors in decoding lowercase__ : Optional[Any] = bytes_to_unicode() lowercase__ : Optional[Any] = {v: k for k, v in self.byte_encoder.items()} with open(a , encoding='utf-8' ) as merges_handle: lowercase__ : Optional[Any] = merges_handle.read().split('\n' )[1:-1] lowercase__ : Optional[int] = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ : Union[str, Any] = dict(zip(a , range(len(a ) ) ) ) lowercase__ : Tuple = {} lowercase__ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ : List[Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _UpperCAmelCase ( self ) -> List[Any]: return len(self.encoder ) def _UpperCAmelCase ( self ) -> str: return dict(self.encoder , **self.added_tokens_encoder ) def _UpperCAmelCase ( self , a ) -> List[str]: if token in self.cache: return self.cache[token] lowercase__ : Optional[Any] = tuple(a ) lowercase__ : int = get_pairs(a ) if not pairs: return token while True: lowercase__ : List[str] = min(a , key=lambda a : self.bpe_ranks.get(a , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ : List[str] = bigram lowercase__ : Union[str, Any] = [] lowercase__ : List[Any] = 0 while i < len(a ): try: lowercase__ : str = word.index(a , a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : Optional[int] = j if word[i] == first and i < len(a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : int = tuple(a ) lowercase__ : Dict = new_word if len(a ) == 1: break else: lowercase__ : Any = get_pairs(a ) lowercase__ : List[str] = ' '.join(a ) lowercase__ : Optional[Any] = word return word def _UpperCAmelCase ( self , a ) -> Union[str, Any]: lowercase__ : Tuple = [] for token in re.findall(self.pat , a ): lowercase__ : Union[str, Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(a ).split(' ' ) ) return bpe_tokens def _UpperCAmelCase ( self , a ) -> Optional[Any]: return self.encoder.get(a , self.encoder.get(self.unk_token ) ) def _UpperCAmelCase ( self , a ) -> Optional[int]: return self.decoder.get(a ) def _UpperCAmelCase ( self , a ) -> str: lowercase__ : Any = ''.join(a ) lowercase__ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Any = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : str = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(a , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a , ensure_ascii=a ) + '\n' ) lowercase__ : List[Any] = 0 with open(a , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) lowercase__ : Union[str, Any] = token_index writer.write(' '.join(a ) + '\n' ) index += 1 return vocab_file, merge_file def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ : Union[str, Any] = [self.cls_token_id] lowercase__ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) + [1] def _UpperCAmelCase ( self , a , a = None ) -> List[int]: lowercase__ : Dict = [self.sep_token_id] lowercase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCAmelCase ( self , a , a=False , **a ) -> Optional[int]: lowercase__ : Tuple = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(a ) > 0 and not text[0].isspace()): lowercase__ : List[str] = ' ' + text return (text, kwargs) def _UpperCAmelCase ( self , a , a = None , a = PaddingStrategy.DO_NOT_PAD , a = None , a = None , ) -> dict: lowercase__ : Dict = super()._pad( encoded_inputs=a , max_length=a , padding_strategy=a , pad_to_multiple_of=a , return_attention_mask=a , ) # Load from model defaults if return_attention_mask is None: lowercase__ : Union[str, Any] = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ : Any = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ : Tuple = len(encoded_inputs['global_attention_mask'] ) != len(a ) if needs_to_be_padded: lowercase__ : str = len(a ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ : Union[str, Any] = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowercase__ : List[str] = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
77
1
"""simple docstring""" import math import tensorflow as tf from packaging import version def a_ ( _lowerCAmelCase : Dict ): '''simple docstring''' lowercase__ : int = tf.convert_to_tensor(_lowerCAmelCase ) lowercase__ : List[str] = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def a_ ( _lowerCAmelCase : Dict ): '''simple docstring''' lowercase__ : int = tf.convert_to_tensor(_lowerCAmelCase ) lowercase__ : str = tf.cast(math.pi , x.dtype ) lowercase__ : str = tf.cast(0.0_4_4_7_1_5 , x.dtype ) lowercase__ : Union[str, Any] = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(_lowerCAmelCase , 3 )) )) return x * cdf def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' lowercase__ : int = tf.convert_to_tensor(_lowerCAmelCase ) return x * tf.tanh(tf.math.softplus(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : Optional[Any] = tf.convert_to_tensor(_lowerCAmelCase ) lowercase__ : List[Any] = tf.cast(0.0_4_4_7_1_5 , x.dtype ) lowercase__ : str = tf.cast(0.7_9_7_8_8_4_5_6_0_8 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def a_ ( _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : Optional[int] = tf.convert_to_tensor(_lowerCAmelCase ) lowercase__ : Any = tf.cast(1.7_0_2 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' return tf.clip_by_value(_gelu(_lowerCAmelCase ) , -10 , 10 ) def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int]=-1 ): '''simple docstring''' lowercase__ , lowercase__ : Union[str, Any] = tf.split(_lowerCAmelCase , 2 , axis=_lowerCAmelCase ) return a * tf.math.sigmoid(_lowerCAmelCase ) if version.parse(tf.version.VERSION) >= version.parse("2.4"): def a_ ( _lowerCAmelCase : List[str] ): '''simple docstring''' return tf.keras.activations.gelu(_lowerCAmelCase , approximate=_lowerCAmelCase ) _UpperCamelCase : Union[str, Any] = tf.keras.activations.gelu _UpperCamelCase : int = approximate_gelu_wrap else: _UpperCamelCase : Union[str, Any] = _gelu _UpperCamelCase : Optional[Any] = _gelu_new _UpperCamelCase : Union[str, Any] = { "gelu": gelu, "gelu_10": gelu_aa, "gelu_fast": gelu_fast, "gelu_new": gelu_new, "glu": glu, "mish": mish, "quick_gelu": quick_gelu, "relu": tf.keras.activations.relu, "sigmoid": tf.keras.activations.sigmoid, "silu": tf.keras.activations.swish, "swish": tf.keras.activations.swish, "tanh": tf.keras.activations.tanh, } def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(f"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
77
"""simple docstring""" _UpperCamelCase : Union[str, Any] = 8.3_1_4_4_5_9_8 def a_ ( _lowerCAmelCase : float , _lowerCAmelCase : float ): '''simple docstring''' if temperature < 0: raise Exception('Temperature cannot be less than 0 K' ) if molar_mass <= 0: raise Exception('Molar mass cannot be less than or equal to 0 kg/mol' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _UpperCamelCase : List[Any] = 3_00 _UpperCamelCase : Tuple = 28 _UpperCamelCase : Any = rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
77
1
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCamelCase : int = logging.get_logger(__name__) _UpperCamelCase : Union[str, Any] = { "Helsinki-NLP/opus-mt-en-de": "https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json", # See all Marian models at https://huggingface.co/models?filter=marian } class UpperCAmelCase_ ( _a): lowerCamelCase__ : List[Any] = "marian" lowerCamelCase__ : int = ["past_key_values"] lowerCamelCase__ : List[Any] = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self , a=5_8_1_0_1 , a=None , a=1_0_2_4 , a=1_2 , a=4_0_9_6 , a=1_6 , a=1_2 , a=4_0_9_6 , a=1_6 , a=0.0 , a=0.0 , a=True , a=True , a="gelu" , a=1_0_2_4 , a=0.1 , a=0.0 , a=0.0 , a=0.02 , a=5_8_1_0_0 , a=False , a=5_8_1_0_0 , a=0 , a=0 , a=True , **a , ) -> Any: lowercase__ : str = vocab_size lowercase__ : str = decoder_vocab_size or vocab_size lowercase__ : Optional[int] = max_position_embeddings lowercase__ : List[str] = d_model lowercase__ : str = encoder_ffn_dim lowercase__ : Optional[int] = encoder_layers lowercase__ : int = encoder_attention_heads lowercase__ : Tuple = decoder_ffn_dim lowercase__ : int = decoder_layers lowercase__ : Optional[Any] = decoder_attention_heads lowercase__ : str = dropout lowercase__ : Union[str, Any] = attention_dropout lowercase__ : Tuple = activation_dropout lowercase__ : Optional[int] = activation_function lowercase__ : int = init_std lowercase__ : str = encoder_layerdrop lowercase__ : int = decoder_layerdrop lowercase__ : Union[str, Any] = use_cache lowercase__ : int = encoder_layers lowercase__ : str = scale_embedding # scale factor will be sqrt(d_model) if True lowercase__ : Tuple = share_encoder_decoder_embeddings super().__init__( pad_token_id=a , eos_token_id=a , is_encoder_decoder=a , decoder_start_token_id=a , forced_eos_token_id=a , **a , ) class UpperCAmelCase_ ( _a): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowercase__ : Tuple = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: lowercase__ : List[Any] = {0: 'batch'} lowercase__ : int = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: lowercase__ : Dict = {0: 'batch', 1: 'decoder_sequence'} lowercase__ : Tuple = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(a , direction='inputs' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase__ : List[Any] = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: lowercase__ , lowercase__ : Any = self.num_layers for i in range(a ): lowercase__ : Union[str, Any] = {0: 'batch', 2: 'past_sequence + sequence'} lowercase__ : int = {0: 'batch', 2: 'past_sequence + sequence'} else: lowercase__ : Optional[int] = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: lowercase__ : Dict = super().outputs else: lowercase__ : List[Any] = super(a , self ).outputs if self.use_past: lowercase__ , lowercase__ : Any = self.num_layers for i in range(a ): lowercase__ : Optional[int] = {0: 'batch', 2: 'past_sequence + sequence'} lowercase__ : Tuple = {0: 'batch', 2: 'past_sequence + sequence'} return common_outputs def _UpperCAmelCase ( self , a , a = -1 , a = -1 , a = False , a = None , ) -> Mapping[str, Any]: lowercase__ : Dict = self._generate_dummy_inputs_for_encoder_and_decoder( a , a , a , a , a ) # Generate decoder inputs lowercase__ : Union[str, Any] = seq_length if not self.use_past else 1 lowercase__ : Optional[Any] = self._generate_dummy_inputs_for_encoder_and_decoder( a , a , a , a , a ) lowercase__ : Any = {f"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} lowercase__ : List[str] = dict(**a , **a ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch lowercase__ , lowercase__ : List[str] = common_inputs['input_ids'].shape lowercase__ : Dict = common_inputs['decoder_input_ids'].shape[1] lowercase__ , lowercase__ : List[str] = self.num_attention_heads lowercase__ : Dict = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase__ : Any = decoder_seq_length + 3 lowercase__ : Dict = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase__ : Any = torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(a , a )] , dim=1 ) lowercase__ : Optional[int] = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase__ , lowercase__ : Any = self.num_layers lowercase__ : Optional[int] = min(a , a ) lowercase__ : Optional[int] = max(a , a ) - min_num_layers lowercase__ : Tuple = 'encoder' if num_encoder_layers > num_decoder_layers else 'decoder' for _ in range(a ): common_inputs["past_key_values"].append( ( torch.zeros(a ), torch.zeros(a ), torch.zeros(a ), torch.zeros(a ), ) ) # TODO: test this. lowercase__ : int = encoder_shape if remaining_side_name == 'encoder' else decoder_shape for _ in range(a , a ): common_inputs["past_key_values"].append((torch.zeros(a ), torch.zeros(a )) ) return common_inputs def _UpperCAmelCase ( self , a , a = -1 , a = -1 , a = False , a = None , ) -> Mapping[str, Any]: lowercase__ : Any = self._generate_dummy_inputs_for_encoder_and_decoder( a , a , a , a , a ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch lowercase__ , lowercase__ : int = common_inputs['input_ids'].shape # Not using the same length for past_key_values lowercase__ : Union[str, Any] = seqlen + 2 lowercase__ , lowercase__ : Optional[int] = self.num_layers lowercase__ , lowercase__ : Any = self.num_attention_heads lowercase__ : Tuple = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase__ : List[Any] = common_inputs['attention_mask'].dtype lowercase__ : int = torch.cat( [common_inputs['attention_mask'], torch.ones(a , a , dtype=a )] , dim=1 ) lowercase__ : Union[str, Any] = [ (torch.zeros(a ), torch.zeros(a )) for _ in range(a ) ] return common_inputs def _UpperCAmelCase ( self , a , a = -1 , a = -1 , a = False , a = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase__ : Any = compute_effective_axis_dimension( a , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase__ : Optional[int] = tokenizer.num_special_tokens_to_add(a ) lowercase__ : Any = compute_effective_axis_dimension( a , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=a ) # Generate dummy inputs according to compute batch and sequence lowercase__ : Tuple = [' '.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase__ : Any = dict(tokenizer(a , return_tensors=a ) ) return common_inputs def _UpperCAmelCase ( self , a , a = -1 , a = -1 , a = False , a = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: lowercase__ : List[Any] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( a , batch_size=a , seq_length=a , is_pair=a , framework=a ) else: lowercase__ : Union[str, Any] = self._generate_dummy_inputs_for_causal_lm( a , batch_size=a , seq_length=a , is_pair=a , framework=a ) return common_inputs def _UpperCAmelCase ( self , a , a , a , a ) -> List[str]: if self.task in ["default", "seq2seq-lm"]: lowercase__ : Union[str, Any] = super()._flatten_past_key_values_(a , a , a , a ) else: lowercase__ : Union[str, Any] = super(a , self )._flatten_past_key_values_( a , a , a , a ) @property def _UpperCAmelCase ( self ) -> float: return 1e-4
77
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization lowerCamelCase__ : str = field(default="text-classification" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({"labels": ClassLabel}) lowerCamelCase__ : str = "text" lowerCamelCase__ : str = "labels" def _UpperCAmelCase ( self , a ) -> Tuple: if self.label_column not in features: raise ValueError(f"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , a ): raise ValueError(f"""Column {self.label_column} is not a ClassLabel.""" ) lowercase__ : Optional[Any] = copy.deepcopy(self ) lowercase__ : Optional[Any] = self.label_schema.copy() lowercase__ : Any = features[self.label_column] lowercase__ : Optional[Any] = label_schema return task_template @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return { self.text_column: "text", self.label_column: "labels", }
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int ): '''simple docstring''' return abs(_lowerCAmelCase ) if a == 0 else greatest_common_divisor(b % a , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int ): '''simple docstring''' while y: # --> when y=0 then loop will terminate and return x as final GCD. lowercase__ , lowercase__ : Optional[int] = y, x % y return abs(_lowerCAmelCase ) def a_ ( ): '''simple docstring''' try: lowercase__ : int = input('Enter two integers separated by comma (,): ' ).split(',' ) lowercase__ : Tuple = int(nums[0] ) lowercase__ : str = int(nums[1] ) print( f"""greatest_common_divisor({num_a}, {num_a}) = """ f"""{greatest_common_divisor(_lowerCAmelCase , _lowerCAmelCase )}""" ) print(f"""By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(_lowerCAmelCase , _lowerCAmelCase )}""" ) except (IndexError, UnboundLocalError, ValueError): print('Wrong input' ) if __name__ == "__main__": main()
77
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np _UpperCamelCase : Any = re.compile(r"\b(a|an|the)\b", re.UNICODE) _UpperCamelCase : Union[str, Any] = None def a_ ( ): '''simple docstring''' lowercase__ : Optional[int] = argparse.ArgumentParser('Official evaluation script for SQuAD version 2.0.' ) parser.add_argument('data_file' , metavar='data.json' , help='Input data JSON file.' ) parser.add_argument('pred_file' , metavar='pred.json' , help='Model predictions.' ) parser.add_argument( '--out-file' , '-o' , metavar='eval.json' , help='Write accuracy metrics to file (default is stdout).' ) parser.add_argument( '--na-prob-file' , '-n' , metavar='na_prob.json' , help='Model estimates of probability of no answer.' ) parser.add_argument( '--na-prob-thresh' , '-t' , type=_lowerCAmelCase , default=1.0 , help='Predict "" if no-answer probability exceeds this (default = 1.0).' , ) parser.add_argument( '--out-image-dir' , '-p' , metavar='out_images' , default=_lowerCAmelCase , help='Save precision-recall curves to directory.' ) parser.add_argument('--verbose' , '-v' , action='store_true' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Tuple = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Optional[int] = bool(qa['answers']['text'] ) return qid_to_has_ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' def remove_articles(_lowerCAmelCase : int ): return ARTICLES_REGEX.sub(' ' , _lowerCAmelCase ) def white_space_fix(_lowerCAmelCase : str ): return " ".join(text.split() ) def remove_punc(_lowerCAmelCase : List[Any] ): lowercase__ : int = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_lowerCAmelCase : List[str] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_lowerCAmelCase ) ) ) ) def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' if not s: return [] return normalize_answer(_lowerCAmelCase ).split() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ): '''simple docstring''' return int(normalize_answer(_lowerCAmelCase ) == normalize_answer(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ): '''simple docstring''' lowercase__ : Dict = get_tokens(_lowerCAmelCase ) lowercase__ : List[str] = get_tokens(_lowerCAmelCase ) lowercase__ : List[Any] = collections.Counter(_lowerCAmelCase ) & collections.Counter(_lowerCAmelCase ) lowercase__ : int = sum(common.values() ) if len(_lowerCAmelCase ) == 0 or len(_lowerCAmelCase ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 lowercase__ : Any = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Dict = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Any = (2 * precision * recall) / (precision + recall) return fa def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Optional[int] = {} lowercase__ : Union[str, Any] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Any = qa['id'] lowercase__ : Union[str, Any] = [t for t in qa['answers']['text'] if normalize_answer(_lowerCAmelCase )] if not gold_answers: # For unanswerable questions, only correct answer is empty string lowercase__ : Dict = [''] if qid not in preds: print(f"""Missing prediction for {qid}""" ) continue lowercase__ : Optional[int] = preds[qid] # Take max over all gold answers lowercase__ : int = max(compute_exact(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) lowercase__ : Optional[Any] = max(compute_fa(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) return exact_scores, fa_scores def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : str = {} for qid, s in scores.items(): lowercase__ : int = na_probs[qid] > na_prob_thresh if pred_na: lowercase__ : Optional[Any] = float(not qid_to_has_ans[qid] ) else: lowercase__ : Optional[Any] = s return new_scores def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str=None ): '''simple docstring''' if not qid_list: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores.values() ) / total), ('f1', 1_0_0.0 * sum(fa_scores.values() ) / total), ('total', total), ] ) else: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores[k] for k in qid_list ) / total), ('f1', 1_0_0.0 * sum(fa_scores[k] for k in qid_list ) / total), ('total', total), ] ) def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' for k in new_eval: lowercase__ : int = new_eval[k] def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple ): '''simple docstring''' plt.step(_lowerCAmelCase , _lowerCAmelCase , color='b' , alpha=0.2 , where='post' ) plt.fill_between(_lowerCAmelCase , _lowerCAmelCase , step='post' , alpha=0.2 , color='b' ) plt.xlabel('Recall' ) plt.ylabel('Precision' ) plt.xlim([0.0, 1.0_5] ) plt.ylim([0.0, 1.0_5] ) plt.title(_lowerCAmelCase ) plt.savefig(_lowerCAmelCase ) plt.clf() def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : str , _lowerCAmelCase : Any=None , _lowerCAmelCase : List[str]=None ): '''simple docstring''' lowercase__ : Optional[int] = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) lowercase__ : Tuple = 0.0 lowercase__ : List[str] = 1.0 lowercase__ : List[str] = 0.0 lowercase__ : Union[str, Any] = [1.0] lowercase__ : List[Any] = [0.0] lowercase__ : Optional[int] = 0.0 for i, qid in enumerate(_lowerCAmelCase ): if qid_to_has_ans[qid]: true_pos += scores[qid] lowercase__ : Tuple = true_pos / float(i + 1 ) lowercase__ : Union[str, Any] = true_pos / float(_lowerCAmelCase ) if i == len(_lowerCAmelCase ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_lowerCAmelCase ) recalls.append(_lowerCAmelCase ) if out_image: plot_pr_curve(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return {"ap": 1_0_0.0 * avg_prec} def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple ): '''simple docstring''' if out_image_dir and not os.path.exists(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) lowercase__ : List[str] = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return lowercase__ : Dict = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_exact.png' ) , title='Precision-Recall curve for Exact Match score' , ) lowercase__ : Tuple = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_f1.png' ) , title='Precision-Recall curve for F1 score' , ) lowercase__ : List[Any] = {k: float(_lowerCAmelCase ) for k, v in qid_to_has_ans.items()} lowercase__ : Any = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_oracle.png' ) , title='Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)' , ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_exact' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_f1' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_oracle' ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' if not qid_list: return lowercase__ : List[str] = [na_probs[k] for k in qid_list] lowercase__ : Tuple = np.ones_like(_lowerCAmelCase ) / float(len(_lowerCAmelCase ) ) plt.hist(_lowerCAmelCase , weights=_lowerCAmelCase , bins=20 , range=(0.0, 1.0) ) plt.xlabel('Model probability of no-answer' ) plt.ylabel('Proportion of dataset' ) plt.title(f"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(_lowerCAmelCase , f"""na_prob_hist_{name}.png""" ) ) plt.clf() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : Tuple = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) lowercase__ : int = num_no_ans lowercase__ : Optional[int] = cur_score lowercase__ : Tuple = 0.0 lowercase__ : Dict = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) for i, qid in enumerate(_lowerCAmelCase ): if qid not in scores: continue if qid_to_has_ans[qid]: lowercase__ : Optional[int] = scores[qid] else: if preds[qid]: lowercase__ : List[Any] = -1 else: lowercase__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: lowercase__ : Dict = cur_score lowercase__ : Optional[int] = na_probs[qid] return 1_0_0.0 * best_score / len(_lowerCAmelCase ), best_thresh def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str ): '''simple docstring''' lowercase__ , lowercase__ : List[Any] = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ , lowercase__ : Dict = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Any = best_exact lowercase__ : Tuple = exact_thresh lowercase__ : Optional[Any] = best_fa lowercase__ : Any = fa_thresh def a_ ( ): '''simple docstring''' with open(OPTS.data_file ) as f: lowercase__ : List[Any] = json.load(_lowerCAmelCase ) lowercase__ : Union[str, Any] = dataset_json['data'] with open(OPTS.pred_file ) as f: lowercase__ : str = json.load(_lowerCAmelCase ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: lowercase__ : Union[str, Any] = json.load(_lowerCAmelCase ) else: lowercase__ : str = {k: 0.0 for k in preds} lowercase__ : int = make_qid_to_has_ans(_lowerCAmelCase ) # maps qid to True/False lowercase__ : List[str] = [k for k, v in qid_to_has_ans.items() if v] lowercase__ : Any = [k for k, v in qid_to_has_ans.items() if not v] lowercase__ , lowercase__ : Any = get_raw_scores(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Optional[Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Union[str, Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Tuple = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase ) if has_ans_qids: lowercase__ : int = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'HasAns' ) if no_ans_qids: lowercase__ : Optional[Any] = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'NoAns' ) if OPTS.na_prob_file: find_all_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'hasAns' ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'noAns' ) if OPTS.out_file: with open(OPTS.out_file , 'w' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) else: print(json.dumps(_lowerCAmelCase , indent=2 ) ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
77
1
"""simple docstring""" from math import sqrt def a_ ( _lowerCAmelCase : int = 100_0000 ): '''simple docstring''' lowercase__ : int = 0 lowercase__ : int = 0 lowercase__ : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(_lowerCAmelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
77
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = 0 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Tuple = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Any: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : str = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : int = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = CLIPConfig() # Create a dummy config file with image_proceesor_type lowercase__ : Optional[int] = Path(a ) / 'preprocessor_config.json' lowercase__ : Optional[int] = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowercase__ : int = AutoImageProcessor.from_pretrained(a ).to_dict() config_dict.pop('image_processor_type' ) lowercase__ : Tuple = CLIPImageProcessor(**a ) # save in new folder model_config.save_pretrained(a ) config.save_pretrained(a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) # make sure private variable is not incorrectly saved lowercase__ : Optional[int] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = Path(a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'clip-base is not a local folder and is not a valid model identifier' ): lowercase__ : Any = AutoImageProcessor.from_pretrained('clip-base' ) def _UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex( a , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowercase__ : Dict = AutoImageProcessor.from_pretrained(a , revision='aaaaaa' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def _UpperCAmelCase ( self ) -> Optional[int]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(a ): lowercase__ : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : str = AutoImageProcessor.from_pretrained(a , trust_remote_code=a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def _UpperCAmelCase ( self ) -> int: try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoImageProcessor.register(a , a ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Optional[Any] = Path(a ) / 'preprocessor_config.json' lowercase__ : List[Any] = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = CustomImageProcessor.from_pretrained(a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCAmelCase ( self ) -> Dict: class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = True try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # If remote code is not set, the default is to use local lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowercase__ : int = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
77
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCamelCase : Optional[Any] = logging.get_logger(__name__) class UpperCAmelCase_ ( _a): def __init__( self , *a , **a ) -> None: warnings.warn( 'The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use YolosImageProcessor instead.' , a , ) super().__init__(*a , **a )
77
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class UpperCAmelCase_ ( _a): lowerCamelCase__ : Dict = ["image_processor", "tokenizer"] lowerCamelCase__ : Dict = "BlipImageProcessor" lowerCamelCase__ : Union[str, Any] = "AutoTokenizer" def __init__( self , a , a , a ) -> Optional[int]: super().__init__(a , a ) # add QFormer tokenizer lowercase__ : Dict = qformer_tokenizer def __call__( self , a = None , a = None , a = True , a = False , a = None , a = None , a = 0 , a = None , a = None , a = False , a = False , a = False , a = False , a = False , a = True , a = None , **a , ) -> BatchFeature: if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) lowercase__ : List[Any] = BatchFeature() if text is not None: lowercase__ : Optional[int] = self.tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) encoding.update(a ) lowercase__ : Optional[int] = self.qformer_tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) lowercase__ : List[str] = qformer_text_encoding.pop('input_ids' ) lowercase__ : Any = qformer_text_encoding.pop('attention_mask' ) if images is not None: lowercase__ : List[Any] = self.image_processor(a , return_tensors=a ) encoding.update(a ) return encoding def _UpperCAmelCase ( self , *a , **a ) -> List[str]: return self.tokenizer.batch_decode(*a , **a ) def _UpperCAmelCase ( self , *a , **a ) -> Tuple: return self.tokenizer.decode(*a , **a ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : str = self.tokenizer.model_input_names lowercase__ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: if os.path.isfile(a ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(a , exist_ok=a ) lowercase__ : int = os.path.join(a , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(a ) return super().save_pretrained(a , **a ) @classmethod def _UpperCAmelCase ( cls , a , **a ) -> str: lowercase__ : str = AutoTokenizer.from_pretrained(a , subfolder='qformer_tokenizer' ) lowercase__ : int = cls._get_arguments_from_pretrained(a , **a ) args.append(a ) return cls(*a )
77
1
"""simple docstring""" import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup _UpperCamelCase : Dict = logging.get_logger(__name__) class UpperCAmelCase_ ( _a): def __init__( self , **a ) -> Tuple: requires_backends(self , ['bs4'] ) super().__init__(**a ) def _UpperCAmelCase ( self , a ) -> Optional[int]: lowercase__ : int = [] lowercase__ : List[str] = [] lowercase__ : Tuple = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag lowercase__ : Tuple = parent.find_all(child.name , recursive=a ) xpath_tags.append(child.name ) xpath_subscripts.append( 0 if 1 == len(a ) else next(i for i, s in enumerate(a , 1 ) if s is child ) ) lowercase__ : Dict = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : Any = BeautifulSoup(a , 'html.parser' ) lowercase__ : str = [] lowercase__ : Optional[int] = [] lowercase__ : str = [] for element in html_code.descendants: if type(a ) == bsa.element.NavigableString: if type(element.parent ) != bsa.element.Tag: continue lowercase__ : List[str] = html.unescape(a ).strip() if not text_in_this_tag: continue all_doc_strings.append(a ) lowercase__ , lowercase__ : Optional[Any] = self.xpath_soup(a ) stringaxtag_seq.append(a ) stringaxsubs_seq.append(a ) if len(a ) != len(a ): raise ValueError('Number of doc strings and xtags does not correspond' ) if len(a ) != len(a ): raise ValueError('Number of doc strings and xsubs does not correspond' ) return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def _UpperCAmelCase ( self , a , a ) -> Optional[Any]: lowercase__ : str = '' for tagname, subs in zip(a , a ): xpath += f"""/{tagname}""" if subs != 0: xpath += f"""[{subs}]""" return xpath def __call__( self , a ) -> BatchFeature: lowercase__ : Union[str, Any] = False # Check that strings has a valid type if isinstance(a , a ): lowercase__ : Union[str, Any] = True elif isinstance(a , (list, tuple) ): if len(a ) == 0 or isinstance(html_strings[0] , a ): lowercase__ : List[str] = True if not valid_strings: raise ValueError( 'HTML strings must of type `str`, `List[str]` (batch of examples), ' f"""but is of type {type(a )}.""" ) lowercase__ : Union[str, Any] = bool(isinstance(a , (list, tuple) ) and (isinstance(html_strings[0] , a )) ) if not is_batched: lowercase__ : Any = [html_strings] # Get nodes + xpaths lowercase__ : Dict = [] lowercase__ : List[str] = [] for html_string in html_strings: lowercase__ , lowercase__ , lowercase__ : Tuple = self.get_three_from_single(a ) nodes.append(a ) lowercase__ : Tuple = [] for node, tag_list, sub_list in zip(a , a , a ): lowercase__ : List[Any] = self.construct_xpath(a , a ) xpath_strings.append(a ) xpaths.append(a ) # return as Dict lowercase__ : List[Any] = {'nodes': nodes, 'xpaths': xpaths} lowercase__ : List[Any] = BatchFeature(data=a , tensor_type=a ) return encoded_inputs
77
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({}) lowerCamelCase__ : str = "text" @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return {self.text_column: "text"}
77
1
"""simple docstring""" import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow _UpperCamelCase : Optional[int] = False class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self , a=3_2 ) -> Any: set_seed(0 ) lowercase__ : int = UNetaDModel(sample_size=a , in_channels=3 , out_channels=3 ) lowercase__ : Tuple = torch.optim.SGD(model.parameters() , lr=0.0_001 ) return model, optimizer @slow def _UpperCAmelCase ( self ) -> Any: lowercase__ : str = 'cpu' # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable lowercase__ : Optional[int] = DDPMScheduler( num_train_timesteps=1_0_0_0 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=a , ) lowercase__ : Dict = DDIMScheduler( num_train_timesteps=1_0_0_0 , beta_start=0.0_001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=a , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) lowercase__ : str = [torch.randn((4, 3, 3_2, 3_2) ).clip(-1 , 1 ).to(a ) for _ in range(4 )] lowercase__ : str = [torch.randn((4, 3, 3_2, 3_2) ).to(a ) for _ in range(4 )] lowercase__ : Any = [torch.randint(0 , 1_0_0_0 , (4,) ).long().to(a ) for _ in range(4 )] # train with a DDPM scheduler lowercase__ , lowercase__ : Any = self.get_model_optimizer(resolution=3_2 ) model.train().to(a ) for i in range(4 ): optimizer.zero_grad() lowercase__ : Optional[Any] = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowercase__ : Optional[int] = model(a , timesteps[i] ).sample lowercase__ : List[str] = torch.nn.functional.mse_loss(a , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM lowercase__ , lowercase__ : Any = self.get_model_optimizer(resolution=3_2 ) model.train().to(a ) for i in range(4 ): optimizer.zero_grad() lowercase__ : str = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) lowercase__ : Tuple = model(a , timesteps[i] ).sample lowercase__ : Union[str, Any] = torch.nn.functional.mse_loss(a , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(a , a , atol=1e-5 ) ) self.assertTrue(torch.allclose(a , a , atol=1e-5 ) )
77
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
77
1
"""simple docstring""" import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py _UpperCamelCase : Optional[Any] = "\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Automatic Evaluation of Machine Translation},\n booktitle = {},\n year = {2002},\n pages = {311--318}\n}\n@inproceedings{lin-och-2004-orange,\n title = \"{ORANGE}: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation\",\n author = \"Lin, Chin-Yew and\n Och, Franz Josef\",\n booktitle = \"{COLING} 2004: Proceedings of the 20th International Conference on Computational Linguistics\",\n month = \"aug 23{--}aug 27\",\n year = \"2004\",\n address = \"Geneva, Switzerland\",\n publisher = \"COLING\",\n url = \"https://www.aclweb.org/anthology/C04-1072\",\n pages = \"501--507\",\n}\n" _UpperCamelCase : Dict = "\\nBLEU (bilingual evaluation understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another.\nQuality is considered to be the correspondence between a machine's output and that of a human: \"the closer a machine translation is to a professional human translation,\nthe better it is\" – this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and\nremains one of the most popular automated and inexpensive metrics.\n\nScores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations.\nThose scores are then averaged over the whole corpus to reach an estimate of the translation's overall quality. Intelligibility or grammatical correctness\nare not taken into account[citation needed].\n\nBLEU's output is always a number between 0 and 1. This value indicates how similar the candidate text is to the reference texts, with values closer to 1\nrepresenting more similar texts. Few human translations will attain a score of 1, since this would indicate that the candidate is identical to one of the\nreference translations. For this reason, it is not necessary to attain a score of 1. Because there are more opportunities to match, adding additional\nreference translations will increase the BLEU score.\n" _UpperCamelCase : Optional[int] = "\nComputes BLEU score of translated segments against one or more references.\nArgs:\n predictions: list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n max_order: Maximum n-gram order to use when computing BLEU score.\n smooth: Whether or not to apply Lin et al. 2004 smoothing.\nReturns:\n 'bleu': bleu score,\n 'precisions': geometric mean of n-gram precisions,\n 'brevity_penalty': brevity penalty,\n 'length_ratio': ratio of lengths,\n 'translation_length': translation_length,\n 'reference_length': reference_length\nExamples:\n\n >>> predictions = [\n ... [\"hello\", \"there\", \"general\", \"kenobi\"], # tokenized prediction of the first sample\n ... [\"foo\", \"bar\", \"foobar\"] # tokenized prediction of the second sample\n ... ]\n >>> references = [\n ... [[\"hello\", \"there\", \"general\", \"kenobi\"], [\"hello\", \"there\", \"!\"]], # tokenized references for the first sample (2 references)\n ... [[\"foo\", \"bar\", \"foobar\"]] # tokenized references for the second sample (1 reference)\n ... ]\n >>> bleu = datasets.load_metric(\"bleu\")\n >>> results = bleu.compute(predictions=predictions, references=references)\n >>> print(results[\"bleu\"])\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class UpperCAmelCase_ ( datasets.Metric): def _UpperCAmelCase ( self ) -> Optional[int]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' , id='token' ) , id='sequence' ), 'references': datasets.Sequence( datasets.Sequence(datasets.Value('string' , id='token' ) , id='sequence' ) , id='references' ), } ) , codebase_urls=['https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py'] , reference_urls=[ 'https://en.wikipedia.org/wiki/BLEU', 'https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213', ] , ) def _UpperCAmelCase ( self , a , a , a=4 , a=False ) -> Optional[int]: lowercase__ : Any = compute_bleu( reference_corpus=a , translation_corpus=a , max_order=a , smooth=a ) ((lowercase__) , (lowercase__) , (lowercase__) , (lowercase__) , (lowercase__) , (lowercase__)) : Tuple = score return { "bleu": bleu, "precisions": precisions, "brevity_penalty": bp, "length_ratio": ratio, "translation_length": translation_length, "reference_length": reference_length, }
77
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING _UpperCamelCase : Dict = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , **a ) -> Dict: super().__init__(**a ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type(a ) def __call__( self , a , a = None , **a , ) -> List[str]: if "text_queries" in kwargs: lowercase__ : Optional[Any] = kwargs.pop('text_queries' ) if isinstance(a , (str, Image.Image) ): lowercase__ : Optional[Any] = {'image': image, 'candidate_labels': candidate_labels} else: lowercase__ : List[str] = image lowercase__ : Optional[Any] = super().__call__(a , **a ) return results def _UpperCAmelCase ( self , **a ) -> Dict: lowercase__ : Optional[Any] = {} if "threshold" in kwargs: lowercase__ : Tuple = kwargs['threshold'] if "top_k" in kwargs: lowercase__ : List[Any] = kwargs['top_k'] return {}, {}, postprocess_params def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : Any = load_image(inputs['image'] ) lowercase__ : Optional[int] = inputs['candidate_labels'] if isinstance(a , a ): lowercase__ : Optional[int] = candidate_labels.split(',' ) lowercase__ : Optional[int] = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(a ): lowercase__ : List[str] = self.tokenizer(a , return_tensors=self.framework ) lowercase__ : List[Any] = self.image_processor(a , return_tensors=self.framework ) yield { "is_last": i == len(a ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : List[Any] = model_inputs.pop('target_size' ) lowercase__ : Dict = model_inputs.pop('candidate_label' ) lowercase__ : Dict = model_inputs.pop('is_last' ) lowercase__ : Optional[int] = self.model(**a ) lowercase__ : Any = {'target_size': target_size, 'candidate_label': candidate_label, 'is_last': is_last, **outputs} return model_outputs def _UpperCAmelCase ( self , a , a=0.1 , a=None ) -> Union[str, Any]: lowercase__ : Dict = [] for model_output in model_outputs: lowercase__ : List[Any] = model_output['candidate_label'] lowercase__ : Optional[int] = BaseModelOutput(a ) lowercase__ : Any = self.image_processor.post_process_object_detection( outputs=a , threshold=a , target_sizes=model_output['target_size'] )[0] for index in outputs["scores"].nonzero(): lowercase__ : Union[str, Any] = outputs['scores'][index].item() lowercase__ : Tuple = self._get_bounding_box(outputs['boxes'][index][0] ) lowercase__ : Tuple = {'score': score, 'label': label, 'box': box} results.append(a ) lowercase__ : Dict = sorted(a , key=lambda a : x["score"] , reverse=a ) if top_k: lowercase__ : Dict = results[:top_k] return results def _UpperCAmelCase ( self , a ) -> Dict[str, int]: if self.framework != "pt": raise ValueError('The ZeroShotObjectDetectionPipeline is only available in PyTorch.' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ : Tuple = box.int().tolist() lowercase__ : Any = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
77
1
"""simple docstring""" # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position _UpperCamelCase : Optional[int] = "2.13.1" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("3.7"): raise ImportWarning( "To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition." ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( "To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n" "If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`." ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip _UpperCamelCase : Any = concatenate_datasets _UpperCamelCase : Dict = DownloadConfig _UpperCamelCase : Dict = DownloadManager _UpperCamelCase : Dict = DownloadMode _UpperCamelCase : int = DownloadConfig _UpperCamelCase : Union[str, Any] = DownloadMode _UpperCamelCase : Dict = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
77
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ : def __init__( self , a , a=1_2 , a=7 , a=True , a=True , a=True , a=9_9 , a=3_2 , a=3_2 , a=2 , a=4 , a=3_7 , a=0.1 , a=0.1 , a=5_1_2 , a=0.02 , a=0 , a=None , ) -> Union[str, Any]: lowercase__ : Any = parent lowercase__ : str = batch_size lowercase__ : List[Any] = seq_length lowercase__ : Union[str, Any] = is_training lowercase__ : List[str] = use_input_mask lowercase__ : int = use_labels lowercase__ : List[Any] = vocab_size lowercase__ : str = hidden_size lowercase__ : int = projection_dim lowercase__ : Optional[int] = num_hidden_layers lowercase__ : Any = num_attention_heads lowercase__ : Optional[Any] = intermediate_size lowercase__ : Optional[Any] = dropout lowercase__ : Optional[int] = attention_dropout lowercase__ : Optional[int] = max_position_embeddings lowercase__ : str = initializer_range lowercase__ : Tuple = scope lowercase__ : int = bos_token_id def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : int = None if self.use_input_mask: lowercase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowercase__ : int = input_mask.numpy() lowercase__ , lowercase__ : Tuple = input_mask.shape lowercase__ : List[str] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a ): lowercase__ : Dict = 1 lowercase__ : Union[str, Any] = 0 lowercase__ : Tuple = self.get_config() return config, input_ids, tf.convert_to_tensor(a ) def _UpperCAmelCase ( self ) -> List[Any]: return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _UpperCAmelCase ( self , a , a , a ) -> Any: lowercase__ : List[Any] = TFBlipTextModel(config=a ) lowercase__ : Optional[int] = model(a , attention_mask=a , training=a ) lowercase__ : List[str] = model(a , training=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _UpperCAmelCase ( self ) -> Any: lowercase__ : Optional[Any] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ : Any = config_and_inputs lowercase__ : Any = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : Dict = (TFBlipTextModel,) if is_tf_available() else () lowerCamelCase__ : Optional[Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Optional[int] = BlipTextModelTester(self ) lowercase__ : int = ConfigTester(self , config_class=a , hidden_size=3_7 ) def _UpperCAmelCase ( self ) -> Tuple: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> int: lowercase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _UpperCAmelCase ( self ) -> Optional[Any]: pass def _UpperCAmelCase ( self ) -> Optional[int]: pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def _UpperCAmelCase ( self ) -> List[str]: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> Dict: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> str: pass @slow def _UpperCAmelCase ( self ) -> int: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Any = TFBlipTextModel.from_pretrained(a ) self.assertIsNotNone(a ) def _UpperCAmelCase ( self , a=True ) -> List[str]: super().test_pt_tf_model_equivalence(allow_missing_keys=a )
77
1
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : str = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all LED models at https://huggingface.co/models?filter=LED _UpperCamelCase : Optional[Any] = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } _UpperCamelCase : Optional[int] = { "allenai/led-base-16384": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def a_ ( ): '''simple docstring''' lowercase__ : int = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowercase__ : Union[str, Any] = bs[:] lowercase__ : str = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCAmelCase ) cs.append(2**8 + n ) n += 1 lowercase__ : str = [chr(_lowerCAmelCase ) for n in cs] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Dict = set() lowercase__ : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Optional[Any] = char return pairs class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = VOCAB_FILES_NAMES lowerCamelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , a , a , a="replace" , a="<s>" , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a=False , **a , ) -> Any: lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else bos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else eos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else sep_token lowercase__ : Dict = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else cls_token lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else unk_token lowercase__ : Tuple = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( errors=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , add_prefix_space=a , **a , ) with open(a , encoding='utf-8' ) as vocab_handle: lowercase__ : Tuple = json.load(a ) lowercase__ : Dict = {v: k for k, v in self.encoder.items()} lowercase__ : str = errors # how to handle errors in decoding lowercase__ : Optional[Any] = bytes_to_unicode() lowercase__ : Optional[Any] = {v: k for k, v in self.byte_encoder.items()} with open(a , encoding='utf-8' ) as merges_handle: lowercase__ : Optional[Any] = merges_handle.read().split('\n' )[1:-1] lowercase__ : Optional[int] = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ : Union[str, Any] = dict(zip(a , range(len(a ) ) ) ) lowercase__ : Tuple = {} lowercase__ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ : List[Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _UpperCAmelCase ( self ) -> List[Any]: return len(self.encoder ) def _UpperCAmelCase ( self ) -> str: return dict(self.encoder , **self.added_tokens_encoder ) def _UpperCAmelCase ( self , a ) -> List[str]: if token in self.cache: return self.cache[token] lowercase__ : Optional[Any] = tuple(a ) lowercase__ : int = get_pairs(a ) if not pairs: return token while True: lowercase__ : List[str] = min(a , key=lambda a : self.bpe_ranks.get(a , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ : List[str] = bigram lowercase__ : Union[str, Any] = [] lowercase__ : List[Any] = 0 while i < len(a ): try: lowercase__ : str = word.index(a , a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : Optional[int] = j if word[i] == first and i < len(a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : int = tuple(a ) lowercase__ : Dict = new_word if len(a ) == 1: break else: lowercase__ : Any = get_pairs(a ) lowercase__ : List[str] = ' '.join(a ) lowercase__ : Optional[Any] = word return word def _UpperCAmelCase ( self , a ) -> Union[str, Any]: lowercase__ : Tuple = [] for token in re.findall(self.pat , a ): lowercase__ : Union[str, Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(a ).split(' ' ) ) return bpe_tokens def _UpperCAmelCase ( self , a ) -> Optional[Any]: return self.encoder.get(a , self.encoder.get(self.unk_token ) ) def _UpperCAmelCase ( self , a ) -> Optional[int]: return self.decoder.get(a ) def _UpperCAmelCase ( self , a ) -> str: lowercase__ : Any = ''.join(a ) lowercase__ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Any = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : str = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(a , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a , ensure_ascii=a ) + '\n' ) lowercase__ : List[Any] = 0 with open(a , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) lowercase__ : Union[str, Any] = token_index writer.write(' '.join(a ) + '\n' ) index += 1 return vocab_file, merge_file def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ : Union[str, Any] = [self.cls_token_id] lowercase__ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) + [1] def _UpperCAmelCase ( self , a , a = None ) -> List[int]: lowercase__ : Dict = [self.sep_token_id] lowercase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCAmelCase ( self , a , a=False , **a ) -> Optional[int]: lowercase__ : Tuple = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(a ) > 0 and not text[0].isspace()): lowercase__ : List[str] = ' ' + text return (text, kwargs) def _UpperCAmelCase ( self , a , a = None , a = PaddingStrategy.DO_NOT_PAD , a = None , a = None , ) -> dict: lowercase__ : Dict = super()._pad( encoded_inputs=a , max_length=a , padding_strategy=a , pad_to_multiple_of=a , return_attention_mask=a , ) # Load from model defaults if return_attention_mask is None: lowercase__ : Union[str, Any] = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ : Any = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ : Tuple = len(encoded_inputs['global_attention_mask'] ) != len(a ) if needs_to_be_padded: lowercase__ : str = len(a ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ : Union[str, Any] = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowercase__ : List[str] = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
77
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase : Tuple = logging.get_logger(__name__) _UpperCamelCase : Union[str, Any] = { "alibaba-damo/mgp-str-base": "https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json", } class UpperCAmelCase_ ( _a): lowerCamelCase__ : List[Any] = "mgp-str" def __init__( self , a=[3_2, 1_2_8] , a=4 , a=3 , a=2_7 , a=3_8 , a=5_0_2_5_7 , a=3_0_5_2_2 , a=7_6_8 , a=1_2 , a=1_2 , a=4.0 , a=True , a=False , a=1e-5 , a=0.0 , a=0.0 , a=0.0 , a=False , a=0.02 , **a , ) -> Tuple: super().__init__(**a ) lowercase__ : int = image_size lowercase__ : List[Any] = patch_size lowercase__ : Optional[Any] = num_channels lowercase__ : Optional[Any] = max_token_length lowercase__ : Dict = num_character_labels lowercase__ : Optional[int] = num_bpe_labels lowercase__ : Dict = num_wordpiece_labels lowercase__ : Tuple = hidden_size lowercase__ : List[Any] = num_hidden_layers lowercase__ : Union[str, Any] = num_attention_heads lowercase__ : Dict = mlp_ratio lowercase__ : Optional[int] = distilled lowercase__ : Optional[int] = layer_norm_eps lowercase__ : Optional[int] = drop_rate lowercase__ : List[str] = qkv_bias lowercase__ : Optional[int] = attn_drop_rate lowercase__ : Any = drop_path_rate lowercase__ : List[Any] = output_aa_attentions lowercase__ : Tuple = initializer_range
77
1
"""simple docstring""" import re import string import numpy as np import datasets _UpperCamelCase : List[str] = "\nReturns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list.\n" _UpperCamelCase : Optional[Any] = "\nArgs:\n predictions: List of predicted texts.\n references: List of reference texts.\n regexes_to_ignore: List, defaults to None. Regex expressions of characters to\n ignore when calculating the exact matches. Note: these regexes are removed\n from the input data before the changes based on the options below (e.g. ignore_case,\n ignore_punctuation, ignore_numbers) are applied.\n ignore_case: Boolean, defaults to False. If true, turns everything\n to lowercase so that capitalization differences are ignored.\n ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\n ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\nReturns:\n exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive.\nExamples:\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results[\"exact_match\"], 1))\n 25.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results[\"exact_match\"], 1))\n 50.0\n\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results[\"exact_match\"], 1))\n 75.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True)\n >>> print(round(results[\"exact_match\"], 1))\n 100.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"The cat sat on the mat.\", \"Theaters are great.\", \"It's like comparing oranges and apples.\"]\n >>> preds = [\"The cat sat on the mat?\", \"Theaters are great.\", \"It's like comparing apples and oranges.\"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results[\"exact_match\"], 1))\n 33.3\n\n" _UpperCamelCase : Optional[Any] = "\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class UpperCAmelCase_ ( datasets.Metric): def _UpperCAmelCase ( self ) -> Optional[int]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , reference_urls=[] , ) def _UpperCAmelCase ( self , a , a , a=None , a=False , a=False , a=False , ) -> Union[str, Any]: if regexes_to_ignore is not None: for s in regexes_to_ignore: lowercase__ : List[Any] = np.array([re.sub(a , '' , a ) for x in predictions] ) lowercase__ : Tuple = np.array([re.sub(a , '' , a ) for x in references] ) else: lowercase__ : Optional[int] = np.asarray(a ) lowercase__ : Union[str, Any] = np.asarray(a ) if ignore_case: lowercase__ : Tuple = np.char.lower(a ) lowercase__ : List[Any] = np.char.lower(a ) if ignore_punctuation: lowercase__ : Any = string.punctuation.maketrans('' , '' , string.punctuation ) lowercase__ : Union[str, Any] = np.char.translate(a , table=a ) lowercase__ : Optional[int] = np.char.translate(a , table=a ) if ignore_numbers: lowercase__ : Any = string.digits.maketrans('' , '' , string.digits ) lowercase__ : Union[str, Any] = np.char.translate(a , table=a ) lowercase__ : List[Any] = np.char.translate(a , table=a ) lowercase__ : Any = predictions == references return {"exact_match": np.mean(a ) * 1_0_0}
77
"""simple docstring""" import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : int = FunnelConfig.from_json_file(_lowerCAmelCase ) print(f"""Building PyTorch model from configuration: {config}""" ) lowercase__ : List[Any] = FunnelBaseModel(_lowerCAmelCase ) if base_model else FunnelModel(_lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_funnel(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _lowerCAmelCase ) if __name__ == "__main__": _UpperCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--base_model", action="store_true", help="Whether you want just the base model (no decoder) or not." ) _UpperCamelCase : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
77
1
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _UpperCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , *a , **a ) -> Union[str, Any]: super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _UpperCAmelCase ( self , a=None ) -> Dict: lowercase__ : Any = {} if top_k is not None: lowercase__ : List[str] = top_k return {}, {}, postprocess_params def __call__( self , a , **a ) -> Tuple: return super().__call__(a , **a ) def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : List[Any] = load_image(a ) lowercase__ : Union[str, Any] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _UpperCAmelCase ( self , a ) -> List[str]: lowercase__ : Dict = self.model(**a ) return model_outputs def _UpperCAmelCase ( self , a , a=5 ) -> Dict: if top_k > self.model.config.num_labels: lowercase__ : List[Any] = self.model.config.num_labels if self.framework == "pt": lowercase__ : Tuple = model_outputs.logits.softmax(-1 )[0] lowercase__ , lowercase__ : Optional[Any] = probs.topk(a ) elif self.framework == "tf": lowercase__ : Union[str, Any] = stable_softmax(model_outputs.logits , axis=-1 )[0] lowercase__ : str = tf.math.top_k(a , k=a ) lowercase__ , lowercase__ : Dict = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) lowercase__ : Dict = scores.tolist() lowercase__ : Dict = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
77
"""simple docstring""" import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _UpperCamelCase : Any = logging.getLogger(__name__) _UpperCamelCase : Optional[Any] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _UpperCamelCase : Any = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The model checkpoint for weights initialization. Leave None if you want to train a model from" " scratch." ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(_a)} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained config name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "The input training data file (a text file)."}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The input training data files (multiple files in glob format). " "Very often splitting large files to smaller files can prevent tokenizer going out of memory" ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input train ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input eval ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Whether distinct lines of text in the dataset are to be handled as distinct sequences."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Train with masked-language modeling loss instead of language modeling."}) lowerCamelCase__ : bool = field(default=_a , metadata={"help": "Whether ot not to use whole word mask."}) lowerCamelCase__ : float = field( default=0.15 , metadata={"help": "Ratio of tokens to mask for masked language modeling loss"}) lowerCamelCase__ : float = field( default=1 / 6 , metadata={ "help": ( "Ratio of length of a span of masked tokens to surrounding context length for permutation language" " modeling." ) } , ) lowerCamelCase__ : int = field( default=5 , metadata={"help": "Maximum length of a span of masked tokens for permutation language modeling."}) lowerCamelCase__ : int = field( default=-1 , metadata={ "help": ( "Optional input sequence length after tokenization." "The training dataset will be truncated in block of this size for training." "Default to the model max input length for single sentence inputs (take into account special tokens)." ) } , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Overwrite the cached training and evaluation sets"}) def a_ ( _lowerCAmelCase : DataTrainingArguments , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[str] = None , ): '''simple docstring''' def _dataset(_lowerCAmelCase : Any , _lowerCAmelCase : Any=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('You need to set world whole masking and mlm to True for Chinese Whole Word Mask' ) return LineByLineWithRefDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , ref_path=_lowerCAmelCase , ) return LineByLineTextDataset(tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size ) else: return TextDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=_lowerCAmelCase , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(_lowerCAmelCase ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def a_ ( ): '''simple docstring''' lowercase__ : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ : List[Any] = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( 'Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ' 'or remove the --do_eval argument.' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s' , _lowerCAmelCase ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: lowercase__ : List[Any] = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : Any = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: lowercase__ : Union[str, Any] = CONFIG_MAPPING[model_args.model_type]() logger.warning('You are instantiating a new config instance from scratch.' ) if model_args.tokenizer_name: lowercase__ : Dict = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : int = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another' ' script, save it,and load it from here, using --tokenizer_name' ) if model_args.model_name_or_path: lowercase__ : Optional[Any] = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=_lowerCAmelCase , cache_dir=model_args.cache_dir , ) else: logger.info('Training new model from scratch' ) lowercase__ : int = AutoModelWithLMHead.from_config(_lowerCAmelCase ) model.resize_token_embeddings(len(_lowerCAmelCase ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( 'BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the' '--mlm flag (masked language modeling).' ) if data_args.block_size <= 0: lowercase__ : List[str] = tokenizer.max_len # Our input block size will be the max possible for the model else: lowercase__ : int = min(data_args.block_size , tokenizer.max_len ) # Get datasets lowercase__ : Tuple = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) lowercase__ : Optional[Any] = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , evaluate=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": lowercase__ : List[Any] = DataCollatorForPermutationLanguageModeling( tokenizer=_lowerCAmelCase , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: lowercase__ : List[str] = DataCollatorForWholeWordMask( tokenizer=_lowerCAmelCase , mlm_probability=data_args.mlm_probability ) else: lowercase__ : str = DataCollatorForLanguageModeling( tokenizer=_lowerCAmelCase , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer lowercase__ : Optional[int] = Trainer( model=_lowerCAmelCase , args=_lowerCAmelCase , data_collator=_lowerCAmelCase , train_dataset=_lowerCAmelCase , eval_dataset=_lowerCAmelCase , prediction_loss_only=_lowerCAmelCase , ) # Training if training_args.do_train: lowercase__ : List[str] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=_lowerCAmelCase ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase__ : List[Any] = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) lowercase__ : Dict = trainer.evaluate() lowercase__ : List[Any] = math.exp(eval_output['eval_loss'] ) lowercase__ : int = {'perplexity': perplexity} lowercase__ : int = os.path.join(training_args.output_dir , 'eval_results_lm.txt' ) if trainer.is_world_master(): with open(_lowerCAmelCase , 'w' ) as writer: logger.info('***** Eval results *****' ) for key in sorted(result.keys() ): logger.info(' %s = %s' , _lowerCAmelCase , str(result[key] ) ) writer.write('%s = %s\n' % (key, str(result[key] )) ) results.update(_lowerCAmelCase ) return results def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' main() if __name__ == "__main__": main()
77
1
"""simple docstring""" import os import re import shutil import sys import tempfile import unittest import black _UpperCamelCase : Union[str, Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. _UpperCamelCase : Union[str, Any] = " def __init__(self, config):\n super().__init__()\n self.transform = BertPredictionHeadTransform(config)\n\n # The output weights are the same as the input embeddings, but there is\n # an output-only bias for each token.\n self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False)\n\n self.bias = nn.Parameter(torch.zeros(config.vocab_size))\n\n # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings`\n self.decoder.bias = self.bias\n\n def forward(self, hidden_states):\n hidden_states = self.transform(hidden_states)\n hidden_states = self.decoder(hidden_states)\n return hidden_states\n" class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Any: lowercase__ : Union[str, Any] = tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , 'models/bert/' ) ) lowercase__ : List[str] = self.transformer_dir shutil.copy( os.path.join(a , 'src/transformers/models/bert/modeling_bert.py' ) , os.path.join(self.transformer_dir , 'models/bert/modeling_bert.py' ) , ) def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : int = 'src/transformers' shutil.rmtree(self.transformer_dir ) def _UpperCAmelCase ( self , a , a , a , a=None ) -> int: lowercase__ : Tuple = comment + f"""\nclass {class_name}(nn.Module):\n""" + class_code if overwrite_result is not None: lowercase__ : Any = comment + f"""\nclass {class_name}(nn.Module):\n""" + overwrite_result lowercase__ : int = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_1_9 ) lowercase__ : int = black.format_str(a , mode=a ) lowercase__ : Any = os.path.join(self.transformer_dir , 'new_code.py' ) with open(a , 'w' , newline='\n' ) as f: f.write(a ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(a ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=a ) with open(a , 'r' ) as f: self.assertTrue(f.read() , a ) def _UpperCAmelCase ( self ) -> int: lowercase__ : Optional[int] = check_copies.find_code_in_transformers('models.bert.modeling_bert.BertLMPredictionHead' ) self.assertEqual(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: # Base copy consistency self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead' , 'BertLMPredictionHead' , REFERENCE_CODE + '\n' , ) # With no empty line at the end self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead' , 'BertLMPredictionHead' , a , ) # Copy consistency with rename self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel' , 'TestModelLMPredictionHead' , re.sub('Bert' , 'TestModel' , a ) , ) # Copy consistency with a really long name lowercase__ : int = 'TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason' self.check_copy_consistency( f"""# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}""" , f"""{long_class_name}LMPredictionHead""" , re.sub('Bert' , a , a ) , ) # Copy consistency with overwrite self.check_copy_consistency( '# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel' , 'TestModelLMPredictionHead' , a , overwrite_result=re.sub('Bert' , 'TestModel' , a ) , ) def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : List[Any] = check_copies.LOCALIZED_READMES['README_zh-hans.md'] lowercase__ : Tuple = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the' ' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for' ' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong' ' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1.' ' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace),' ' released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and' ' lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same' ' method has been applied to compress GPT2 into' ' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into' ' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),' ' Multilingual BERT into' ' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German' ' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**' ' (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders' ' as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang' ' Luong, Quoc V. Le, Christopher D. Manning.' ) lowercase__ : List[Any] = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the' ' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n' ) lowercase__ : str = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the' ' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1.' ' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文' ' [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and' ' lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same' ' method has been applied to compress GPT2 into' ' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into' ' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),' ' Multilingual BERT into' ' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German' ' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自' ' Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather' ' than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le,' ' Christopher D. Manning 发布。\n' ) lowercase__ , lowercase__ : Optional[int] = check_copies.convert_to_localized_md( a , a , localized_readme['format_model_list'] ) self.assertFalse(a ) self.assertEqual(a , a ) lowercase__ , lowercase__ : Any = check_copies.convert_to_localized_md( a , a , localized_readme['format_model_list'] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(a ) lowercase__ : Dict = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the' ' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for' ' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong' ' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.' ) lowercase__ : Optional[int] = ( '1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and' ' the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n' ) lowercase__ : str = ( '1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the' ' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of' ' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian' ' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n' ) lowercase__ , lowercase__ : List[str] = check_copies.convert_to_localized_md( a , a , localized_readme['format_model_list'] ) # Check if the model link is synchronized. self.assertEqual(a , a )
77
"""simple docstring""" import math import flax.linen as nn import jax.numpy as jnp def a_ ( _lowerCAmelCase : jnp.ndarray , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1.0E4 , _lowerCAmelCase : bool = False , _lowerCAmelCase : float = 1.0 , ): '''simple docstring''' assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, f"""Embedding dimension {embedding_dim} should be even""" lowercase__ : Optional[Any] = float(embedding_dim // 2 ) lowercase__ : Optional[Any] = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) lowercase__ : Any = min_timescale * jnp.exp(jnp.arange(_lowerCAmelCase , dtype=jnp.floataa ) * -log_timescale_increment ) lowercase__ : Dict = jnp.expand_dims(_lowerCAmelCase , 1 ) * jnp.expand_dims(_lowerCAmelCase , 0 ) # scale embeddings lowercase__ : List[str] = scale * emb if flip_sin_to_cos: lowercase__ : Dict = jnp.concatenate([jnp.cos(_lowerCAmelCase ), jnp.sin(_lowerCAmelCase )] , axis=1 ) else: lowercase__ : Optional[int] = jnp.concatenate([jnp.sin(_lowerCAmelCase ), jnp.cos(_lowerCAmelCase )] , axis=1 ) lowercase__ : List[Any] = jnp.reshape(_lowerCAmelCase , [jnp.shape(_lowerCAmelCase )[0], embedding_dim] ) return signal class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : jnp.dtype = jnp.floataa @nn.compact def __call__( self , a ) -> Any: lowercase__ : str = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_1' )(a ) lowercase__ : Union[str, Any] = nn.silu(a ) lowercase__ : Optional[Any] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_2' )(a ) return temb class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : bool = False lowerCamelCase__ : float = 1 @nn.compact def __call__( self , a ) -> str: return get_sinusoidal_embeddings( a , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
77
1
"""simple docstring""" import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : Tuple = { "google/owlvit-base-patch32": "https://huggingface.co/google/owlvit-base-patch32/resolve/main/config.json", "google/owlvit-base-patch16": "https://huggingface.co/google/owlvit-base-patch16/resolve/main/config.json", "google/owlvit-large-patch14": "https://huggingface.co/google/owlvit-large-patch14/resolve/main/config.json", } class UpperCAmelCase_ ( _a): lowerCamelCase__ : Any = "owlvit_text_model" def __init__( self , a=4_9_4_0_8 , a=5_1_2 , a=2_0_4_8 , a=1_2 , a=8 , a=1_6 , a="quick_gelu" , a=1e-5 , a=0.0 , a=0.02 , a=1.0 , a=0 , a=4_9_4_0_6 , a=4_9_4_0_7 , **a , ) -> List[str]: super().__init__(pad_token_id=a , bos_token_id=a , eos_token_id=a , **a ) lowercase__ : List[Any] = vocab_size lowercase__ : Union[str, Any] = hidden_size lowercase__ : str = intermediate_size lowercase__ : Optional[int] = num_hidden_layers lowercase__ : Optional[int] = num_attention_heads lowercase__ : Optional[int] = max_position_embeddings lowercase__ : Tuple = hidden_act lowercase__ : Dict = layer_norm_eps lowercase__ : List[str] = attention_dropout lowercase__ : Any = initializer_range lowercase__ : Optional[int] = initializer_factor @classmethod def _UpperCAmelCase ( cls , a , **a ) -> "PretrainedConfig": cls._set_token_in_kwargs(a ) lowercase__ , lowercase__ : int = cls.get_config_dict(a , **a ) # get the text config dict if we are loading from OwlViTConfig if config_dict.get('model_type' ) == "owlvit": lowercase__ : Union[str, Any] = config_dict['text_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(a , **a ) class UpperCAmelCase_ ( _a): lowerCamelCase__ : List[str] = "owlvit_vision_model" def __init__( self , a=7_6_8 , a=3_0_7_2 , a=1_2 , a=1_2 , a=3 , a=7_6_8 , a=3_2 , a="quick_gelu" , a=1e-5 , a=0.0 , a=0.02 , a=1.0 , **a , ) -> Dict: super().__init__(**a ) lowercase__ : List[str] = hidden_size lowercase__ : Optional[int] = intermediate_size lowercase__ : Optional[Any] = num_hidden_layers lowercase__ : str = num_attention_heads lowercase__ : Optional[Any] = num_channels lowercase__ : Any = image_size lowercase__ : Optional[int] = patch_size lowercase__ : List[Any] = hidden_act lowercase__ : str = layer_norm_eps lowercase__ : Dict = attention_dropout lowercase__ : Optional[Any] = initializer_range lowercase__ : str = initializer_factor @classmethod def _UpperCAmelCase ( cls , a , **a ) -> "PretrainedConfig": cls._set_token_in_kwargs(a ) lowercase__ , lowercase__ : List[Any] = cls.get_config_dict(a , **a ) # get the vision config dict if we are loading from OwlViTConfig if config_dict.get('model_type' ) == "owlvit": lowercase__ : Dict = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(a , **a ) class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = "owlvit" lowerCamelCase__ : Optional[Any] = True def __init__( self , a=None , a=None , a=5_1_2 , a=2.6_592 , a=True , **a , ) -> List[Any]: super().__init__(**a ) if text_config is None: lowercase__ : Dict = {} logger.info('text_config is None. Initializing the OwlViTTextConfig with default values.' ) if vision_config is None: lowercase__ : int = {} logger.info('vision_config is None. initializing the OwlViTVisionConfig with default values.' ) lowercase__ : Optional[Any] = OwlViTTextConfig(**a ) lowercase__ : int = OwlViTVisionConfig(**a ) lowercase__ : Optional[int] = projection_dim lowercase__ : Union[str, Any] = logit_scale_init_value lowercase__ : Optional[Any] = return_dict lowercase__ : Dict = 1.0 @classmethod def _UpperCAmelCase ( cls , a , **a ) -> "PretrainedConfig": cls._set_token_in_kwargs(a ) lowercase__ , lowercase__ : str = cls.get_config_dict(a , **a ) if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(a , **a ) @classmethod def _UpperCAmelCase ( cls , a , a , **a ) -> Union[str, Any]: lowercase__ : List[str] = {} lowercase__ : str = text_config lowercase__ : Any = vision_config return cls.from_dict(a , **a ) def _UpperCAmelCase ( self ) -> Any: lowercase__ : Union[str, Any] = copy.deepcopy(self.__dict__ ) lowercase__ : Union[str, Any] = self.text_config.to_dict() lowercase__ : Tuple = self.vision_config.to_dict() lowercase__ : Union[str, Any] = self.__class__.model_type return output class UpperCAmelCase_ ( _a): @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ('attention_mask', {0: 'batch', 1: 'sequence'}), ] ) @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('logits_per_image', {0: 'batch'}), ('logits_per_text', {0: 'batch'}), ('text_embeds', {0: 'batch'}), ('image_embeds', {0: 'batch'}), ] ) @property def _UpperCAmelCase ( self ) -> float: return 1e-4 def _UpperCAmelCase ( self , a , a = -1 , a = -1 , a = None , ) -> Mapping[str, Any]: lowercase__ : Optional[Any] = super().generate_dummy_inputs( processor.tokenizer , batch_size=a , seq_length=a , framework=a ) lowercase__ : str = super().generate_dummy_inputs( processor.image_processor , batch_size=a , framework=a ) return {**text_input_dict, **image_input_dict} @property def _UpperCAmelCase ( self ) -> int: return 1_4
77
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def a_ ( _lowerCAmelCase : Callable[[int | float], int | float] , _lowerCAmelCase : int | float , _lowerCAmelCase : int | float , _lowerCAmelCase : int = 100 , ): '''simple docstring''' lowercase__ : Dict = x_start lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) lowercase__ : Optional[Any] = 0.0 for _ in range(_lowerCAmelCase ): # Approximates curve as a sequence of linear lines and sums their length lowercase__ : Union[str, Any] = (x_end - x_start) / steps + xa lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step lowercase__ : Union[str, Any] = xa lowercase__ : int = fxa return length if __name__ == "__main__": def a_ ( _lowerCAmelCase : List[Any] ): '''simple docstring''' return math.sin(10 * x ) print("f(x) = sin(10 * x)") print("The length of the curve from x = -10 to x = 10 is:") _UpperCamelCase : str = 10 while i <= 10_00_00: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
77
1
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _UpperCamelCase : Optional[Any] = {"configuration_mra": ["MRA_PRETRAINED_CONFIG_ARCHIVE_MAP", "MraConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Any = [ "MRA_PRETRAINED_MODEL_ARCHIVE_LIST", "MraForMaskedLM", "MraForMultipleChoice", "MraForQuestionAnswering", "MraForSequenceClassification", "MraForTokenClassification", "MraLayer", "MraModel", "MraPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys _UpperCamelCase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure)
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase : Tuple = { "configuration_whisper": ["WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP", "WhisperConfig", "WhisperOnnxConfig"], "feature_extraction_whisper": ["WhisperFeatureExtractor"], "processing_whisper": ["WhisperProcessor"], "tokenization_whisper": ["WhisperTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = ["WhisperTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : str = [ "WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "WhisperForConditionalGeneration", "WhisperModel", "WhisperPreTrainedModel", "WhisperForAudioClassification", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Any = [ "TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFWhisperForConditionalGeneration", "TFWhisperModel", "TFWhisperPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Tuple = [ "FlaxWhisperForConditionalGeneration", "FlaxWhisperModel", "FlaxWhisperPreTrainedModel", "FlaxWhisperForAudioClassification", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _UpperCamelCase : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () _UpperCamelCase : Optional[Any] = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). _UpperCamelCase : Any = [0, 25, 50] _UpperCamelCase : List[Any] = [25, 50, 75] _UpperCamelCase : str = fuzz.membership.trimf(X, abca) _UpperCamelCase : str = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. _UpperCamelCase : List[str] = np.ones(75) _UpperCamelCase : Optional[Any] = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) _UpperCamelCase : Dict = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) _UpperCamelCase : Dict = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) _UpperCamelCase : Tuple = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) _UpperCamelCase : int = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] _UpperCamelCase : Union[str, Any] = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) _UpperCamelCase : Optional[Any] = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] _UpperCamelCase : int = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] _UpperCamelCase : Optional[int] = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title("Young") plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title("Middle aged") plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title("union") plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title("intersection") plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title("complement_a") plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title("difference a/b") plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title("alg_sum") plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title("alg_product") plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title("bdd_sum") plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title("bdd_difference") plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
77
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a_ ( _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : int = args.pruning_method lowercase__ : Tuple = args.threshold lowercase__ : str = args.model_name_or_path.rstrip('/' ) lowercase__ : List[Any] = args.target_model_path print(f"""Load fine-pruned model from {model_name_or_path}""" ) lowercase__ : Optional[Any] = torch.load(os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) lowercase__ : List[str] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowercase__ : Tuple = tensor print(f"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowercase__ : List[str] = tensor print(f"""Copied layer {name}""" ) elif "bias" in name: lowercase__ : Optional[Any] = tensor print(f"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowercase__ : Optional[Any] = MagnitudeBinarizer.apply(inputs=_lowerCAmelCase , threshold=_lowerCAmelCase ) lowercase__ : Optional[int] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowercase__ : Optional[Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ : Any = TopKBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[Any] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowercase__ : Any = name[:-6] lowercase__ : Optional[Any] = model[f"""{prefix_}mask_scores"""] lowercase__ : Tuple = ThresholdBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[str] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowercase__ : Union[str, Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ , lowercase__ : Tuple = -0.1, 1.1 lowercase__ : Optional[Any] = torch.sigmoid(_lowerCAmelCase ) lowercase__ : Optional[Any] = s * (r - l) + l lowercase__ : Optional[Any] = s_bar.clamp(min=0.0 , max=1.0 ) lowercase__ : Union[str, Any] = tensor * mask print(f"""Pruned layer {name}""" ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: lowercase__ : Union[str, Any] = os.path.join( os.path.dirname(_lowerCAmelCase ) , f"""bertarized_{os.path.basename(_lowerCAmelCase )}""" ) if not os.path.isdir(_lowerCAmelCase ): shutil.copytree(_lowerCAmelCase , _lowerCAmelCase ) print(f"""\nCreated folder {target_model_path}""" ) torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": _UpperCamelCase : int = argparse.ArgumentParser() parser.add_argument( "--pruning_method", choices=["l0", "magnitude", "topK", "sigmoied_threshold"], type=str, required=True, help=( "Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning," " sigmoied_threshold = Soft movement pruning)" ), ) parser.add_argument( "--threshold", type=float, required=False, help=( "For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model." "For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared." "Not needed for `l0`" ), ) parser.add_argument( "--model_name_or_path", type=str, required=True, help="Folder containing the model that was previously fine-pruned", ) parser.add_argument( "--target_model_path", default=None, type=str, required=False, help="Folder containing the model that was previously fine-pruned", ) _UpperCamelCase : Dict = parser.parse_args() main(args)
77
1
"""simple docstring""" import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version(">=", FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpoint.default_planner import DefaultLoadPlanner, DefaultSavePlanner from torch.distributed.checkpoint.optimizer import load_sharded_optimizer_state_dict from torch.distributed.fsdp.fully_sharded_data_parallel import FullyShardedDataParallel as FSDP from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType _UpperCamelCase : Dict = get_logger(__name__) def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : List[str]=0 ): '''simple docstring''' os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) with FSDP.state_dict_type( _lowerCAmelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): lowercase__ : Any = model.state_dict() if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: lowercase__ : Tuple = f"""{MODEL_NAME}.bin""" if model_index == 0 else f"""{MODEL_NAME}_{model_index}.bin""" lowercase__ : Dict = os.path.join(_lowerCAmelCase , _lowerCAmelCase ) if accelerator.process_index == 0: logger.info(f"""Saving model to {output_model_file}""" ) torch.save(_lowerCAmelCase , _lowerCAmelCase ) logger.info(f"""Model saved to {output_model_file}""" ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: lowercase__ : Any = ( f"""{MODEL_NAME}_rank{accelerator.process_index}.bin""" if model_index == 0 else f"""{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin""" ) lowercase__ : Tuple = os.path.join(_lowerCAmelCase , _lowerCAmelCase ) logger.info(f"""Saving model to {output_model_file}""" ) torch.save(_lowerCAmelCase , _lowerCAmelCase ) logger.info(f"""Model saved to {output_model_file}""" ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: lowercase__ : int = os.path.join(_lowerCAmelCase , f"""{MODEL_NAME}_{model_index}""" ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) logger.info(f"""Saving model to {ckpt_dir}""" ) lowercase__ : str = {'model': state_dict} dist_cp.save_state_dict( state_dict=_lowerCAmelCase , storage_writer=dist_cp.FileSystemWriter(_lowerCAmelCase ) , planner=DefaultSavePlanner() , ) logger.info(f"""Model saved to {ckpt_dir}""" ) def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : str=0 ): '''simple docstring''' accelerator.wait_for_everyone() with FSDP.state_dict_type( _lowerCAmelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if type(_lowerCAmelCase ) != FSDP and accelerator.process_index != 0: if not fsdp_plugin.sync_module_states: raise ValueError( 'Set the `sync_module_states` flag to `True` so that model states are synced across processes when ' 'initializing FSDP object' ) return lowercase__ : Dict = f"""{MODEL_NAME}.bin""" if model_index == 0 else f"""{MODEL_NAME}_{model_index}.bin""" lowercase__ : Dict = os.path.join(_lowerCAmelCase , _lowerCAmelCase ) logger.info(f"""Loading model from {input_model_file}""" ) lowercase__ : Dict = torch.load(_lowerCAmelCase ) logger.info(f"""Model loaded from {input_model_file}""" ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: lowercase__ : int = ( f"""{MODEL_NAME}_rank{accelerator.process_index}.bin""" if model_index == 0 else f"""{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin""" ) lowercase__ : Union[str, Any] = os.path.join(_lowerCAmelCase , _lowerCAmelCase ) logger.info(f"""Loading model from {input_model_file}""" ) lowercase__ : List[Any] = torch.load(_lowerCAmelCase ) logger.info(f"""Model loaded from {input_model_file}""" ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: lowercase__ : str = ( os.path.join(_lowerCAmelCase , f"""{MODEL_NAME}_{model_index}""" ) if f"""{MODEL_NAME}""" not in input_dir else input_dir ) logger.info(f"""Loading model from {ckpt_dir}""" ) lowercase__ : int = {'model': model.state_dict()} dist_cp.load_state_dict( state_dict=_lowerCAmelCase , storage_reader=dist_cp.FileSystemReader(_lowerCAmelCase ) , planner=DefaultLoadPlanner() , ) lowercase__ : List[Any] = state_dict['model'] logger.info(f"""Model loaded from {ckpt_dir}""" ) model.load_state_dict(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any]=0 ): '''simple docstring''' os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) with FSDP.state_dict_type( _lowerCAmelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): lowercase__ : List[str] = FSDP.optim_state_dict(_lowerCAmelCase , _lowerCAmelCase ) if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if accelerator.process_index == 0: lowercase__ : List[str] = ( f"""{OPTIMIZER_NAME}.bin""" if optimizer_index == 0 else f"""{OPTIMIZER_NAME}_{optimizer_index}.bin""" ) lowercase__ : Any = os.path.join(_lowerCAmelCase , _lowerCAmelCase ) logger.info(f"""Saving Optimizer state to {output_optimizer_file}""" ) torch.save(_lowerCAmelCase , _lowerCAmelCase ) logger.info(f"""Optimizer state saved in {output_optimizer_file}""" ) else: lowercase__ : List[Any] = os.path.join(_lowerCAmelCase , f"""{OPTIMIZER_NAME}_{optimizer_index}""" ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) logger.info(f"""Saving Optimizer state to {ckpt_dir}""" ) dist_cp.save_state_dict( state_dict={'optimizer': optim_state} , storage_writer=dist_cp.FileSystemWriter(_lowerCAmelCase ) , planner=DefaultSavePlanner() , ) logger.info(f"""Optimizer state saved in {ckpt_dir}""" ) def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : int , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any]=0 ): '''simple docstring''' accelerator.wait_for_everyone() with FSDP.state_dict_type( _lowerCAmelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: lowercase__ : Any = None # below check should work but currently it isn't working (mostly opytorch issue), # in the meantime disabling it at the cost of excess memory usage # if accelerator.process_index == 0 or not fsdp_plugin.optim_state_dict_config.rank0_only: lowercase__ : Union[str, Any] = ( f"""{OPTIMIZER_NAME}.bin""" if optimizer_index == 0 else f"""{OPTIMIZER_NAME}_{optimizer_index}.bin""" ) lowercase__ : Optional[int] = os.path.join(_lowerCAmelCase , _lowerCAmelCase ) logger.info(f"""Loading Optimizer state from {input_optimizer_file}""" ) lowercase__ : str = torch.load(_lowerCAmelCase ) logger.info(f"""Optimizer state loaded from {input_optimizer_file}""" ) else: lowercase__ : Any = ( os.path.join(_lowerCAmelCase , f"""{OPTIMIZER_NAME}_{optimizer_index}""" ) if f"""{OPTIMIZER_NAME}""" not in input_dir else input_dir ) logger.info(f"""Loading Optimizer from {ckpt_dir}""" ) lowercase__ : Optional[Any] = load_sharded_optimizer_state_dict( model_state_dict=model.state_dict() , optimizer_key='optimizer' , storage_reader=dist_cp.FileSystemReader(_lowerCAmelCase ) , ) lowercase__ : str = optim_state['optimizer'] logger.info(f"""Optimizer loaded from {ckpt_dir}""" ) lowercase__ : str = FSDP.optim_state_dict_to_load(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) optimizer.load_state_dict(_lowerCAmelCase )
77
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[List[PIL.Image.Image], np.ndarray] lowerCamelCase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : np.ndarray lowerCamelCase__ : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int ): '''simple docstring''' if divisor % 5 == 0 or divisor % 2 == 0: return 0 lowercase__ : Dict = 1 lowercase__ : Dict = 1 while repunit: lowercase__ : List[Any] = (10 * repunit + 1) % divisor repunit_index += 1 return repunit_index def a_ ( _lowerCAmelCase : int = 100_0000 ): '''simple docstring''' lowercase__ : str = limit - 1 if divisor % 2 == 0: divisor += 1 while least_divisible_repunit(_lowerCAmelCase ) <= limit: divisor += 2 return divisor if __name__ == "__main__": print(f'''{solution() = }''')
77
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class UpperCAmelCase_ : def __init__( self , a , a , a , a , a , a=0.2 , a=0.2 ) -> Dict: lowercase__ : Any = bp_numa lowercase__ : Optional[int] = bp_numa lowercase__ : Tuple = bp_numa lowercase__ : Optional[Any] = conva_get[:2] lowercase__ : Optional[int] = conva_get[2] lowercase__ : Optional[Any] = size_pa lowercase__ : Union[str, Any] = rate_w lowercase__ : Union[str, Any] = rate_t lowercase__ : List[Any] = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ : Optional[Any] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : int = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : Any = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 def _UpperCAmelCase ( self , a ) -> Union[str, Any]: # save model dict with pickle lowercase__ : Optional[Any] = { 'num_bp1': self.num_bpa, 'num_bp2': self.num_bpa, 'num_bp3': self.num_bpa, 'conv1': self.conva, 'step_conv1': self.step_conva, 'size_pooling1': self.size_poolinga, 'rate_weight': self.rate_weight, 'rate_thre': self.rate_thre, 'w_conv1': self.w_conva, 'wkj': self.wkj, 'vji': self.vji, 'thre_conv1': self.thre_conva, 'thre_bp2': self.thre_bpa, 'thre_bp3': self.thre_bpa, } with open(a , 'wb' ) as f: pickle.dump(a , a ) print(f"""Model saved: {save_path}""" ) @classmethod def _UpperCAmelCase ( cls , a ) -> Any: # read saved model with open(a , 'rb' ) as f: lowercase__ : Optional[int] = pickle.load(a ) # noqa: S301 lowercase__ : Optional[int] = model_dic.get('conv1' ) conv_get.append(model_dic.get('step_conv1' ) ) lowercase__ : List[Any] = model_dic.get('size_pooling1' ) lowercase__ : Tuple = model_dic.get('num_bp1' ) lowercase__ : int = model_dic.get('num_bp2' ) lowercase__ : int = model_dic.get('num_bp3' ) lowercase__ : Union[str, Any] = model_dic.get('rate_weight' ) lowercase__ : Tuple = model_dic.get('rate_thre' ) # create model instance lowercase__ : Tuple = CNN(a , a , a , a , a , a , a ) # modify model parameter lowercase__ : str = model_dic.get('w_conv1' ) lowercase__ : Optional[int] = model_dic.get('wkj' ) lowercase__ : Tuple = model_dic.get('vji' ) lowercase__ : str = model_dic.get('thre_conv1' ) lowercase__ : Union[str, Any] = model_dic.get('thre_bp2' ) lowercase__ : List[str] = model_dic.get('thre_bp3' ) return conv_ins def _UpperCAmelCase ( self , a ) -> str: return 1 / (1 + np.exp(-1 * x )) def _UpperCAmelCase ( self , a ) -> Any: return round(a , 3 ) def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[str]: # convolution process lowercase__ : int = convs[0] lowercase__ : Optional[Any] = convs[1] lowercase__ : int = np.shape(a )[0] # get the data slice of original image data, data_focus lowercase__ : Optional[Any] = [] for i_focus in range(0 , size_data - size_conv + 1 , a ): for j_focus in range(0 , size_data - size_conv + 1 , a ): lowercase__ : Optional[int] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(a ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ : Union[str, Any] = [] lowercase__ : Dict = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(a ): lowercase__ : Any = [] for i_focus in range(len(a ) ): lowercase__ : Tuple = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(a ) ) lowercase__ : Optional[Any] = np.asmatrix(a ).reshape( a , a ) data_featuremap.append(a ) # expanding the data slice to One dimenssion lowercase__ : str = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(a ) ) lowercase__ : int = np.asarray(a ) return focus_list, data_featuremap def _UpperCAmelCase ( self , a , a , a="average_pool" ) -> str: # pooling process lowercase__ : List[str] = len(featuremaps[0] ) lowercase__ : List[str] = int(size_map / size_pooling ) lowercase__ : str = [] for i_map in range(len(a ) ): lowercase__ : List[str] = featuremaps[i_map] lowercase__ : Optional[int] = [] for i_focus in range(0 , a , a ): for j_focus in range(0 , a , a ): lowercase__ : List[Any] = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(a ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(a ) ) lowercase__ : List[Any] = np.asmatrix(a ).reshape(a , a ) featuremap_pooled.append(a ) return featuremap_pooled def _UpperCAmelCase ( self , a ) -> List[str]: # expanding three dimension data to one dimension list lowercase__ : Any = [] for i in range(len(a ) ): lowercase__ : Optional[int] = np.shape(data[i] ) lowercase__ : int = data[i].reshape(1 , shapes[0] * shapes[1] ) lowercase__ : str = data_listed.getA().tolist()[0] data_expanded.extend(a ) lowercase__ : int = np.asarray(a ) return data_expanded def _UpperCAmelCase ( self , a ) -> Dict: # expanding matrix to one dimension list lowercase__ : Dict = np.asarray(a ) lowercase__ : Union[str, Any] = np.shape(a ) lowercase__ : Optional[Any] = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[Any]: lowercase__ : Dict = [] lowercase__ : int = 0 for i_map in range(a ): lowercase__ : str = np.ones((size_map, size_map) ) for i in range(0 , a , a ): for j in range(0 , a , a ): lowercase__ : Optional[Any] = pd_pool[ i_pool ] lowercase__ : Union[str, Any] = i_pool + 1 lowercase__ : List[Any] = np.multiply( a , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(a ) return pd_all def _UpperCAmelCase ( self , a , a , a , a , a , a=bool ) -> str: # model traning print('----------------------Start Training-------------------------' ) print((' - - Shape: Train_Data ', np.shape(a )) ) print((' - - Shape: Teach_Data ', np.shape(a )) ) lowercase__ : int = 0 lowercase__ : List[Any] = [] lowercase__ : Union[str, Any] = 1_0_0_0_0 while rp < n_repeat and mse >= error_accuracy: lowercase__ : List[Any] = 0 print(f"""-------------Learning Time {rp}--------------""" ) for p in range(len(a ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ : Optional[int] = np.asmatrix(datas_train[p] ) lowercase__ : int = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ : Union[str, Any] = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Optional[Any] = self.pooling(a , self.size_poolinga ) lowercase__ : Tuple = np.shape(a ) lowercase__ : List[str] = self._expand(a ) lowercase__ : Optional[int] = data_bp_input lowercase__ : Optional[Any] = np.dot(a , self.vji.T ) - self.thre_bpa lowercase__ : str = self.sig(a ) lowercase__ : Tuple = np.dot(a , self.wkj.T ) - self.thre_bpa lowercase__ : Any = self.sig(a ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ : int = np.multiply( (data_teach - bp_outa) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Any = np.multiply( np.dot(a , self.wkj ) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Optional[int] = np.dot(a , self.vji ) lowercase__ : Union[str, Any] = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ : Any = pd_conva_pooled.T.getA().tolist() lowercase__ : List[str] = self._calculate_gradient_from_pool( a , a , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ : Optional[int] = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ : Tuple = self.rate_weight * np.dot(a , a ) lowercase__ : Union[str, Any] = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ : Any = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ : Tuple = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ : Optional[Any] = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ : Dict = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ : str = rp + 1 lowercase__ : List[str] = error_count / patterns all_mse.append(a ) def draw_error(): lowercase__ : Any = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(a , '+-' ) plt.plot(a , 'r--' ) plt.xlabel('Learning Times' ) plt.ylabel('All_mse' ) plt.grid(a , alpha=0.5 ) plt.show() print('------------------Training Complished---------------------' ) print((' - - Training epoch: ', rp, f""" - - Mse: {mse:.6f}""") ) if draw_e: draw_error() return mse def _UpperCAmelCase ( self , a ) -> List[Any]: # model predict lowercase__ : Optional[int] = [] print('-------------------Start Testing-------------------------' ) print((' - - Shape: Test_Data ', np.shape(a )) ) for p in range(len(a ) ): lowercase__ : List[str] = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ : Tuple = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Any = self.pooling(a , self.size_poolinga ) lowercase__ : Union[str, Any] = self._expand(a ) lowercase__ : Optional[Any] = data_bp_input lowercase__ : str = bp_outa * self.vji.T - self.thre_bpa lowercase__ : Optional[Any] = self.sig(a ) lowercase__ : Dict = bp_outa * self.wkj.T - self.thre_bpa lowercase__ : List[str] = self.sig(a ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ : Optional[int] = [list(map(self.do_round , a ) ) for each in produce_out] return np.asarray(a ) def _UpperCAmelCase ( self , a ) -> List[str]: # return the data of image after convoluting process so we can check it out lowercase__ : Any = np.asmatrix(a ) lowercase__ , lowercase__ : str = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Tuple = self.pooling(a , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
77
1
"""simple docstring""" import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class UpperCAmelCase_ : def __init__( self , a , a=1_3 , a=7 , a=True , a=True , a=False , a=True , a=9_9 , a=3_2 , a=5 , a=4 , a=3_7 , a="gelu" , a=0.1 , a=0.1 , a=5_1_2 , a=1_6 , a=2 , a=0.02 , a=3 , a=4 , a=None , ) -> Union[str, Any]: lowercase__ : Any = parent lowercase__ : List[str] = batch_size lowercase__ : str = seq_length lowercase__ : Optional[Any] = is_training lowercase__ : Tuple = use_input_mask lowercase__ : List[Any] = use_token_type_ids lowercase__ : Dict = use_labels lowercase__ : List[str] = vocab_size lowercase__ : Tuple = hidden_size lowercase__ : Optional[Any] = num_hidden_layers lowercase__ : str = num_attention_heads lowercase__ : Union[str, Any] = intermediate_size lowercase__ : Optional[int] = hidden_act lowercase__ : Optional[Any] = hidden_dropout_prob lowercase__ : Optional[int] = attention_probs_dropout_prob lowercase__ : List[Any] = max_position_embeddings lowercase__ : Optional[Any] = type_vocab_size lowercase__ : Optional[int] = type_sequence_label_size lowercase__ : str = initializer_range lowercase__ : Tuple = num_labels lowercase__ : int = num_choices lowercase__ : Union[str, Any] = scope def _UpperCAmelCase ( self ) -> Any: lowercase__ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : List[str] = None if self.use_input_mask: lowercase__ : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__ : Dict = None if self.use_token_type_ids: lowercase__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase__ : Tuple = None lowercase__ : Tuple = None lowercase__ : Tuple = None if self.use_labels: lowercase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase__ : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase__ : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) lowercase__ : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCAmelCase ( self ) -> str: return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a , initializer_range=self.initializer_range , ) def _UpperCAmelCase ( self , a , a , a , a , a , a , a ) -> Optional[Any]: lowercase__ : int = LlamaModel(config=a ) model.to(a ) model.eval() lowercase__ : Union[str, Any] = model(a , attention_mask=a ) lowercase__ : int = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self , a , a , a , a , a , a , a , a , a , ) -> Any: lowercase__ : Union[str, Any] = True lowercase__ : Tuple = LlamaModel(a ) model.to(a ) model.eval() lowercase__ : Any = model( a , attention_mask=a , encoder_hidden_states=a , encoder_attention_mask=a , ) lowercase__ : Dict = model( a , attention_mask=a , encoder_hidden_states=a , ) lowercase__ : int = model(a , attention_mask=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self , a , a , a , a , a , a , a , a , a , ) -> Optional[Any]: lowercase__ : Dict = LlamaForCausalLM(config=a ) model.to(a ) model.eval() lowercase__ : Optional[Any] = model(a , attention_mask=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCAmelCase ( self , a , a , a , a , a , a , a , a , a , ) -> Union[str, Any]: lowercase__ : Any = True lowercase__ : List[Any] = True lowercase__ : Union[str, Any] = LlamaForCausalLM(config=a ) model.to(a ) model.eval() # first forward pass lowercase__ : Optional[Any] = model( a , attention_mask=a , encoder_hidden_states=a , encoder_attention_mask=a , use_cache=a , ) lowercase__ : int = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids lowercase__ : Optional[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowercase__ : Optional[Any] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and lowercase__ : Optional[int] = torch.cat([input_ids, next_tokens] , dim=-1 ) lowercase__ : Tuple = torch.cat([input_mask, next_mask] , dim=-1 ) lowercase__ : Tuple = model( a , attention_mask=a , encoder_hidden_states=a , encoder_attention_mask=a , output_hidden_states=a , )['hidden_states'][0] lowercase__ : Optional[int] = model( a , attention_mask=a , encoder_hidden_states=a , encoder_attention_mask=a , past_key_values=a , output_hidden_states=a , )['hidden_states'][0] # select random slice lowercase__ : int = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowercase__ : List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() lowercase__ : Optional[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a , a , atol=1e-3 ) ) def _UpperCAmelCase ( self ) -> int: lowercase__ : Dict = self.prepare_config_and_inputs() ( ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ) : Dict = config_and_inputs lowercase__ : Any = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class UpperCAmelCase_ ( _a , _a , _a , unittest.TestCase): lowerCamelCase__ : Any = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () lowerCamelCase__ : List[Any] = (LlamaForCausalLM,) if is_torch_available() else () lowerCamelCase__ : str = ( { "feature-extraction": LlamaModel, "text-classification": LlamaForSequenceClassification, "text-generation": LlamaForCausalLM, "zero-shot": LlamaForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase__ : Tuple = False lowerCamelCase__ : str = False def _UpperCAmelCase ( self ) -> Dict: lowercase__ : Tuple = LlamaModelTester(self ) lowercase__ : List[Any] = ConfigTester(self , config_class=a , hidden_size=3_7 ) def _UpperCAmelCase ( self ) -> int: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : str = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase__ : Any = type self.model_tester.create_and_check_model(*a ) def _UpperCAmelCase ( self ) -> int: lowercase__ , lowercase__ : Any = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ : Tuple = 3 lowercase__ : str = input_dict['input_ids'] lowercase__ : str = input_ids.ne(1 ).to(a ) lowercase__ : Tuple = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowercase__ : Optional[int] = LlamaForSequenceClassification(a ) model.to(a ) model.eval() lowercase__ : Tuple = model(a , attention_mask=a , labels=a ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCAmelCase ( self ) -> int: lowercase__ , lowercase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ : Tuple = 3 lowercase__ : str = 'single_label_classification' lowercase__ : int = input_dict['input_ids'] lowercase__ : List[Any] = input_ids.ne(1 ).to(a ) lowercase__ : Any = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowercase__ : Any = LlamaForSequenceClassification(a ) model.to(a ) model.eval() lowercase__ : Dict = model(a , attention_mask=a , labels=a ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ , lowercase__ : int = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ : Union[str, Any] = 3 lowercase__ : str = 'multi_label_classification' lowercase__ : List[str] = input_dict['input_ids'] lowercase__ : Optional[Any] = input_ids.ne(1 ).to(a ) lowercase__ : List[Any] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) lowercase__ : Any = LlamaForSequenceClassification(a ) model.to(a ) model.eval() lowercase__ : List[Any] = model(a , attention_mask=a , labels=a ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('LLaMA buffers include complex numbers, which breaks this test' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: pass @parameterized.expand([('linear',), ('dynamic',)] ) def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ , lowercase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ : Optional[int] = ids_tensor([1, 1_0] , config.vocab_size ) lowercase__ : Tuple = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights lowercase__ : int = LlamaModel(a ) original_model.to(a ) original_model.eval() lowercase__ : int = original_model(a ).last_hidden_state lowercase__ : Optional[int] = original_model(a ).last_hidden_state set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights lowercase__ : Union[str, Any] = {'type': scaling_type, 'factor': 10.0} lowercase__ : Optional[int] = LlamaModel(a ) scaled_model.to(a ) scaled_model.eval() lowercase__ : str = scaled_model(a ).last_hidden_state lowercase__ : Union[str, Any] = scaled_model(a ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(a , a , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(a , a , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(a , a , atol=1e-5 ) ) @require_torch class UpperCAmelCase_ ( unittest.TestCase): @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Dict = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] lowercase__ : Dict = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-7b-hf' , device_map='auto' ) lowercase__ : List[str] = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 lowercase__ : Dict = torch.tensor([[-6.6_550, -4.1_227, -4.9_859, -3.2_406, 0.8_262, -3.0_033, 1.2_964, -3.3_699]] ) torch.testing.assert_close(out.mean(-1 ) , a , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off lowercase__ : int = torch.tensor([-12.8_281, -7.4_453, -0.4_639, -8.0_625, -7.2_500, -8.0_000, -6.4_883, -7.7_695, -7.8_438, -7.0_312, -6.2_188, -7.1_328, -1.8_496, 1.9_961, -8.6_250, -6.7_227, -12.8_281, -6.9_492, -7.0_742, -7.7_852, -7.5_820, -7.9_062, -6.9_375, -7.9_805, -8.3_438, -8.1_562, -8.0_469, -7.6_250, -7.7_422, -7.3_398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] , a , atol=1e-5 , rtol=1e-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Optional[int] = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] lowercase__ : Optional[Any] = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-hf' , device_map='auto' ) lowercase__ : Any = model(torch.tensor(a ) ) # Expected mean on dim = -1 lowercase__ : Optional[int] = torch.tensor([[-2.0_622, -1.2_794, -1.1_638, -0.9_788, -1.4_603, -1.0_238, -1.7_893, -1.4_411]] ) torch.testing.assert_close(out.mean(-1 ) , a , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off lowercase__ : Tuple = torch.tensor([-8.1_406, -8.0_547, 2.7_461, -1.2_344, -0.1_448, -1.8_262, -1.0_020, -1.8_154, -1.6_895, -1.8_516, -2.3_574, -0.9_277, 3.7_598, 6.5_742, -1.2_998, -0.1_177, -8.1_406, -2.9_688, -2.9_199, -3.1_699, -3.5_254, -2.3_555, -2.7_988, -3.4_141, -2.8_262, -4.5_195, -3.3_379, -3.3_164, -2.7_832, -3.0_273] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] , a , atol=1e-5 , rtol=1e-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def _UpperCAmelCase ( self ) -> int: lowercase__ : Optional[Any] = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] lowercase__ : Union[str, Any] = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-chat-hf' , device_map='auto' ) lowercase__ : int = model(torch.tensor(a ) ) # Expected mean on dim = -1 lowercase__ : Tuple = torch.tensor([[-0.8_562, -1.8_520, -0.7_551, -0.4_162, -1.5_161, -1.2_038, -2.4_823, -2.3_254]] ) torch.testing.assert_close(out.mean(-1 ) , a , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off lowercase__ : int = torch.tensor([-2.2_227, 4.8_828, 0.9_023, -0.4_578, -0.7_871, -0.1_033, -0.6_221, -0.5_786, -0.7_803, -1.0_674, -1.2_920, -0.1_570, 0.8_008, 2.0_723, -0.9_497, 0.2_771, -2.2_227, -0.7_612, -1.4_346, -1.2_061, -1.6_426, -0.3_000, -0.7_139, -1.1_934, -1.8_691, -1.6_973, -1.5_947, -1.2_705, -0.3_523, -0.5_513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , a , atol=1e-2 , rtol=1e-2 ) @unittest.skip( 'Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test' ) @slow def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : str = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] lowercase__ : Optional[Any] = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-70b-hf' , device_map='auto' ) lowercase__ : Optional[int] = model(torch.tensor(a ) ) lowercase__ : int = torch.tensor( [[-4.2_327, -3.3_360, -4.6_665, -4.7_631, -1.8_180, -3.4_170, -1.4_211, -3.1_810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , a , atol=1e-2 , rtol=1e-2 ) # fmt: off lowercase__ : Optional[Any] = torch.tensor([-9.4_922, -3.9_551, 1.7_998, -5.6_758, -5.1_055, -5.8_984, -4.8_320, -6.8_086, -6.5_391, -5.6_172, -5.5_820, -5.5_352, 1.7_881, 3.6_289, -6.5_117, -3.4_785, -9.5_000, -6.0_352, -6.8_125, -6.0_195, -6.6_836, -5.4_727, -6.2_812, -6.0_391, -7.3_398, -7.4_297, -7.4_844, -6.5_820, -5.8_789, -5.5_312] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] , a , atol=1e-5 , rtol=1e-5 ) @unittest.skip('Model is curently gated' ) @slow def _UpperCAmelCase ( self ) -> Any: lowercase__ : str = 'Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi' lowercase__ : Optional[int] = 'Simply put, the theory of relativity states that ' lowercase__ : List[Any] = LlamaTokenizer.from_pretrained('meta-llama/Llama-2-13b-chat-hf' ) lowercase__ : Optional[int] = tokenizer.encode(a , return_tensors='pt' ) lowercase__ : int = LlamaForCausalLM.from_pretrained( 'meta-llama/Llama-2-13b-chat-hf' , device_map='sequential' , use_safetensors=a ) # greedy generation outputs lowercase__ : str = model.generate(a , max_new_tokens=6_4 , top_p=a , temperature=1 , do_sample=a ) lowercase__ : Optional[Any] = tokenizer.decode(generated_ids[0] , skip_special_tokens=a ) self.assertEqual(a , a )
77
"""simple docstring""" from collections.abc import Generator def a_ ( ): '''simple docstring''' lowercase__ , lowercase__ : List[str] = 0, 1 while True: lowercase__ , lowercase__ : Optional[int] = b, a + b yield b def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' lowercase__ : List[Any] = 1 lowercase__ : Any = fibonacci_generator() while len(str(next(_lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
77
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCamelCase : str = logging.get_logger(__name__) _UpperCamelCase : Tuple = { "facebook/data2vec-vision-base-ft": ( "https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json" ), } class UpperCAmelCase_ ( _a): lowerCamelCase__ : int = "data2vec-vision" def __init__( self , a=7_6_8 , a=1_2 , a=1_2 , a=3_0_7_2 , a="gelu" , a=0.0 , a=0.0 , a=0.02 , a=1e-12 , a=2_2_4 , a=1_6 , a=3 , a=False , a=False , a=False , a=False , a=0.1 , a=0.1 , a=True , a=[3, 5, 7, 1_1] , a=[1, 2, 3, 6] , a=True , a=0.4 , a=2_5_6 , a=1 , a=False , a=2_5_5 , **a , ) -> Optional[int]: super().__init__(**a ) lowercase__ : Any = hidden_size lowercase__ : Tuple = num_hidden_layers lowercase__ : List[str] = num_attention_heads lowercase__ : Tuple = intermediate_size lowercase__ : Dict = hidden_act lowercase__ : Any = hidden_dropout_prob lowercase__ : Optional[Any] = attention_probs_dropout_prob lowercase__ : Tuple = initializer_range lowercase__ : Tuple = layer_norm_eps lowercase__ : Union[str, Any] = image_size lowercase__ : str = patch_size lowercase__ : Optional[int] = num_channels lowercase__ : Dict = use_mask_token lowercase__ : List[str] = use_absolute_position_embeddings lowercase__ : List[Any] = use_relative_position_bias lowercase__ : List[str] = use_shared_relative_position_bias lowercase__ : Tuple = layer_scale_init_value lowercase__ : int = drop_path_rate lowercase__ : List[str] = use_mean_pooling # decode head attributes (semantic segmentation) lowercase__ : Tuple = out_indices lowercase__ : Dict = pool_scales # auxiliary head attributes (semantic segmentation) lowercase__ : Optional[int] = use_auxiliary_head lowercase__ : int = auxiliary_loss_weight lowercase__ : List[str] = auxiliary_channels lowercase__ : str = auxiliary_num_convs lowercase__ : int = auxiliary_concat_input lowercase__ : Dict = semantic_loss_ignore_index class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = version.parse("1.11") @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def _UpperCAmelCase ( self ) -> float: return 1e-4
77
"""simple docstring""" import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class UpperCAmelCase_ : def __init__( self , a ) -> List[str]: if isinstance(a , a ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden lowercase__ : Optional[Any] = deepcopy(a ) elif os.path.exists(a ): with io.open(a , 'r' , encoding='utf-8' ) as f: lowercase__ : List[Any] = json.load(a ) else: try: lowercase__ : Optional[int] = baseaa.urlsafe_baadecode(a ).decode('utf-8' ) lowercase__ : List[str] = json.loads(a ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""" ) lowercase__ : Any = config self.set_stage_and_offload() def _UpperCAmelCase ( self ) -> Dict: # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. lowercase__ : Tuple = self.get_value('zero_optimization.stage' , -1 ) # offload lowercase__ : int = False if self.is_zeroa() or self.is_zeroa(): lowercase__ : str = set(['cpu', 'nvme'] ) lowercase__ : Optional[Any] = set( [ self.get_value('zero_optimization.offload_optimizer.device' ), self.get_value('zero_optimization.offload_param.device' ), ] ) if len(offload_devices & offload_devices_valid ) > 0: lowercase__ : Optional[Any] = True def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Dict = self.config # find the config node of interest if it exists lowercase__ : int = ds_key_long.split('.' ) lowercase__ : Dict = nodes.pop() for node in nodes: lowercase__ : Optional[Any] = config.get(a ) if config is None: return None, ds_key return config, ds_key def _UpperCAmelCase ( self , a , a=None ) -> Union[str, Any]: lowercase__ , lowercase__ : Tuple = self.find_config_node(a ) if config is None: return default return config.get(a , a ) def _UpperCAmelCase ( self , a , a=False ) -> Any: lowercase__ : str = self.config # find the config node of interest if it exists lowercase__ : List[Any] = ds_key_long.split('.' ) for node in nodes: lowercase__ : str = config lowercase__ : str = config.get(a ) if config is None: if must_exist: raise ValueError(f"""Can't find {ds_key_long} entry in the config: {self.config}""" ) else: return # if found remove it if parent_config is not None: parent_config.pop(a ) def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : Union[str, Any] = self.get_value(a ) return False if value is None else bool(a ) def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Any = self.get_value(a ) return False if value is None else not bool(a ) def _UpperCAmelCase ( self ) -> Tuple: return self._stage == 2 def _UpperCAmelCase ( self ) -> List[Any]: return self._stage == 3 def _UpperCAmelCase ( self ) -> str: return self._offload class UpperCAmelCase_ : def __init__( self , a ) -> str: lowercase__ : Tuple = engine def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: # runs backpropagation and handles mixed precision self.engine.backward(a , **a ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class UpperCAmelCase_ ( _a): def __init__( self , a ) -> Dict: super().__init__(a , device_placement=a , scaler=a ) lowercase__ : Union[str, Any] = hasattr(self.optimizer , 'overflow' ) def _UpperCAmelCase ( self , a=None ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def _UpperCAmelCase ( self ) -> Optional[int]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def _UpperCAmelCase ( self ) -> Tuple: if self.__has_overflow__: return self.optimizer.overflow return False class UpperCAmelCase_ ( _a): def __init__( self , a , a ) -> Any: super().__init__(a , a ) def _UpperCAmelCase ( self ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class UpperCAmelCase_ : def __init__( self , a , a=0.001 , a=0 , **a ) -> Tuple: lowercase__ : List[Any] = params lowercase__ : int = lr lowercase__ : int = weight_decay lowercase__ : Union[str, Any] = kwargs class UpperCAmelCase_ : def __init__( self , a , a=None , a=0 , **a ) -> Tuple: lowercase__ : Dict = optimizer lowercase__ : List[str] = total_num_steps lowercase__ : Optional[int] = warmup_num_steps lowercase__ : List[Any] = kwargs
77
1
"""simple docstring""" from manim import * class UpperCAmelCase_ ( _a): def _UpperCAmelCase ( self ) -> int: lowercase__ : str = Rectangle(height=0.5 , width=0.5 ) lowercase__ : str = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) lowercase__ : List[str] = [mem.copy() for i in range(6 )] lowercase__ : int = [mem.copy() for i in range(6 )] lowercase__ : int = VGroup(*a ).arrange(a , buff=0 ) lowercase__ : Optional[int] = VGroup(*a ).arrange(a , buff=0 ) lowercase__ : List[str] = VGroup(a , a ).arrange(a , buff=0 ) lowercase__ : Union[str, Any] = Text('CPU' , font_size=2_4 ) lowercase__ : Union[str, Any] = Group(a , a ).arrange(a , buff=0.5 , aligned_edge=a ) cpu.move_to([-2.5, -0.5, 0] ) self.add(a ) lowercase__ : Union[str, Any] = [mem.copy() for i in range(4 )] lowercase__ : List[Any] = VGroup(*a ).arrange(a , buff=0 ) lowercase__ : Tuple = Text('GPU' , font_size=2_4 ) lowercase__ : Optional[int] = Group(a , a ).arrange(a , buff=0.5 , aligned_edge=a ) gpu.move_to([-1, -1, 0] ) self.add(a ) lowercase__ : int = [mem.copy() for i in range(6 )] lowercase__ : List[str] = VGroup(*a ).arrange(a , buff=0 ) lowercase__ : int = Text('Model' , font_size=2_4 ) lowercase__ : Optional[Any] = Group(a , a ).arrange(a , buff=0.5 , aligned_edge=a ) model.move_to([3, -1.0, 0] ) self.add(a ) lowercase__ : Dict = [] for i, rect in enumerate(a ): rect.set_stroke(a ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) lowercase__ : Dict = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(a , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=a ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=a , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=a , buff=0.0 ) self.add(a ) cpu_targs.append(a ) lowercase__ : Any = [mem.copy() for i in range(6 )] lowercase__ : Optional[Any] = VGroup(*a ).arrange(a , buff=0 ) lowercase__ : Optional[int] = Text('Loaded Checkpoint' , font_size=2_4 ) lowercase__ : Dict = Group(a , a ).arrange(a , aligned_edge=a , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) lowercase__ : str = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowercase__ : Any = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=1_8 , ) key_text.move_to([-5, 2.4, 0] ) self.add(a , a ) lowercase__ : List[Any] = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=1_8 , ) blue_text.next_to(a , DOWN * 2.4 , aligned_edge=key_text.get_left() ) lowercase__ : List[str] = MarkupText( f"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(a ) , Write(a ) ) self.play(Write(a , run_time=1 ) , Create(a , run_time=1 ) ) lowercase__ : str = [] lowercase__ : Any = [] for i, rect in enumerate(a ): lowercase__ : Optional[int] = fill.copy().set_fill(a , opacity=0.7 ) target.move_to(a ) first_animations.append(GrowFromCenter(a , run_time=1 ) ) lowercase__ : Optional[Any] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(a , run_time=1.5 ) ) self.play(*a ) self.play(*a ) self.wait()
77
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _UpperCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , *a , **a ) -> Union[str, Any]: super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _UpperCAmelCase ( self , a=None ) -> Dict: lowercase__ : Any = {} if top_k is not None: lowercase__ : List[str] = top_k return {}, {}, postprocess_params def __call__( self , a , **a ) -> Tuple: return super().__call__(a , **a ) def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : List[Any] = load_image(a ) lowercase__ : Union[str, Any] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _UpperCAmelCase ( self , a ) -> List[str]: lowercase__ : Dict = self.model(**a ) return model_outputs def _UpperCAmelCase ( self , a , a=5 ) -> Dict: if top_k > self.model.config.num_labels: lowercase__ : List[Any] = self.model.config.num_labels if self.framework == "pt": lowercase__ : Tuple = model_outputs.logits.softmax(-1 )[0] lowercase__ , lowercase__ : Optional[Any] = probs.topk(a ) elif self.framework == "tf": lowercase__ : Union[str, Any] = stable_softmax(model_outputs.logits , axis=-1 )[0] lowercase__ : str = tf.math.top_k(a , k=a ) lowercase__ , lowercase__ : Dict = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) lowercase__ : Dict = scores.tolist() lowercase__ : Dict = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig class UpperCAmelCase_ ( _a): lowerCamelCase__ : int = "bert-generation" def __init__( self , a=5_0_3_5_8 , a=1_0_2_4 , a=2_4 , a=1_6 , a=4_0_9_6 , a="gelu" , a=0.1 , a=0.1 , a=5_1_2 , a=0.02 , a=1e-12 , a=0 , a=2 , a=1 , a="absolute" , a=True , **a , ) -> Optional[int]: super().__init__(pad_token_id=a , bos_token_id=a , eos_token_id=a , **a ) lowercase__ : List[Any] = vocab_size lowercase__ : List[str] = hidden_size lowercase__ : Any = num_hidden_layers lowercase__ : str = num_attention_heads lowercase__ : List[str] = hidden_act lowercase__ : str = intermediate_size lowercase__ : List[Any] = hidden_dropout_prob lowercase__ : str = attention_probs_dropout_prob lowercase__ : Union[str, Any] = max_position_embeddings lowercase__ : Optional[Any] = initializer_range lowercase__ : str = layer_norm_eps lowercase__ : Tuple = position_embedding_type lowercase__ : Tuple = use_cache
77
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : str = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all LED models at https://huggingface.co/models?filter=LED _UpperCamelCase : Optional[Any] = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } _UpperCamelCase : Optional[int] = { "allenai/led-base-16384": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def a_ ( ): '''simple docstring''' lowercase__ : int = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowercase__ : Union[str, Any] = bs[:] lowercase__ : str = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCAmelCase ) cs.append(2**8 + n ) n += 1 lowercase__ : str = [chr(_lowerCAmelCase ) for n in cs] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Dict = set() lowercase__ : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Optional[Any] = char return pairs class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = VOCAB_FILES_NAMES lowerCamelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , a , a , a="replace" , a="<s>" , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a=False , **a , ) -> Any: lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else bos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else eos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else sep_token lowercase__ : Dict = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else cls_token lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else unk_token lowercase__ : Tuple = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( errors=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , add_prefix_space=a , **a , ) with open(a , encoding='utf-8' ) as vocab_handle: lowercase__ : Tuple = json.load(a ) lowercase__ : Dict = {v: k for k, v in self.encoder.items()} lowercase__ : str = errors # how to handle errors in decoding lowercase__ : Optional[Any] = bytes_to_unicode() lowercase__ : Optional[Any] = {v: k for k, v in self.byte_encoder.items()} with open(a , encoding='utf-8' ) as merges_handle: lowercase__ : Optional[Any] = merges_handle.read().split('\n' )[1:-1] lowercase__ : Optional[int] = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ : Union[str, Any] = dict(zip(a , range(len(a ) ) ) ) lowercase__ : Tuple = {} lowercase__ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ : List[Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _UpperCAmelCase ( self ) -> List[Any]: return len(self.encoder ) def _UpperCAmelCase ( self ) -> str: return dict(self.encoder , **self.added_tokens_encoder ) def _UpperCAmelCase ( self , a ) -> List[str]: if token in self.cache: return self.cache[token] lowercase__ : Optional[Any] = tuple(a ) lowercase__ : int = get_pairs(a ) if not pairs: return token while True: lowercase__ : List[str] = min(a , key=lambda a : self.bpe_ranks.get(a , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ : List[str] = bigram lowercase__ : Union[str, Any] = [] lowercase__ : List[Any] = 0 while i < len(a ): try: lowercase__ : str = word.index(a , a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : Optional[int] = j if word[i] == first and i < len(a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : int = tuple(a ) lowercase__ : Dict = new_word if len(a ) == 1: break else: lowercase__ : Any = get_pairs(a ) lowercase__ : List[str] = ' '.join(a ) lowercase__ : Optional[Any] = word return word def _UpperCAmelCase ( self , a ) -> Union[str, Any]: lowercase__ : Tuple = [] for token in re.findall(self.pat , a ): lowercase__ : Union[str, Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(a ).split(' ' ) ) return bpe_tokens def _UpperCAmelCase ( self , a ) -> Optional[Any]: return self.encoder.get(a , self.encoder.get(self.unk_token ) ) def _UpperCAmelCase ( self , a ) -> Optional[int]: return self.decoder.get(a ) def _UpperCAmelCase ( self , a ) -> str: lowercase__ : Any = ''.join(a ) lowercase__ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Any = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : str = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(a , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a , ensure_ascii=a ) + '\n' ) lowercase__ : List[Any] = 0 with open(a , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) lowercase__ : Union[str, Any] = token_index writer.write(' '.join(a ) + '\n' ) index += 1 return vocab_file, merge_file def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ : Union[str, Any] = [self.cls_token_id] lowercase__ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) + [1] def _UpperCAmelCase ( self , a , a = None ) -> List[int]: lowercase__ : Dict = [self.sep_token_id] lowercase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCAmelCase ( self , a , a=False , **a ) -> Optional[int]: lowercase__ : Tuple = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(a ) > 0 and not text[0].isspace()): lowercase__ : List[str] = ' ' + text return (text, kwargs) def _UpperCAmelCase ( self , a , a = None , a = PaddingStrategy.DO_NOT_PAD , a = None , a = None , ) -> dict: lowercase__ : Dict = super()._pad( encoded_inputs=a , max_length=a , padding_strategy=a , pad_to_multiple_of=a , return_attention_mask=a , ) # Load from model defaults if return_attention_mask is None: lowercase__ : Union[str, Any] = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ : Any = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ : Tuple = len(encoded_inputs['global_attention_mask'] ) != len(a ) if needs_to_be_padded: lowercase__ : str = len(a ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ : Union[str, Any] = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowercase__ : List[str] = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
77
1
"""simple docstring""" import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase_ ( unittest.TestCase): @property def _UpperCAmelCase ( self ) -> Optional[int]: torch.manual_seed(0 ) lowercase__ : Optional[int] = UNetaDModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Dict = self.dummy_uncond_unet lowercase__ : Dict = PNDMScheduler() lowercase__ : Any = PNDMPipeline(unet=a , scheduler=a ) pndm.to(a ) pndm.set_progress_bar_config(disable=a ) lowercase__ : Optional[Any] = torch.manual_seed(0 ) lowercase__ : Optional[Any] = pndm(generator=a , num_inference_steps=2_0 , output_type='numpy' ).images lowercase__ : Optional[int] = torch.manual_seed(0 ) lowercase__ : Union[str, Any] = pndm(generator=a , num_inference_steps=2_0 , output_type='numpy' , return_dict=a )[0] lowercase__ : Tuple = image[0, -3:, -3:, -1] lowercase__ : Optional[int] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) lowercase__ : Optional[int] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : List[Any] = 'google/ddpm-cifar10-32' lowercase__ : Any = UNetaDModel.from_pretrained(a ) lowercase__ : str = PNDMScheduler() lowercase__ : Any = PNDMPipeline(unet=a , scheduler=a ) pndm.to(a ) pndm.set_progress_bar_config(disable=a ) lowercase__ : Optional[int] = torch.manual_seed(0 ) lowercase__ : str = pndm(generator=a , output_type='numpy' ).images lowercase__ : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) lowercase__ : str = np.array([0.1_564, 0.14_645, 0.1_406, 0.14_715, 0.12_425, 0.14_045, 0.13_115, 0.12_175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
77
"""simple docstring""" _UpperCamelCase : Union[str, Any] = 8.3_1_4_4_5_9_8 def a_ ( _lowerCAmelCase : float , _lowerCAmelCase : float ): '''simple docstring''' if temperature < 0: raise Exception('Temperature cannot be less than 0 K' ) if molar_mass <= 0: raise Exception('Molar mass cannot be less than or equal to 0 kg/mol' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _UpperCamelCase : List[Any] = 3_00 _UpperCamelCase : Tuple = 28 _UpperCamelCase : Any = rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
77
1
"""simple docstring""" import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _UpperCAmelCase ( self ) -> int: lowercase__ : Optional[int] = 1 lowercase__ : int = 3 lowercase__ : Optional[Any] = (3_2, 3_2) lowercase__ : int = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(a ) return image @property def _UpperCAmelCase ( self ) -> int: torch.manual_seed(0 ) lowercase__ : str = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) return model @property def _UpperCAmelCase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) lowercase__ : int = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) return model @property def _UpperCAmelCase ( self ) -> Optional[Any]: torch.manual_seed(0 ) lowercase__ : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModel(a ) @property def _UpperCAmelCase ( self ) -> Union[str, Any]: def extract(*a , **a ): class UpperCAmelCase_ : def __init__( self ) -> List[str]: lowercase__ : List[Any] = torch.ones([0] ) def _UpperCAmelCase ( self , a ) -> Dict: self.pixel_values.to(a ) return self return Out() return extract def _UpperCAmelCase ( self ) -> Any: lowercase__ : Optional[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator lowercase__ : Optional[int] = self.dummy_cond_unet lowercase__ : int = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=a , set_alpha_to_one=a , ) lowercase__ : Any = self.dummy_vae lowercase__ : Optional[Any] = self.dummy_text_encoder lowercase__ : Union[str, Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) # make sure here that pndm scheduler skips prk lowercase__ : List[Any] = StableDiffusionPipeline( unet=a , scheduler=a , vae=a , text_encoder=a , tokenizer=a , safety_checker=a , feature_extractor=self.dummy_extractor , ) lowercase__ : str = sd_pipe.to(a ) sd_pipe.set_progress_bar_config(disable=a ) lowercase__ : int = 'A painting of a squirrel eating a burger' lowercase__ : str = torch.Generator(device=a ).manual_seed(0 ) lowercase__ : List[Any] = sd_pipe([prompt] , generator=a , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' ) lowercase__ : List[Any] = output.images lowercase__ : Dict = torch.Generator(device=a ).manual_seed(0 ) lowercase__ : str = sd_pipe( [prompt] , generator=a , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' , return_dict=a , )[0] lowercase__ : List[str] = image[0, -3:, -3:, -1] lowercase__ : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) lowercase__ : Dict = np.array([0.5_756, 0.6_118, 0.5_005, 0.5_041, 0.5_471, 0.4_726, 0.4_976, 0.4_865, 0.4_864] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : List[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator lowercase__ : Dict = self.dummy_cond_unet lowercase__ : Union[str, Any] = PNDMScheduler(skip_prk_steps=a ) lowercase__ : int = self.dummy_vae lowercase__ : List[Any] = self.dummy_text_encoder lowercase__ : List[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) # make sure here that pndm scheduler skips prk lowercase__ : Tuple = StableDiffusionPipeline( unet=a , scheduler=a , vae=a , text_encoder=a , tokenizer=a , safety_checker=a , feature_extractor=self.dummy_extractor , ) lowercase__ : Optional[int] = sd_pipe.to(a ) sd_pipe.set_progress_bar_config(disable=a ) lowercase__ : Dict = 'A painting of a squirrel eating a burger' lowercase__ : Optional[int] = torch.Generator(device=a ).manual_seed(0 ) lowercase__ : List[Any] = sd_pipe([prompt] , generator=a , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' ) lowercase__ : str = output.images lowercase__ : Optional[Any] = torch.Generator(device=a ).manual_seed(0 ) lowercase__ : Union[str, Any] = sd_pipe( [prompt] , generator=a , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' , return_dict=a , )[0] lowercase__ : str = image[0, -3:, -3:, -1] lowercase__ : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) lowercase__ : str = np.array([0.5_125, 0.5_716, 0.4_828, 0.5_060, 0.5_650, 0.4_768, 0.5_185, 0.4_895, 0.4_993] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : int = StableDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-lms-pipe' , safety_checker=a ) assert isinstance(a , a ) assert isinstance(pipe.scheduler , a ) assert pipe.safety_checker is None lowercase__ : Optional[int] = pipe('example prompt' , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(a ) lowercase__ : Union[str, Any] = StableDiffusionPipeline.from_pretrained(a ) # sanity check that the pipeline still works assert pipe.safety_checker is None lowercase__ : Union[str, Any] = pipe('example prompt' , num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU' ) def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : Union[str, Any] = self.dummy_cond_unet lowercase__ : Any = PNDMScheduler(skip_prk_steps=a ) lowercase__ : Optional[Any] = self.dummy_vae lowercase__ : int = self.dummy_text_encoder lowercase__ : List[str] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) # put models in fp16 lowercase__ : Any = unet.half() lowercase__ : List[Any] = vae.half() lowercase__ : Optional[Any] = bert.half() # make sure here that pndm scheduler skips prk lowercase__ : Any = StableDiffusionPipeline( unet=a , scheduler=a , vae=a , text_encoder=a , tokenizer=a , safety_checker=a , feature_extractor=self.dummy_extractor , ) lowercase__ : Union[str, Any] = sd_pipe.to(a ) sd_pipe.set_progress_bar_config(disable=a ) lowercase__ : int = 'A painting of a squirrel eating a burger' lowercase__ : List[Any] = sd_pipe([prompt] , num_inference_steps=2 , output_type='np' ).images assert image.shape == (1, 6_4, 6_4, 3) @nightly @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Union[str, Any] = StableDiffusionPipeline.from_pretrained('runwayml/stable-diffusion-v1-5' , safety_checker=a ) lowercase__ : List[str] = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowercase__ : Any = sd_pipe.to(a ) sd_pipe.set_progress_bar_config(disable=a ) lowercase__ : str = ( 'portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle' ' coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with' ' anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and' ' children from bahnhof zoo, detailed ' ) lowercase__ : Any = 4_0_0_3_6_6_0_3_4_6 lowercase__ : Union[str, Any] = 7 # without safety guidance (sld_guidance_scale = 0) lowercase__ : Union[str, Any] = torch.manual_seed(a ) lowercase__ : str = sd_pipe( [prompt] , generator=a , guidance_scale=a , num_inference_steps=5_0 , output_type='np' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=0 , ) lowercase__ : Dict = output.images lowercase__ : Tuple = image[0, -3:, -3:, -1] lowercase__ : List[Any] = [0.2_278, 0.2_231, 0.2_249, 0.2_333, 0.2_303, 0.1_885, 0.2_273, 0.2_144, 0.2_176] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 # without safety guidance (strong configuration) lowercase__ : Any = torch.manual_seed(a ) lowercase__ : Any = sd_pipe( [prompt] , generator=a , guidance_scale=a , num_inference_steps=5_0 , output_type='np' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=2_0_0_0 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowercase__ : Dict = output.images lowercase__ : Dict = image[0, -3:, -3:, -1] lowercase__ : Any = [0.2_383, 0.2_276, 0.236, 0.2_192, 0.2_186, 0.2_053, 0.1_971, 0.1_901, 0.1_719] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : Optional[Any] = StableDiffusionPipeline.from_pretrained('runwayml/stable-diffusion-v1-5' , safety_checker=a ) lowercase__ : List[str] = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowercase__ : int = sd_pipe.to(a ) sd_pipe.set_progress_bar_config(disable=a ) lowercase__ : int = 'padme amidala taking a bath artwork, safe for work, no nudity' lowercase__ : Union[str, Any] = 2_7_3_4_9_7_1_7_5_5 lowercase__ : Any = 7 lowercase__ : Union[str, Any] = torch.manual_seed(a ) lowercase__ : Dict = sd_pipe( [prompt] , generator=a , guidance_scale=a , num_inference_steps=5_0 , output_type='np' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=0 , ) lowercase__ : Tuple = output.images lowercase__ : List[Any] = image[0, -3:, -3:, -1] lowercase__ : Optional[int] = [0.3_502, 0.3_622, 0.3_396, 0.3_642, 0.3_478, 0.3_318, 0.35, 0.3_348, 0.3_297] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 lowercase__ : Tuple = torch.manual_seed(a ) lowercase__ : Union[str, Any] = sd_pipe( [prompt] , generator=a , guidance_scale=a , num_inference_steps=5_0 , output_type='np' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=2_0_0_0 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowercase__ : str = output.images lowercase__ : Tuple = image[0, -3:, -3:, -1] lowercase__ : Tuple = [0.5_531, 0.5_206, 0.4_895, 0.5_156, 0.5_182, 0.4_751, 0.4_802, 0.4_803, 0.4_443] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : int = StableDiffusionPipeline.from_pretrained('runwayml/stable-diffusion-v1-5' ) lowercase__ : Union[str, Any] = sd_pipe.to(a ) sd_pipe.set_progress_bar_config(disable=a ) lowercase__ : int = ( 'the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c.' ' leyendecker' ) lowercase__ : Optional[int] = 1_0_4_4_3_5_5_2_3_4 lowercase__ : List[Any] = 1_2 lowercase__ : List[str] = torch.manual_seed(a ) lowercase__ : int = sd_pipe( [prompt] , generator=a , guidance_scale=a , num_inference_steps=5_0 , output_type='np' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=0 , ) lowercase__ : Tuple = output.images lowercase__ : Any = image[0, -3:, -3:, -1] lowercase__ : Tuple = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-7 lowercase__ : List[Any] = torch.manual_seed(a ) lowercase__ : str = sd_pipe( [prompt] , generator=a , guidance_scale=a , num_inference_steps=5_0 , output_type='np' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=2_0_0_0 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowercase__ : Dict = output.images lowercase__ : Dict = image[0, -3:, -3:, -1] lowercase__ : Optional[Any] = np.array([0.5_818, 0.6_285, 0.6_835, 0.6_019, 0.625, 0.6_754, 0.6_096, 0.6_334, 0.6_561] ) assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
77
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization lowerCamelCase__ : str = field(default="text-classification" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({"labels": ClassLabel}) lowerCamelCase__ : str = "text" lowerCamelCase__ : str = "labels" def _UpperCAmelCase ( self , a ) -> Tuple: if self.label_column not in features: raise ValueError(f"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , a ): raise ValueError(f"""Column {self.label_column} is not a ClassLabel.""" ) lowercase__ : Optional[Any] = copy.deepcopy(self ) lowercase__ : Optional[Any] = self.label_schema.copy() lowercase__ : Any = features[self.label_column] lowercase__ : Optional[Any] = label_schema return task_template @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return { self.text_column: "text", self.label_column: "labels", }
77
1
"""simple docstring""" from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput _UpperCamelCase : List[str] = 8 def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=BITS ): '''simple docstring''' lowercase__ : List[Any] = x.device lowercase__ : List[str] = (x * 255).int().clamp(0 , 255 ) lowercase__ : Union[str, Any] = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_lowerCAmelCase ) lowercase__ : Dict = rearrange(_lowerCAmelCase , 'd -> d 1 1' ) lowercase__ : List[Any] = rearrange(_lowerCAmelCase , 'b c h w -> b c 1 h w' ) lowercase__ : Any = ((x & mask) != 0).float() lowercase__ : Optional[int] = rearrange(_lowerCAmelCase , 'b c d h w -> b (c d) h w' ) lowercase__ : int = bits * 2 - 1 return bits def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any]=BITS ): '''simple docstring''' lowercase__ : Dict = x.device lowercase__ : str = (x > 0).int() lowercase__ : str = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_lowerCAmelCase , dtype=torch.intaa ) lowercase__ : Union[str, Any] = rearrange(_lowerCAmelCase , 'd -> d 1 1' ) lowercase__ : Tuple = rearrange(_lowerCAmelCase , 'b (c d) h w -> b c d h w' , d=8 ) lowercase__ : Union[str, Any] = reduce(x * mask , 'b c d h w -> b c h w' , 'sum' ) return (dec / 255).clamp(0.0 , 1.0 ) def a_ ( self : List[Any] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : float = 0.0 , _lowerCAmelCase : bool = True , _lowerCAmelCase : Optional[int]=None , _lowerCAmelCase : bool = True , ): '''simple docstring''' if self.num_inference_steps is None: raise ValueError( 'Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler' ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) lowercase__ : Union[str, Any] = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas lowercase__ : Optional[Any] = self.alphas_cumprod[timestep] lowercase__ : int = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod lowercase__ : Optional[int] = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf lowercase__ : Tuple = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" lowercase__ : str = self.bit_scale if self.config.clip_sample: lowercase__ : Tuple = torch.clamp(_lowerCAmelCase , -scale , _lowerCAmelCase ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) lowercase__ : int = self._get_variance(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[str] = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide lowercase__ : str = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf lowercase__ : Tuple = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf lowercase__ : Any = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 lowercase__ : Optional[Any] = model_output.device if torch.is_tensor(_lowerCAmelCase ) else 'cpu' lowercase__ : Tuple = torch.randn(model_output.shape , dtype=model_output.dtype , generator=_lowerCAmelCase ).to(_lowerCAmelCase ) lowercase__ : Tuple = self._get_variance(_lowerCAmelCase , _lowerCAmelCase ) ** 0.5 * eta * noise lowercase__ : int = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=_lowerCAmelCase , pred_original_sample=_lowerCAmelCase ) def a_ ( self : List[str] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : int="epsilon" , _lowerCAmelCase : int=None , _lowerCAmelCase : bool = True , ): '''simple docstring''' lowercase__ : Dict = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: lowercase__ , lowercase__ : List[Any] = torch.split(_lowerCAmelCase , sample.shape[1] , dim=1 ) else: lowercase__ : Tuple = None # 1. compute alphas, betas lowercase__ : int = self.alphas_cumprod[t] lowercase__ : Union[str, Any] = self.alphas_cumprod[t - 1] if t > 0 else self.one lowercase__ : Optional[int] = 1 - alpha_prod_t lowercase__ : Any = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": lowercase__ : Optional[int] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": lowercase__ : Optional[Any] = model_output else: raise ValueError(f"""Unsupported prediction_type {prediction_type}.""" ) # 3. Clip "predicted x_0" lowercase__ : Optional[int] = self.bit_scale if self.config.clip_sample: lowercase__ : Any = torch.clamp(_lowerCAmelCase , -scale , _lowerCAmelCase ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ : Union[str, Any] = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t lowercase__ : Tuple = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf lowercase__ : Optional[Any] = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise lowercase__ : Optional[int] = 0 if t > 0: lowercase__ : Dict = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=_lowerCAmelCase ).to(model_output.device ) lowercase__ : List[str] = (self._get_variance(_lowerCAmelCase , predicted_variance=_lowerCAmelCase ) ** 0.5) * noise lowercase__ : Union[str, Any] = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=_lowerCAmelCase , pred_original_sample=_lowerCAmelCase ) class UpperCAmelCase_ ( _a): def __init__( self , a , a , a = 1.0 , ) -> str: super().__init__() lowercase__ : Any = bit_scale lowercase__ : str = ( ddim_bit_scheduler_step if isinstance(a , a ) else ddpm_bit_scheduler_step ) self.register_modules(unet=a , scheduler=a ) @torch.no_grad() def __call__( self , a = 2_5_6 , a = 2_5_6 , a = 5_0 , a = None , a = 1 , a = "pil" , a = True , **a , ) -> Union[Tuple, ImagePipelineOutput]: lowercase__ : Optional[Any] = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=a , ) lowercase__ : Optional[int] = decimal_to_bits(a ) * self.bit_scale lowercase__ : Dict = latents.to(self.device ) self.scheduler.set_timesteps(a ) for t in self.progress_bar(self.scheduler.timesteps ): # predict the noise residual lowercase__ : str = self.unet(a , a ).sample # compute the previous noisy sample x_t -> x_t-1 lowercase__ : Tuple = self.scheduler.step(a , a , a ).prev_sample lowercase__ : List[str] = bits_to_decimal(a ) if output_type == "pil": lowercase__ : Optional[int] = self.numpy_to_pil(a ) if not return_dict: return (image,) return ImagePipelineOutput(images=a )
77
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np _UpperCamelCase : Any = re.compile(r"\b(a|an|the)\b", re.UNICODE) _UpperCamelCase : Union[str, Any] = None def a_ ( ): '''simple docstring''' lowercase__ : Optional[int] = argparse.ArgumentParser('Official evaluation script for SQuAD version 2.0.' ) parser.add_argument('data_file' , metavar='data.json' , help='Input data JSON file.' ) parser.add_argument('pred_file' , metavar='pred.json' , help='Model predictions.' ) parser.add_argument( '--out-file' , '-o' , metavar='eval.json' , help='Write accuracy metrics to file (default is stdout).' ) parser.add_argument( '--na-prob-file' , '-n' , metavar='na_prob.json' , help='Model estimates of probability of no answer.' ) parser.add_argument( '--na-prob-thresh' , '-t' , type=_lowerCAmelCase , default=1.0 , help='Predict "" if no-answer probability exceeds this (default = 1.0).' , ) parser.add_argument( '--out-image-dir' , '-p' , metavar='out_images' , default=_lowerCAmelCase , help='Save precision-recall curves to directory.' ) parser.add_argument('--verbose' , '-v' , action='store_true' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Tuple = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Optional[int] = bool(qa['answers']['text'] ) return qid_to_has_ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' def remove_articles(_lowerCAmelCase : int ): return ARTICLES_REGEX.sub(' ' , _lowerCAmelCase ) def white_space_fix(_lowerCAmelCase : str ): return " ".join(text.split() ) def remove_punc(_lowerCAmelCase : List[Any] ): lowercase__ : int = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_lowerCAmelCase : List[str] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_lowerCAmelCase ) ) ) ) def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' if not s: return [] return normalize_answer(_lowerCAmelCase ).split() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ): '''simple docstring''' return int(normalize_answer(_lowerCAmelCase ) == normalize_answer(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ): '''simple docstring''' lowercase__ : Dict = get_tokens(_lowerCAmelCase ) lowercase__ : List[str] = get_tokens(_lowerCAmelCase ) lowercase__ : List[Any] = collections.Counter(_lowerCAmelCase ) & collections.Counter(_lowerCAmelCase ) lowercase__ : int = sum(common.values() ) if len(_lowerCAmelCase ) == 0 or len(_lowerCAmelCase ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 lowercase__ : Any = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Dict = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Any = (2 * precision * recall) / (precision + recall) return fa def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Optional[int] = {} lowercase__ : Union[str, Any] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Any = qa['id'] lowercase__ : Union[str, Any] = [t for t in qa['answers']['text'] if normalize_answer(_lowerCAmelCase )] if not gold_answers: # For unanswerable questions, only correct answer is empty string lowercase__ : Dict = [''] if qid not in preds: print(f"""Missing prediction for {qid}""" ) continue lowercase__ : Optional[int] = preds[qid] # Take max over all gold answers lowercase__ : int = max(compute_exact(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) lowercase__ : Optional[Any] = max(compute_fa(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) return exact_scores, fa_scores def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : str = {} for qid, s in scores.items(): lowercase__ : int = na_probs[qid] > na_prob_thresh if pred_na: lowercase__ : Optional[Any] = float(not qid_to_has_ans[qid] ) else: lowercase__ : Optional[Any] = s return new_scores def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str=None ): '''simple docstring''' if not qid_list: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores.values() ) / total), ('f1', 1_0_0.0 * sum(fa_scores.values() ) / total), ('total', total), ] ) else: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores[k] for k in qid_list ) / total), ('f1', 1_0_0.0 * sum(fa_scores[k] for k in qid_list ) / total), ('total', total), ] ) def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' for k in new_eval: lowercase__ : int = new_eval[k] def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple ): '''simple docstring''' plt.step(_lowerCAmelCase , _lowerCAmelCase , color='b' , alpha=0.2 , where='post' ) plt.fill_between(_lowerCAmelCase , _lowerCAmelCase , step='post' , alpha=0.2 , color='b' ) plt.xlabel('Recall' ) plt.ylabel('Precision' ) plt.xlim([0.0, 1.0_5] ) plt.ylim([0.0, 1.0_5] ) plt.title(_lowerCAmelCase ) plt.savefig(_lowerCAmelCase ) plt.clf() def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : str , _lowerCAmelCase : Any=None , _lowerCAmelCase : List[str]=None ): '''simple docstring''' lowercase__ : Optional[int] = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) lowercase__ : Tuple = 0.0 lowercase__ : List[str] = 1.0 lowercase__ : List[str] = 0.0 lowercase__ : Union[str, Any] = [1.0] lowercase__ : List[Any] = [0.0] lowercase__ : Optional[int] = 0.0 for i, qid in enumerate(_lowerCAmelCase ): if qid_to_has_ans[qid]: true_pos += scores[qid] lowercase__ : Tuple = true_pos / float(i + 1 ) lowercase__ : Union[str, Any] = true_pos / float(_lowerCAmelCase ) if i == len(_lowerCAmelCase ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_lowerCAmelCase ) recalls.append(_lowerCAmelCase ) if out_image: plot_pr_curve(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return {"ap": 1_0_0.0 * avg_prec} def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple ): '''simple docstring''' if out_image_dir and not os.path.exists(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) lowercase__ : List[str] = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return lowercase__ : Dict = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_exact.png' ) , title='Precision-Recall curve for Exact Match score' , ) lowercase__ : Tuple = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_f1.png' ) , title='Precision-Recall curve for F1 score' , ) lowercase__ : List[Any] = {k: float(_lowerCAmelCase ) for k, v in qid_to_has_ans.items()} lowercase__ : Any = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_oracle.png' ) , title='Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)' , ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_exact' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_f1' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_oracle' ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' if not qid_list: return lowercase__ : List[str] = [na_probs[k] for k in qid_list] lowercase__ : Tuple = np.ones_like(_lowerCAmelCase ) / float(len(_lowerCAmelCase ) ) plt.hist(_lowerCAmelCase , weights=_lowerCAmelCase , bins=20 , range=(0.0, 1.0) ) plt.xlabel('Model probability of no-answer' ) plt.ylabel('Proportion of dataset' ) plt.title(f"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(_lowerCAmelCase , f"""na_prob_hist_{name}.png""" ) ) plt.clf() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : Tuple = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) lowercase__ : int = num_no_ans lowercase__ : Optional[int] = cur_score lowercase__ : Tuple = 0.0 lowercase__ : Dict = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) for i, qid in enumerate(_lowerCAmelCase ): if qid not in scores: continue if qid_to_has_ans[qid]: lowercase__ : Optional[int] = scores[qid] else: if preds[qid]: lowercase__ : List[Any] = -1 else: lowercase__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: lowercase__ : Dict = cur_score lowercase__ : Optional[int] = na_probs[qid] return 1_0_0.0 * best_score / len(_lowerCAmelCase ), best_thresh def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str ): '''simple docstring''' lowercase__ , lowercase__ : List[Any] = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ , lowercase__ : Dict = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Any = best_exact lowercase__ : Tuple = exact_thresh lowercase__ : Optional[Any] = best_fa lowercase__ : Any = fa_thresh def a_ ( ): '''simple docstring''' with open(OPTS.data_file ) as f: lowercase__ : List[Any] = json.load(_lowerCAmelCase ) lowercase__ : Union[str, Any] = dataset_json['data'] with open(OPTS.pred_file ) as f: lowercase__ : str = json.load(_lowerCAmelCase ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: lowercase__ : Union[str, Any] = json.load(_lowerCAmelCase ) else: lowercase__ : str = {k: 0.0 for k in preds} lowercase__ : int = make_qid_to_has_ans(_lowerCAmelCase ) # maps qid to True/False lowercase__ : List[str] = [k for k, v in qid_to_has_ans.items() if v] lowercase__ : Any = [k for k, v in qid_to_has_ans.items() if not v] lowercase__ , lowercase__ : Any = get_raw_scores(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Optional[Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Union[str, Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Tuple = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase ) if has_ans_qids: lowercase__ : int = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'HasAns' ) if no_ans_qids: lowercase__ : Optional[Any] = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'NoAns' ) if OPTS.na_prob_file: find_all_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'hasAns' ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'noAns' ) if OPTS.out_file: with open(OPTS.out_file , 'w' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) else: print(json.dumps(_lowerCAmelCase , indent=2 ) ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int ): '''simple docstring''' return int((input_a, input_a).count(0 ) != 0 ) def a_ ( ): '''simple docstring''' assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
77
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = 0 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Tuple = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Any: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : str = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : int = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = CLIPConfig() # Create a dummy config file with image_proceesor_type lowercase__ : Optional[int] = Path(a ) / 'preprocessor_config.json' lowercase__ : Optional[int] = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowercase__ : int = AutoImageProcessor.from_pretrained(a ).to_dict() config_dict.pop('image_processor_type' ) lowercase__ : Tuple = CLIPImageProcessor(**a ) # save in new folder model_config.save_pretrained(a ) config.save_pretrained(a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) # make sure private variable is not incorrectly saved lowercase__ : Optional[int] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = Path(a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'clip-base is not a local folder and is not a valid model identifier' ): lowercase__ : Any = AutoImageProcessor.from_pretrained('clip-base' ) def _UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex( a , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowercase__ : Dict = AutoImageProcessor.from_pretrained(a , revision='aaaaaa' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def _UpperCAmelCase ( self ) -> Optional[int]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(a ): lowercase__ : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : str = AutoImageProcessor.from_pretrained(a , trust_remote_code=a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def _UpperCAmelCase ( self ) -> int: try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoImageProcessor.register(a , a ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Optional[Any] = Path(a ) / 'preprocessor_config.json' lowercase__ : List[Any] = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = CustomImageProcessor.from_pretrained(a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCAmelCase ( self ) -> Dict: class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = True try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # If remote code is not set, the default is to use local lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowercase__ : int = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
77
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCamelCase : Optional[int] = logging.get_logger(__name__) _UpperCamelCase : Dict = "▁" _UpperCamelCase : Optional[Any] = {"vocab_file": "sentencepiece.bpe.model"} _UpperCamelCase : Dict = { "vocab_file": { "facebook/xglm-564M": "https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model", } } _UpperCamelCase : Tuple = { "facebook/xglm-564M": 20_48, } class UpperCAmelCase_ ( _a): lowerCamelCase__ : Optional[int] = VOCAB_FILES_NAMES lowerCamelCase__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : List[str] = ["input_ids", "attention_mask"] def __init__( self , a , a="<s>" , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a = None , **a , ) -> None: lowercase__ : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer lowercase__ : int = 7 lowercase__ : int = [f"""<madeupword{i}>""" for i in range(self.num_madeup_words )] lowercase__ : Optional[int] = kwargs.get('additional_special_tokens' , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , sp_model_kwargs=self.sp_model_kwargs , **a , ) lowercase__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(a ) ) lowercase__ : Union[str, Any] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab lowercase__ : List[str] = 1 # Mimic fairseq token-to-id alignment for the first 4 token lowercase__ : Union[str, Any] = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} lowercase__ : Tuple = len(self.sp_model ) lowercase__ : List[Any] = {f"""<madeupword{i}>""": sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(a ) lowercase__ : str = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ) -> List[str]: lowercase__ : List[Any] = self.__dict__.copy() lowercase__ : Tuple = None lowercase__ : Dict = self.sp_model.serialized_model_proto() return state def __setstate__( self , a ) -> int: lowercase__ : Optional[Any] = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowercase__ : Tuple = {} lowercase__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return [self.sep_token_id] + token_ids_a lowercase__ : Optional[Any] = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) def _UpperCAmelCase ( self , a , a = None ) -> List[int]: lowercase__ : Union[str, Any] = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def _UpperCAmelCase ( self ) -> Optional[int]: return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def _UpperCAmelCase ( self ) -> Dict: lowercase__ : Optional[Any] = {self.convert_ids_to_tokens(a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _UpperCAmelCase ( self , a ) -> List[str]: return self.sp_model.encode(a , out_type=a ) def _UpperCAmelCase ( self , a ) -> Any: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] lowercase__ : Any = self.sp_model.PieceToId(a ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _UpperCAmelCase ( self , a ) -> List[Any]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : List[Any] = ''.join(a ).replace(a , ' ' ).strip() return out_string def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Any = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , a ) elif not os.path.isfile(self.vocab_file ): with open(a , 'wb' ) as fi: lowercase__ : Any = self.sp_model.serialized_model_proto() fi.write(a ) return (out_vocab_file,)
77
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class UpperCAmelCase_ ( _a): lowerCamelCase__ : Dict = ["image_processor", "tokenizer"] lowerCamelCase__ : Dict = "BlipImageProcessor" lowerCamelCase__ : Union[str, Any] = "AutoTokenizer" def __init__( self , a , a , a ) -> Optional[int]: super().__init__(a , a ) # add QFormer tokenizer lowercase__ : Dict = qformer_tokenizer def __call__( self , a = None , a = None , a = True , a = False , a = None , a = None , a = 0 , a = None , a = None , a = False , a = False , a = False , a = False , a = False , a = True , a = None , **a , ) -> BatchFeature: if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) lowercase__ : List[Any] = BatchFeature() if text is not None: lowercase__ : Optional[int] = self.tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) encoding.update(a ) lowercase__ : Optional[int] = self.qformer_tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) lowercase__ : List[str] = qformer_text_encoding.pop('input_ids' ) lowercase__ : Any = qformer_text_encoding.pop('attention_mask' ) if images is not None: lowercase__ : List[Any] = self.image_processor(a , return_tensors=a ) encoding.update(a ) return encoding def _UpperCAmelCase ( self , *a , **a ) -> List[str]: return self.tokenizer.batch_decode(*a , **a ) def _UpperCAmelCase ( self , *a , **a ) -> Tuple: return self.tokenizer.decode(*a , **a ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : str = self.tokenizer.model_input_names lowercase__ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: if os.path.isfile(a ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(a , exist_ok=a ) lowercase__ : int = os.path.join(a , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(a ) return super().save_pretrained(a , **a ) @classmethod def _UpperCAmelCase ( cls , a , **a ) -> str: lowercase__ : str = AutoTokenizer.from_pretrained(a , subfolder='qformer_tokenizer' ) lowercase__ : int = cls._get_arguments_from_pretrained(a , **a ) args.append(a ) return cls(*a )
77
1
"""simple docstring""" import random def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Optional[int] = num - 1 lowercase__ : Optional[int] = 0 while s % 2 == 0: lowercase__ : Optional[Any] = s // 2 t += 1 for _ in range(5 ): lowercase__ : List[str] = random.randrange(2 , num - 1 ) lowercase__ : Any = pow(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if v != 1: lowercase__ : Optional[Any] = 0 while v != (num - 1): if i == t - 1: return False else: lowercase__ : str = i + 1 lowercase__ : Optional[Any] = (v**2) % num return True def a_ ( _lowerCAmelCase : int ): '''simple docstring''' if num < 2: return False lowercase__ : Dict = [ 2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 101, 103, 107, 109, 113, 127, 131, 137, 139, 149, 151, 157, 163, 167, 173, 179, 181, 191, 193, 197, 199, 211, 223, 227, 229, 233, 239, 241, 251, 257, 263, 269, 271, 277, 281, 283, 293, 307, 311, 313, 317, 331, 337, 347, 349, 353, 359, 367, 373, 379, 383, 389, 397, 401, 409, 419, 421, 431, 433, 439, 443, 449, 457, 461, 463, 467, 479, 487, 491, 499, 503, 509, 521, 523, 541, 547, 557, 563, 569, 571, 577, 587, 593, 599, 601, 607, 613, 617, 619, 631, 641, 643, 647, 653, 659, 661, 673, 677, 683, 691, 701, 709, 719, 727, 733, 739, 743, 751, 757, 761, 769, 773, 787, 797, 809, 811, 821, 823, 827, 829, 839, 853, 857, 859, 863, 877, 881, 883, 887, 907, 911, 919, 929, 937, 941, 947, 953, 967, 971, 977, 983, 991, 997, ] if num in low_primes: return True for prime in low_primes: if (num % prime) == 0: return False return rabin_miller(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : int = 1024 ): '''simple docstring''' while True: lowercase__ : Tuple = random.randrange(2 ** (keysize - 1) , 2 ** (keysize) ) if is_prime_low_num(_lowerCAmelCase ): return num if __name__ == "__main__": _UpperCamelCase : Any = generate_large_prime() print(("Prime number:", num)) print(("is_prime_low_num:", is_prime_low_num(num)))
77
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({}) lowerCamelCase__ : str = "text" @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return {self.text_column: "text"}
77
1
"""simple docstring""" import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : str = VideoToVideoSDPipeline lowerCamelCase__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({"video"}) - {"image", "width", "height"} lowerCamelCase__ : Tuple = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"video"}) - {"image"} lowerCamelCase__ : Union[str, Any] = PipelineTesterMixin.required_optional_params - {"latents"} lowerCamelCase__ : List[Any] = False # No `output_type`. lowerCamelCase__ : List[Any] = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ]) def _UpperCAmelCase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) lowercase__ : Optional[int] = UNetaDConditionModel( block_out_channels=(3_2, 6_4, 6_4, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'DownBlock3D') , up_block_types=('UpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D') , cross_attention_dim=3_2 , attention_head_dim=4 , ) lowercase__ : Union[str, Any] = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=a , set_alpha_to_one=a , ) torch.manual_seed(0 ) lowercase__ : Tuple = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) lowercase__ : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='gelu' , projection_dim=5_1_2 , ) lowercase__ : Union[str, Any] = CLIPTextModel(a ) lowercase__ : Optional[int] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) lowercase__ : List[str] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def _UpperCAmelCase ( self , a , a=0 ) -> Union[str, Any]: # 3 frames lowercase__ : List[Any] = floats_tensor((1, 3, 3, 3_2, 3_2) , rng=random.Random(a ) ).to(a ) if str(a ).startswith('mps' ): lowercase__ : Tuple = torch.manual_seed(a ) else: lowercase__ : Any = torch.Generator(device=a ).manual_seed(a ) lowercase__ : Any = { 'prompt': 'A painting of a squirrel eating a burger', 'video': video, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'pt', } return inputs def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : List[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator lowercase__ : Optional[Any] = self.get_dummy_components() lowercase__ : List[Any] = VideoToVideoSDPipeline(**a ) lowercase__ : Optional[Any] = sd_pipe.to(a ) sd_pipe.set_progress_bar_config(disable=a ) lowercase__ : List[Any] = self.get_dummy_inputs(a ) lowercase__ : Any = 'np' lowercase__ : Any = sd_pipe(**a ).frames lowercase__ : List[Any] = frames[0][-3:, -3:, -1] assert frames[0].shape == (3_2, 3_2, 3) lowercase__ : Optional[Any] = np.array([1_0_6, 1_1_7, 1_1_3, 1_7_4, 1_3_7, 1_1_2, 1_4_8, 1_5_1, 1_3_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def _UpperCAmelCase ( self ) -> int: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=a , expected_max_diff=5e-3 ) @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def _UpperCAmelCase ( self ) -> Dict: pass @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def _UpperCAmelCase ( self ) -> int: pass @unittest.skip(reason='`num_images_per_prompt` argument is not supported for this pipeline.' ) def _UpperCAmelCase ( self ) -> int: pass def _UpperCAmelCase ( self ) -> Union[str, Any]: return super().test_progress_bar() @slow @skip_mps class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> int: lowercase__ : Dict = VideoToVideoSDPipeline.from_pretrained('cerspense/zeroscope_v2_XL' , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames lowercase__ : Any = torch.Generator(device='cpu' ).manual_seed(0 ) lowercase__ : int = torch.randn((1, 1_0, 3, 1_0_2_4, 5_7_6) , generator=a ) lowercase__ : Optional[int] = video.to('cuda' ) lowercase__ : Optional[int] = 'Spiderman is surfing' lowercase__ : Union[str, Any] = pipe(a , video=a , generator=a , num_inference_steps=3 , output_type='pt' ).frames lowercase__ : Any = np.array([-1.0_458_984, -1.1_279_297, -0.9_663_086, -0.91_503_906, -0.75_097_656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
77
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
77
1
"""simple docstring""" from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf _UpperCamelCase : Optional[int] = logging.get_logger(__name__) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Tuple = [ "no_inference", "no_cuda", "no_tpu", "no_speed", "no_memory", "no_env_print", "no_multi_process", ] def __init__( self , **a ) -> List[Any]: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: lowercase__ : Optional[Any] = deprecated_arg[3:] lowercase__ : Dict = not kwargs.pop(a ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no-{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) lowercase__ : Dict = kwargs.pop('tpu_name' , self.tpu_name ) lowercase__ : Optional[int] = kwargs.pop('device_idx' , self.device_idx ) lowercase__ : List[Any] = kwargs.pop('eager_mode' , self.eager_mode ) lowercase__ : Union[str, Any] = kwargs.pop('use_xla' , self.use_xla ) super().__init__(**a ) lowerCamelCase__ : str = field( default=_a , metadata={"help": "Name of TPU"} , ) lowerCamelCase__ : int = field( default=0 , metadata={"help": "CPU / GPU device index. Defaults to 0."} , ) lowerCamelCase__ : bool = field(default=_a , metadata={"help": "Benchmark models in eager model."}) lowerCamelCase__ : bool = field( default=_a , metadata={ "help": "Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`." } , ) @cached_property def _UpperCAmelCase ( self ) -> Tuple["tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ['tf'] ) lowercase__ : Dict = None if self.tpu: try: if self.tpu_name: lowercase__ : Tuple = tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: lowercase__ : Optional[int] = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: lowercase__ : str = None return tpu @cached_property def _UpperCAmelCase ( self ) -> Tuple["tf.distribute.Strategy", "tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ['tf'] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) lowercase__ : Union[str, Any] = tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , 'GPU' ) lowercase__ : str = tf.distribute.OneDeviceStrategy(device=f"""/gpu:{self.device_idx}""" ) else: tf.config.set_visible_devices([] , 'GPU' ) # disable GPU lowercase__ : Dict = tf.distribute.OneDeviceStrategy(device=f"""/cpu:{self.device_idx}""" ) return strategy @property def _UpperCAmelCase ( self ) -> bool: requires_backends(self , ['tf'] ) return self._setup_tpu is not None @property def _UpperCAmelCase ( self ) -> "tf.distribute.Strategy": requires_backends(self , ['tf'] ) return self._setup_strategy @property def _UpperCAmelCase ( self ) -> Tuple: requires_backends(self , ['tf'] ) return tf.config.list_physical_devices('GPU' ) @property def _UpperCAmelCase ( self ) -> int: requires_backends(self , ['tf'] ) if self.cuda: return len(self.gpu_list ) return 0 @property def _UpperCAmelCase ( self ) -> bool: return self.n_gpu > 0
77
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING _UpperCamelCase : Dict = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , **a ) -> Dict: super().__init__(**a ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type(a ) def __call__( self , a , a = None , **a , ) -> List[str]: if "text_queries" in kwargs: lowercase__ : Optional[Any] = kwargs.pop('text_queries' ) if isinstance(a , (str, Image.Image) ): lowercase__ : Optional[Any] = {'image': image, 'candidate_labels': candidate_labels} else: lowercase__ : List[str] = image lowercase__ : Optional[Any] = super().__call__(a , **a ) return results def _UpperCAmelCase ( self , **a ) -> Dict: lowercase__ : Optional[Any] = {} if "threshold" in kwargs: lowercase__ : Tuple = kwargs['threshold'] if "top_k" in kwargs: lowercase__ : List[Any] = kwargs['top_k'] return {}, {}, postprocess_params def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : Any = load_image(inputs['image'] ) lowercase__ : Optional[int] = inputs['candidate_labels'] if isinstance(a , a ): lowercase__ : Optional[int] = candidate_labels.split(',' ) lowercase__ : Optional[int] = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(a ): lowercase__ : List[str] = self.tokenizer(a , return_tensors=self.framework ) lowercase__ : List[Any] = self.image_processor(a , return_tensors=self.framework ) yield { "is_last": i == len(a ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : List[Any] = model_inputs.pop('target_size' ) lowercase__ : Dict = model_inputs.pop('candidate_label' ) lowercase__ : Dict = model_inputs.pop('is_last' ) lowercase__ : Optional[int] = self.model(**a ) lowercase__ : Any = {'target_size': target_size, 'candidate_label': candidate_label, 'is_last': is_last, **outputs} return model_outputs def _UpperCAmelCase ( self , a , a=0.1 , a=None ) -> Union[str, Any]: lowercase__ : Dict = [] for model_output in model_outputs: lowercase__ : List[Any] = model_output['candidate_label'] lowercase__ : Optional[int] = BaseModelOutput(a ) lowercase__ : Any = self.image_processor.post_process_object_detection( outputs=a , threshold=a , target_sizes=model_output['target_size'] )[0] for index in outputs["scores"].nonzero(): lowercase__ : Union[str, Any] = outputs['scores'][index].item() lowercase__ : Tuple = self._get_bounding_box(outputs['boxes'][index][0] ) lowercase__ : Tuple = {'score': score, 'label': label, 'box': box} results.append(a ) lowercase__ : Dict = sorted(a , key=lambda a : x["score"] , reverse=a ) if top_k: lowercase__ : Dict = results[:top_k] return results def _UpperCAmelCase ( self , a ) -> Dict[str, int]: if self.framework != "pt": raise ValueError('The ZeroShotObjectDetectionPipeline is only available in PyTorch.' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ : Tuple = box.int().tolist() lowercase__ : Any = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
77
1
"""simple docstring""" import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class UpperCAmelCase_ ( _a): def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Optional[int] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(a , 'hidden_sizes' ) ) self.parent.assertTrue(hasattr(a , 'neck_hidden_sizes' ) ) self.parent.assertTrue(hasattr(a , 'num_attention_heads' ) ) class UpperCAmelCase_ : def __init__( self , a , a=1_3 , a=3_2 , a=2 , a=3 , a=6_4_0 , a=4 , a="silu" , a=3 , a=3_2 , a=0.1 , a=0.1 , a=0.1 , a=0.02 , a=True , a=True , a=1_0 , a=None , ) -> Dict: lowercase__ : int = parent lowercase__ : List[Any] = batch_size lowercase__ : Tuple = image_size lowercase__ : Any = patch_size lowercase__ : Union[str, Any] = num_channels lowercase__ : Any = last_hidden_size lowercase__ : Union[str, Any] = num_attention_heads lowercase__ : List[Any] = hidden_act lowercase__ : Dict = conv_kernel_size lowercase__ : List[str] = output_stride lowercase__ : Optional[int] = hidden_dropout_prob lowercase__ : Tuple = attention_probs_dropout_prob lowercase__ : Union[str, Any] = classifier_dropout_prob lowercase__ : List[Any] = use_labels lowercase__ : int = is_training lowercase__ : Dict = num_labels lowercase__ : Union[str, Any] = initializer_range lowercase__ : List[str] = scope def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase__ : Optional[int] = None lowercase__ : int = None if self.use_labels: lowercase__ : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) lowercase__ : str = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowercase__ : Union[str, Any] = self.get_config() return config, pixel_values, labels, pixel_labels def _UpperCAmelCase ( self ) -> Optional[Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def _UpperCAmelCase ( self , a , a , a , a ) -> Union[str, Any]: lowercase__ : Optional[int] = MobileViTModel(config=a ) model.to(a ) model.eval() lowercase__ : Optional[Any] = model(a ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _UpperCAmelCase ( self , a , a , a , a ) -> Optional[int]: lowercase__ : Dict = self.num_labels lowercase__ : int = MobileViTForImageClassification(a ) model.to(a ) model.eval() lowercase__ : Tuple = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCAmelCase ( self , a , a , a , a ) -> Any: lowercase__ : List[str] = self.num_labels lowercase__ : int = MobileViTForSemanticSegmentation(a ) model.to(a ) model.eval() lowercase__ : str = model(a ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) lowercase__ : Any = model(a , labels=a ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _UpperCAmelCase ( self ) -> Dict: lowercase__ : Union[str, Any] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ , lowercase__ : str = config_and_inputs lowercase__ : List[Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase_ ( _a , _a , unittest.TestCase): lowerCamelCase__ : Optional[Any] = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) lowerCamelCase__ : Dict = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) lowerCamelCase__ : Optional[int] = False lowerCamelCase__ : Optional[Any] = False lowerCamelCase__ : Tuple = False lowerCamelCase__ : List[str] = False def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : List[str] = MobileViTModelTester(self ) lowercase__ : Tuple = MobileViTConfigTester(self , config_class=a , has_text_modality=a ) def _UpperCAmelCase ( self ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason='MobileViT does not use inputs_embeds' ) def _UpperCAmelCase ( self ) -> int: pass @unittest.skip(reason='MobileViT does not support input and output embeddings' ) def _UpperCAmelCase ( self ) -> List[str]: pass @unittest.skip(reason='MobileViT does not output attentions' ) def _UpperCAmelCase ( self ) -> int: pass def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ , lowercase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ : Tuple = model_class(a ) lowercase__ : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ : Union[str, Any] = [*signature.parameters.keys()] lowercase__ : Tuple = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def _UpperCAmelCase ( self ) -> List[str]: pass def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _UpperCAmelCase ( self ) -> Any: def check_hidden_states_output(a , a , a ): lowercase__ : Optional[int] = model_class(a ) model.to(a ) model.eval() with torch.no_grad(): lowercase__ : Tuple = model(**self._prepare_for_class(a , a ) ) lowercase__ : Tuple = outputs.hidden_states lowercase__ : Tuple = 5 self.assertEqual(len(a ) , a ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. lowercase__ : Tuple = 2 for i in range(len(a ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) lowercase__ , lowercase__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ : Any = True check_hidden_states_output(a , a , a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase__ : Dict = True check_hidden_states_output(a , a , a ) def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) def _UpperCAmelCase ( self ) -> int: lowercase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*a ) @slow def _UpperCAmelCase ( self ) -> Union[str, Any]: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : int = MobileViTModel.from_pretrained(a ) self.assertIsNotNone(a ) def a_ ( ): '''simple docstring''' lowercase__ : Any = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class UpperCAmelCase_ ( unittest.TestCase): @cached_property def _UpperCAmelCase ( self ) -> str: return MobileViTImageProcessor.from_pretrained('apple/mobilevit-xx-small' ) if is_vision_available() else None @slow def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : List[str] = MobileViTForImageClassification.from_pretrained('apple/mobilevit-xx-small' ).to(a ) lowercase__ : Optional[int] = self.default_image_processor lowercase__ : str = prepare_img() lowercase__ : Dict = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowercase__ : str = model(**a ) # verify the logits lowercase__ : str = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , a ) lowercase__ : Union[str, Any] = torch.tensor([-1.9_364, -1.2_327, -0.4_653] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1e-4 ) ) @slow def _UpperCAmelCase ( self ) -> Dict: lowercase__ : str = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) lowercase__ : Any = model.to(a ) lowercase__ : Any = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) lowercase__ : Tuple = prepare_img() lowercase__ : str = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowercase__ : Optional[Any] = model(**a ) lowercase__ : int = outputs.logits # verify the logits lowercase__ : List[str] = torch.Size((1, 2_1, 3_2, 3_2) ) self.assertEqual(logits.shape , a ) lowercase__ : str = torch.tensor( [ [[6.9_713, 6.9_786, 7.2_422], [7.2_893, 7.2_825, 7.4_446], [7.6_580, 7.8_797, 7.9_420]], [[-10.6_869, -10.3_250, -10.3_471], [-10.4_228, -9.9_868, -9.7_132], [-11.0_405, -11.0_221, -10.7_318]], [[-3.3_089, -2.8_539, -2.6_740], [-3.2_706, -2.5_621, -2.5_108], [-3.2_534, -2.6_615, -2.6_651]], ] , device=a , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , a , atol=1e-4 ) ) @slow def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Tuple = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) lowercase__ : Optional[Any] = model.to(a ) lowercase__ : Tuple = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) lowercase__ : List[Any] = prepare_img() lowercase__ : Optional[int] = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowercase__ : Optional[Any] = model(**a ) lowercase__ : Optional[Any] = outputs.logits.detach().cpu() lowercase__ : str = image_processor.post_process_semantic_segmentation(outputs=a , target_sizes=[(5_0, 6_0)] ) lowercase__ : Tuple = torch.Size((5_0, 6_0) ) self.assertEqual(segmentation[0].shape , a ) lowercase__ : Optional[Any] = image_processor.post_process_semantic_segmentation(outputs=a ) lowercase__ : str = torch.Size((3_2, 3_2) ) self.assertEqual(segmentation[0].shape , a )
77
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ : def __init__( self , a , a=1_2 , a=7 , a=True , a=True , a=True , a=9_9 , a=3_2 , a=3_2 , a=2 , a=4 , a=3_7 , a=0.1 , a=0.1 , a=5_1_2 , a=0.02 , a=0 , a=None , ) -> Union[str, Any]: lowercase__ : Any = parent lowercase__ : str = batch_size lowercase__ : List[Any] = seq_length lowercase__ : Union[str, Any] = is_training lowercase__ : List[str] = use_input_mask lowercase__ : int = use_labels lowercase__ : List[Any] = vocab_size lowercase__ : str = hidden_size lowercase__ : int = projection_dim lowercase__ : Optional[int] = num_hidden_layers lowercase__ : Any = num_attention_heads lowercase__ : Optional[Any] = intermediate_size lowercase__ : Optional[Any] = dropout lowercase__ : Optional[int] = attention_dropout lowercase__ : Optional[int] = max_position_embeddings lowercase__ : str = initializer_range lowercase__ : Tuple = scope lowercase__ : int = bos_token_id def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : int = None if self.use_input_mask: lowercase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowercase__ : int = input_mask.numpy() lowercase__ , lowercase__ : Tuple = input_mask.shape lowercase__ : List[str] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a ): lowercase__ : Dict = 1 lowercase__ : Union[str, Any] = 0 lowercase__ : Tuple = self.get_config() return config, input_ids, tf.convert_to_tensor(a ) def _UpperCAmelCase ( self ) -> List[Any]: return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _UpperCAmelCase ( self , a , a , a ) -> Any: lowercase__ : List[Any] = TFBlipTextModel(config=a ) lowercase__ : Optional[int] = model(a , attention_mask=a , training=a ) lowercase__ : List[str] = model(a , training=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _UpperCAmelCase ( self ) -> Any: lowercase__ : Optional[Any] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ : Any = config_and_inputs lowercase__ : Any = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : Dict = (TFBlipTextModel,) if is_tf_available() else () lowerCamelCase__ : Optional[Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Optional[int] = BlipTextModelTester(self ) lowercase__ : int = ConfigTester(self , config_class=a , hidden_size=3_7 ) def _UpperCAmelCase ( self ) -> Tuple: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> int: lowercase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _UpperCAmelCase ( self ) -> Optional[Any]: pass def _UpperCAmelCase ( self ) -> Optional[int]: pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def _UpperCAmelCase ( self ) -> List[str]: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> Dict: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> str: pass @slow def _UpperCAmelCase ( self ) -> int: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Any = TFBlipTextModel.from_pretrained(a ) self.assertIsNotNone(a ) def _UpperCAmelCase ( self , a=True ) -> List[str]: super().test_pt_tf_model_equivalence(allow_missing_keys=a )
77
1
"""simple docstring""" import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def a_ ( ): '''simple docstring''' print('Making key files...' ) make_key_files('rsa' , 1024 ) print('Key files generation successful.' ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' print('Generating prime p...' ) lowercase__ : Dict = rabinMiller.generate_large_prime(_lowerCAmelCase ) print('Generating prime q...' ) lowercase__ : List[str] = rabinMiller.generate_large_prime(_lowerCAmelCase ) lowercase__ : Tuple = p * q print('Generating e that is relatively prime to (p - 1) * (q - 1)...' ) while True: lowercase__ : Dict = random.randrange(2 ** (key_size - 1) , 2 ** (key_size) ) if cryptoMath.gcd(_lowerCAmelCase , (p - 1) * (q - 1) ) == 1: break print('Calculating d that is mod inverse of e...' ) lowercase__ : Tuple = cryptoMath.find_mod_inverse(_lowerCAmelCase , (p - 1) * (q - 1) ) lowercase__ : Dict = (n, e) lowercase__ : str = (n, d) return (public_key, private_key) def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : int ): '''simple docstring''' if os.path.exists(f"""{name}_pubkey.txt""" ) or os.path.exists(f"""{name}_privkey.txt""" ): print('\nWARNING:' ) print( f"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" 'Use a different name or delete these files and re-run this program.' ) sys.exit() lowercase__ , lowercase__ : int = generate_key(_lowerCAmelCase ) print(f"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(f"""{name}_pubkey.txt""" , 'w' ) as out_file: out_file.write(f"""{key_size},{public_key[0]},{public_key[1]}""" ) print(f"""Writing private key to file {name}_privkey.txt...""" ) with open(f"""{name}_privkey.txt""" , 'w' ) as out_file: out_file.write(f"""{key_size},{private_key[0]},{private_key[1]}""" ) if __name__ == "__main__": main()
77
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase : Tuple = logging.get_logger(__name__) _UpperCamelCase : Union[str, Any] = { "alibaba-damo/mgp-str-base": "https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json", } class UpperCAmelCase_ ( _a): lowerCamelCase__ : List[Any] = "mgp-str" def __init__( self , a=[3_2, 1_2_8] , a=4 , a=3 , a=2_7 , a=3_8 , a=5_0_2_5_7 , a=3_0_5_2_2 , a=7_6_8 , a=1_2 , a=1_2 , a=4.0 , a=True , a=False , a=1e-5 , a=0.0 , a=0.0 , a=0.0 , a=False , a=0.02 , **a , ) -> Tuple: super().__init__(**a ) lowercase__ : int = image_size lowercase__ : List[Any] = patch_size lowercase__ : Optional[Any] = num_channels lowercase__ : Optional[Any] = max_token_length lowercase__ : Dict = num_character_labels lowercase__ : Optional[int] = num_bpe_labels lowercase__ : Dict = num_wordpiece_labels lowercase__ : Tuple = hidden_size lowercase__ : List[Any] = num_hidden_layers lowercase__ : Union[str, Any] = num_attention_heads lowercase__ : Dict = mlp_ratio lowercase__ : Optional[int] = distilled lowercase__ : Optional[int] = layer_norm_eps lowercase__ : Optional[int] = drop_rate lowercase__ : List[str] = qkv_bias lowercase__ : Optional[int] = attn_drop_rate lowercase__ : Any = drop_path_rate lowercase__ : List[Any] = output_aa_attentions lowercase__ : Tuple = initializer_range
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase : Optional[Any] = logging.get_logger(__name__) _UpperCamelCase : Optional[int] = { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/config.json", } class UpperCAmelCase_ ( _a): lowerCamelCase__ : Dict = "gpt_neox_japanese" def __init__( self , a=3_2_0_0_0 , a=2_5_6_0 , a=3_2 , a=3_2 , a=4 , a="gelu" , a=1.00 , a=1_0_0_0_0 , a=2_0_4_8 , a=0.02 , a=1e-5 , a=True , a=3_1_9_9_6 , a=3_1_9_9_9 , a=0.1 , a=0.0 , **a , ) -> Optional[Any]: super().__init__(bos_token_id=a , eos_token_id=a , **a ) lowercase__ : Optional[int] = vocab_size lowercase__ : int = max_position_embeddings lowercase__ : Optional[Any] = hidden_size lowercase__ : Optional[Any] = num_hidden_layers lowercase__ : Optional[int] = num_attention_heads lowercase__ : Optional[int] = intermediate_multiple_size lowercase__ : Dict = hidden_act lowercase__ : Union[str, Any] = rotary_pct lowercase__ : Tuple = rotary_emb_base lowercase__ : Dict = initializer_range lowercase__ : List[str] = layer_norm_eps lowercase__ : List[Any] = use_cache lowercase__ : Dict = attention_dropout lowercase__ : Tuple = hidden_dropout
77
"""simple docstring""" import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : int = FunnelConfig.from_json_file(_lowerCAmelCase ) print(f"""Building PyTorch model from configuration: {config}""" ) lowercase__ : List[Any] = FunnelBaseModel(_lowerCAmelCase ) if base_model else FunnelModel(_lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_funnel(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _lowerCAmelCase ) if __name__ == "__main__": _UpperCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--base_model", action="store_true", help="Whether you want just the base model (no decoder) or not." ) _UpperCamelCase : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
77
1
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ : def __init__( self , a , a=1_2 , a=7 , a=True , a=True , a=True , a=9_9 , a=3_2 , a=3_2 , a=2 , a=4 , a=3_7 , a=0.1 , a=0.1 , a=5_1_2 , a=0.02 , a=0 , a=None , ) -> Union[str, Any]: lowercase__ : Any = parent lowercase__ : str = batch_size lowercase__ : List[Any] = seq_length lowercase__ : Union[str, Any] = is_training lowercase__ : List[str] = use_input_mask lowercase__ : int = use_labels lowercase__ : List[Any] = vocab_size lowercase__ : str = hidden_size lowercase__ : int = projection_dim lowercase__ : Optional[int] = num_hidden_layers lowercase__ : Any = num_attention_heads lowercase__ : Optional[Any] = intermediate_size lowercase__ : Optional[Any] = dropout lowercase__ : Optional[int] = attention_dropout lowercase__ : Optional[int] = max_position_embeddings lowercase__ : str = initializer_range lowercase__ : Tuple = scope lowercase__ : int = bos_token_id def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : int = None if self.use_input_mask: lowercase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowercase__ : int = input_mask.numpy() lowercase__ , lowercase__ : Tuple = input_mask.shape lowercase__ : List[str] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a ): lowercase__ : Dict = 1 lowercase__ : Union[str, Any] = 0 lowercase__ : Tuple = self.get_config() return config, input_ids, tf.convert_to_tensor(a ) def _UpperCAmelCase ( self ) -> List[Any]: return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _UpperCAmelCase ( self , a , a , a ) -> Any: lowercase__ : List[Any] = TFBlipTextModel(config=a ) lowercase__ : Optional[int] = model(a , attention_mask=a , training=a ) lowercase__ : List[str] = model(a , training=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _UpperCAmelCase ( self ) -> Any: lowercase__ : Optional[Any] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ : Any = config_and_inputs lowercase__ : Any = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : Dict = (TFBlipTextModel,) if is_tf_available() else () lowerCamelCase__ : Optional[Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Optional[int] = BlipTextModelTester(self ) lowercase__ : int = ConfigTester(self , config_class=a , hidden_size=3_7 ) def _UpperCAmelCase ( self ) -> Tuple: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> int: lowercase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _UpperCAmelCase ( self ) -> Optional[Any]: pass def _UpperCAmelCase ( self ) -> Optional[int]: pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def _UpperCAmelCase ( self ) -> List[str]: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> Dict: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> str: pass @slow def _UpperCAmelCase ( self ) -> int: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Any = TFBlipTextModel.from_pretrained(a ) self.assertIsNotNone(a ) def _UpperCAmelCase ( self , a=True ) -> List[str]: super().test_pt_tf_model_equivalence(allow_missing_keys=a )
77
"""simple docstring""" import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _UpperCamelCase : Any = logging.getLogger(__name__) _UpperCamelCase : Optional[Any] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _UpperCamelCase : Any = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The model checkpoint for weights initialization. Leave None if you want to train a model from" " scratch." ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(_a)} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained config name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "The input training data file (a text file)."}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The input training data files (multiple files in glob format). " "Very often splitting large files to smaller files can prevent tokenizer going out of memory" ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input train ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input eval ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Whether distinct lines of text in the dataset are to be handled as distinct sequences."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Train with masked-language modeling loss instead of language modeling."}) lowerCamelCase__ : bool = field(default=_a , metadata={"help": "Whether ot not to use whole word mask."}) lowerCamelCase__ : float = field( default=0.15 , metadata={"help": "Ratio of tokens to mask for masked language modeling loss"}) lowerCamelCase__ : float = field( default=1 / 6 , metadata={ "help": ( "Ratio of length of a span of masked tokens to surrounding context length for permutation language" " modeling." ) } , ) lowerCamelCase__ : int = field( default=5 , metadata={"help": "Maximum length of a span of masked tokens for permutation language modeling."}) lowerCamelCase__ : int = field( default=-1 , metadata={ "help": ( "Optional input sequence length after tokenization." "The training dataset will be truncated in block of this size for training." "Default to the model max input length for single sentence inputs (take into account special tokens)." ) } , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Overwrite the cached training and evaluation sets"}) def a_ ( _lowerCAmelCase : DataTrainingArguments , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[str] = None , ): '''simple docstring''' def _dataset(_lowerCAmelCase : Any , _lowerCAmelCase : Any=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('You need to set world whole masking and mlm to True for Chinese Whole Word Mask' ) return LineByLineWithRefDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , ref_path=_lowerCAmelCase , ) return LineByLineTextDataset(tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size ) else: return TextDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=_lowerCAmelCase , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(_lowerCAmelCase ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def a_ ( ): '''simple docstring''' lowercase__ : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ : List[Any] = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( 'Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ' 'or remove the --do_eval argument.' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s' , _lowerCAmelCase ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: lowercase__ : List[Any] = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : Any = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: lowercase__ : Union[str, Any] = CONFIG_MAPPING[model_args.model_type]() logger.warning('You are instantiating a new config instance from scratch.' ) if model_args.tokenizer_name: lowercase__ : Dict = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : int = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another' ' script, save it,and load it from here, using --tokenizer_name' ) if model_args.model_name_or_path: lowercase__ : Optional[Any] = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=_lowerCAmelCase , cache_dir=model_args.cache_dir , ) else: logger.info('Training new model from scratch' ) lowercase__ : int = AutoModelWithLMHead.from_config(_lowerCAmelCase ) model.resize_token_embeddings(len(_lowerCAmelCase ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( 'BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the' '--mlm flag (masked language modeling).' ) if data_args.block_size <= 0: lowercase__ : List[str] = tokenizer.max_len # Our input block size will be the max possible for the model else: lowercase__ : int = min(data_args.block_size , tokenizer.max_len ) # Get datasets lowercase__ : Tuple = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) lowercase__ : Optional[Any] = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , evaluate=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": lowercase__ : List[Any] = DataCollatorForPermutationLanguageModeling( tokenizer=_lowerCAmelCase , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: lowercase__ : List[str] = DataCollatorForWholeWordMask( tokenizer=_lowerCAmelCase , mlm_probability=data_args.mlm_probability ) else: lowercase__ : str = DataCollatorForLanguageModeling( tokenizer=_lowerCAmelCase , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer lowercase__ : Optional[int] = Trainer( model=_lowerCAmelCase , args=_lowerCAmelCase , data_collator=_lowerCAmelCase , train_dataset=_lowerCAmelCase , eval_dataset=_lowerCAmelCase , prediction_loss_only=_lowerCAmelCase , ) # Training if training_args.do_train: lowercase__ : List[str] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=_lowerCAmelCase ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase__ : List[Any] = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) lowercase__ : Dict = trainer.evaluate() lowercase__ : List[Any] = math.exp(eval_output['eval_loss'] ) lowercase__ : int = {'perplexity': perplexity} lowercase__ : int = os.path.join(training_args.output_dir , 'eval_results_lm.txt' ) if trainer.is_world_master(): with open(_lowerCAmelCase , 'w' ) as writer: logger.info('***** Eval results *****' ) for key in sorted(result.keys() ): logger.info(' %s = %s' , _lowerCAmelCase , str(result[key] ) ) writer.write('%s = %s\n' % (key, str(result[key] )) ) results.update(_lowerCAmelCase ) return results def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' main() if __name__ == "__main__": main()
77
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCamelCase : Any = logging.get_logger(__name__) _UpperCamelCase : Tuple = { "google/mobilenet_v2_1.4_224": "https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json", "google/mobilenet_v2_1.0_224": "https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json", "google/mobilenet_v2_0.75_160": "https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json", "google/mobilenet_v2_0.35_96": "https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json", # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = "mobilenet_v2" def __init__( self , a=3 , a=2_2_4 , a=1.0 , a=8 , a=8 , a=6 , a=3_2 , a=True , a=True , a="relu6" , a=True , a=0.8 , a=0.02 , a=0.001 , a=2_5_5 , **a , ) -> Any: super().__init__(**a ) if depth_multiplier <= 0: raise ValueError('depth_multiplier must be greater than zero.' ) lowercase__ : Union[str, Any] = num_channels lowercase__ : str = image_size lowercase__ : List[Any] = depth_multiplier lowercase__ : int = depth_divisible_by lowercase__ : Any = min_depth lowercase__ : str = expand_ratio lowercase__ : Optional[Any] = output_stride lowercase__ : List[str] = first_layer_is_expansion lowercase__ : Union[str, Any] = finegrained_output lowercase__ : Any = hidden_act lowercase__ : Optional[Any] = tf_padding lowercase__ : Optional[int] = classifier_dropout_prob lowercase__ : Union[str, Any] = initializer_range lowercase__ : Any = layer_norm_eps lowercase__ : Tuple = semantic_loss_ignore_index class UpperCAmelCase_ ( _a): lowerCamelCase__ : Optional[Any] = version.parse("1.11") @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict([('pixel_values', {0: 'batch'})] ) @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "image-classification": return OrderedDict([('logits', {0: 'batch'})] ) else: return OrderedDict([('last_hidden_state', {0: 'batch'}), ('pooler_output', {0: 'batch'})] ) @property def _UpperCAmelCase ( self ) -> float: return 1e-4
77
"""simple docstring""" import math import flax.linen as nn import jax.numpy as jnp def a_ ( _lowerCAmelCase : jnp.ndarray , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1.0E4 , _lowerCAmelCase : bool = False , _lowerCAmelCase : float = 1.0 , ): '''simple docstring''' assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, f"""Embedding dimension {embedding_dim} should be even""" lowercase__ : Optional[Any] = float(embedding_dim // 2 ) lowercase__ : Optional[Any] = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) lowercase__ : Any = min_timescale * jnp.exp(jnp.arange(_lowerCAmelCase , dtype=jnp.floataa ) * -log_timescale_increment ) lowercase__ : Dict = jnp.expand_dims(_lowerCAmelCase , 1 ) * jnp.expand_dims(_lowerCAmelCase , 0 ) # scale embeddings lowercase__ : List[str] = scale * emb if flip_sin_to_cos: lowercase__ : Dict = jnp.concatenate([jnp.cos(_lowerCAmelCase ), jnp.sin(_lowerCAmelCase )] , axis=1 ) else: lowercase__ : Optional[int] = jnp.concatenate([jnp.sin(_lowerCAmelCase ), jnp.cos(_lowerCAmelCase )] , axis=1 ) lowercase__ : List[Any] = jnp.reshape(_lowerCAmelCase , [jnp.shape(_lowerCAmelCase )[0], embedding_dim] ) return signal class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : jnp.dtype = jnp.floataa @nn.compact def __call__( self , a ) -> Any: lowercase__ : str = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_1' )(a ) lowercase__ : Union[str, Any] = nn.silu(a ) lowercase__ : Optional[Any] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_2' )(a ) return temb class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : bool = False lowerCamelCase__ : float = 1 @nn.compact def __call__( self , a ) -> str: return get_sinusoidal_embeddings( a , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' return sum(e for e in range(3 , _lowerCAmelCase ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(f'''{solution() = }''')
77
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def a_ ( _lowerCAmelCase : Callable[[int | float], int | float] , _lowerCAmelCase : int | float , _lowerCAmelCase : int | float , _lowerCAmelCase : int = 100 , ): '''simple docstring''' lowercase__ : Dict = x_start lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) lowercase__ : Optional[Any] = 0.0 for _ in range(_lowerCAmelCase ): # Approximates curve as a sequence of linear lines and sums their length lowercase__ : Union[str, Any] = (x_end - x_start) / steps + xa lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step lowercase__ : Union[str, Any] = xa lowercase__ : int = fxa return length if __name__ == "__main__": def a_ ( _lowerCAmelCase : List[Any] ): '''simple docstring''' return math.sin(10 * x ) print("f(x) = sin(10 * x)") print("The length of the curve from x = -10 to x = 10 is:") _UpperCamelCase : str = 10 while i <= 10_00_00: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
77
1
"""simple docstring""" def a_ ( ): '''simple docstring''' return [list(range(1000 - i , -1000 - i , -1 ) ) for i in range(1000 )] _UpperCamelCase : Optional[int] = generate_large_matrix() _UpperCamelCase : int = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def a_ ( _lowerCAmelCase : list[list[int]] ): '''simple docstring''' assert all(row == sorted(_lowerCAmelCase , reverse=_lowerCAmelCase ) for row in grid ) assert all(list(_lowerCAmelCase ) == sorted(_lowerCAmelCase , reverse=_lowerCAmelCase ) for col in zip(*_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : list[int] ): '''simple docstring''' lowercase__ : List[Any] = 0 lowercase__ : List[Any] = len(_lowerCAmelCase ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: lowercase__ : Optional[Any] = (left + right) // 2 lowercase__ : Optional[Any] = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: lowercase__ : str = mid + 1 else: lowercase__ : int = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : list[list[int]] ): '''simple docstring''' lowercase__ : Optional[int] = 0 lowercase__ : Dict = len(grid[0] ) for i in range(len(_lowerCAmelCase ) ): lowercase__ : List[Any] = find_negative_index(grid[i][:bound] ) total += bound return (len(_lowerCAmelCase ) * len(grid[0] )) - total def a_ ( _lowerCAmelCase : list[list[int]] ): '''simple docstring''' return len([number for row in grid for number in row if number < 0] ) def a_ ( _lowerCAmelCase : list[list[int]] ): '''simple docstring''' lowercase__ : Tuple = 0 for row in grid: for i, number in enumerate(_lowerCAmelCase ): if number < 0: total += len(_lowerCAmelCase ) - i break return total def a_ ( ): '''simple docstring''' from timeit import timeit print('Running benchmarks' ) lowercase__ : Union[str, Any] = ( 'from __main__ import count_negatives_binary_search, ' 'count_negatives_brute_force, count_negatives_brute_force_with_break, grid' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): lowercase__ : Any = timeit(f"""{func}(grid=grid)""" , setup=_lowerCAmelCase , number=500 ) print(f"""{func}() took {time:0.4f} seconds""" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase : Tuple = { "configuration_whisper": ["WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP", "WhisperConfig", "WhisperOnnxConfig"], "feature_extraction_whisper": ["WhisperFeatureExtractor"], "processing_whisper": ["WhisperProcessor"], "tokenization_whisper": ["WhisperTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = ["WhisperTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : str = [ "WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "WhisperForConditionalGeneration", "WhisperModel", "WhisperPreTrainedModel", "WhisperForAudioClassification", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Any = [ "TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFWhisperForConditionalGeneration", "TFWhisperModel", "TFWhisperPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Tuple = [ "FlaxWhisperForConditionalGeneration", "FlaxWhisperModel", "FlaxWhisperPreTrainedModel", "FlaxWhisperForAudioClassification", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _UpperCamelCase : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax _UpperCamelCase : Optional[int] = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , **a ) -> int: super().__init__(**a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING ) def __call__( self , a , **a ) -> str: return super().__call__(a , **a ) def _UpperCAmelCase ( self , **a ) -> Any: lowercase__ : str = {} if "candidate_labels" in kwargs: lowercase__ : List[str] = kwargs['candidate_labels'] if "hypothesis_template" in kwargs: lowercase__ : Optional[Any] = kwargs['hypothesis_template'] return preprocess_params, {}, {} def _UpperCAmelCase ( self , a , a=None , a="This is a photo of {}." ) -> Tuple: lowercase__ : str = load_image(a ) lowercase__ : List[Any] = self.image_processor(images=[image] , return_tensors=self.framework ) lowercase__ : Dict = candidate_labels lowercase__ : Union[str, Any] = [hypothesis_template.format(a ) for x in candidate_labels] lowercase__ : Any = self.tokenizer(a , return_tensors=self.framework , padding=a ) lowercase__ : Dict = [text_inputs] return inputs def _UpperCAmelCase ( self , a ) -> str: lowercase__ : Dict = model_inputs.pop('candidate_labels' ) lowercase__ : int = model_inputs.pop('text_inputs' ) if isinstance(text_inputs[0] , a ): lowercase__ : str = text_inputs[0] else: # Batching case. lowercase__ : Dict = text_inputs[0][0] lowercase__ : Optional[Any] = self.model(**a , **a ) lowercase__ : str = { 'candidate_labels': candidate_labels, 'logits': outputs.logits_per_image, } return model_outputs def _UpperCAmelCase ( self , a ) -> Optional[int]: lowercase__ : Optional[int] = model_outputs.pop('candidate_labels' ) lowercase__ : List[str] = model_outputs['logits'][0] if self.framework == "pt": lowercase__ : str = logits.softmax(dim=-1 ).squeeze(-1 ) lowercase__ : int = probs.tolist() if not isinstance(a , a ): lowercase__ : List[str] = [scores] elif self.framework == "tf": lowercase__ : Optional[Any] = stable_softmax(a , axis=-1 ) lowercase__ : Union[str, Any] = probs.numpy().tolist() else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) lowercase__ : Tuple = [ {'score': score, 'label': candidate_label} for score, candidate_label in sorted(zip(a , a ) , key=lambda a : -x[0] ) ] return result
77
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a_ ( _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : int = args.pruning_method lowercase__ : Tuple = args.threshold lowercase__ : str = args.model_name_or_path.rstrip('/' ) lowercase__ : List[Any] = args.target_model_path print(f"""Load fine-pruned model from {model_name_or_path}""" ) lowercase__ : Optional[Any] = torch.load(os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) lowercase__ : List[str] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowercase__ : Tuple = tensor print(f"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowercase__ : List[str] = tensor print(f"""Copied layer {name}""" ) elif "bias" in name: lowercase__ : Optional[Any] = tensor print(f"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowercase__ : Optional[Any] = MagnitudeBinarizer.apply(inputs=_lowerCAmelCase , threshold=_lowerCAmelCase ) lowercase__ : Optional[int] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowercase__ : Optional[Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ : Any = TopKBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[Any] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowercase__ : Any = name[:-6] lowercase__ : Optional[Any] = model[f"""{prefix_}mask_scores"""] lowercase__ : Tuple = ThresholdBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[str] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowercase__ : Union[str, Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ , lowercase__ : Tuple = -0.1, 1.1 lowercase__ : Optional[Any] = torch.sigmoid(_lowerCAmelCase ) lowercase__ : Optional[Any] = s * (r - l) + l lowercase__ : Optional[Any] = s_bar.clamp(min=0.0 , max=1.0 ) lowercase__ : Union[str, Any] = tensor * mask print(f"""Pruned layer {name}""" ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: lowercase__ : Union[str, Any] = os.path.join( os.path.dirname(_lowerCAmelCase ) , f"""bertarized_{os.path.basename(_lowerCAmelCase )}""" ) if not os.path.isdir(_lowerCAmelCase ): shutil.copytree(_lowerCAmelCase , _lowerCAmelCase ) print(f"""\nCreated folder {target_model_path}""" ) torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": _UpperCamelCase : int = argparse.ArgumentParser() parser.add_argument( "--pruning_method", choices=["l0", "magnitude", "topK", "sigmoied_threshold"], type=str, required=True, help=( "Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning," " sigmoied_threshold = Soft movement pruning)" ), ) parser.add_argument( "--threshold", type=float, required=False, help=( "For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model." "For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared." "Not needed for `l0`" ), ) parser.add_argument( "--model_name_or_path", type=str, required=True, help="Folder containing the model that was previously fine-pruned", ) parser.add_argument( "--target_model_path", default=None, type=str, required=False, help="Folder containing the model that was previously fine-pruned", ) _UpperCamelCase : Dict = parser.parse_args() main(args)
77
1
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _UpperCamelCase : int = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : Optional[int] = ReformerTokenizer lowerCamelCase__ : Tuple = ReformerTokenizerFast lowerCamelCase__ : Optional[int] = True lowerCamelCase__ : List[Any] = False lowerCamelCase__ : Union[str, Any] = True def _UpperCAmelCase ( self ) -> List[Any]: super().setUp() lowercase__ : Union[str, Any] = ReformerTokenizer(a , keep_accents=a ) tokenizer.save_pretrained(self.tmpdirname ) def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : Optional[Any] = '<s>' lowercase__ : Any = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a ) , a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a ) , a ) def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : List[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , 'j' ) self.assertEqual(len(a ) , 1_0_0_0 ) def _UpperCAmelCase ( self ) -> Union[str, Any]: self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_0 ) def _UpperCAmelCase ( self ) -> int: if not self.test_rust_tokenizer: return lowercase__ : List[Any] = self.get_tokenizer() lowercase__ : Dict = self.get_rust_tokenizer() lowercase__ : List[Any] = 'I was born in 92000, and this is falsé.' lowercase__ : Dict = tokenizer.tokenize(a ) lowercase__ : Tuple = rust_tokenizer.tokenize(a ) self.assertListEqual(a , a ) lowercase__ : str = tokenizer.encode(a , add_special_tokens=a ) lowercase__ : Tuple = rust_tokenizer.encode(a , add_special_tokens=a ) self.assertListEqual(a , a ) lowercase__ : Any = self.get_rust_tokenizer() lowercase__ : Optional[int] = tokenizer.encode(a ) lowercase__ : Tuple = rust_tokenizer.encode(a ) self.assertListEqual(a , a ) def _UpperCAmelCase ( self , a=1_5 ) -> str: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowercase__ : Tuple = self.rust_tokenizer_class.from_pretrained(a , **a ) # Simple input lowercase__ : Tuple = 'This is a simple input' lowercase__ : List[str] = ['This is a simple input 1', 'This is a simple input 2'] lowercase__ : Optional[int] = ('This is a simple input', 'This is a pair') lowercase__ : List[Any] = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(a , tokenizer_r.encode , a , max_length=a , padding='max_length' ) # Simple input self.assertRaises(a , tokenizer_r.encode_plus , a , max_length=a , padding='max_length' ) # Simple input self.assertRaises( a , tokenizer_r.batch_encode_plus , a , max_length=a , padding='max_length' , ) # Pair input self.assertRaises(a , tokenizer_r.encode , a , max_length=a , padding='max_length' ) # Pair input self.assertRaises(a , tokenizer_r.encode_plus , a , max_length=a , padding='max_length' ) # Pair input self.assertRaises( a , tokenizer_r.batch_encode_plus , a , max_length=a , padding='max_length' , ) def _UpperCAmelCase ( self ) -> Optional[int]: pass def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Union[str, Any] = ReformerTokenizer(a , keep_accents=a ) lowercase__ : str = tokenizer.tokenize('This is a test' ) self.assertListEqual(a , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a ) , [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2] , ) lowercase__ : int = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( a , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowercase__ : int = tokenizer.convert_tokens_to_ids(a ) self.assertListEqual( a , [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 0, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 0, 4] , ) lowercase__ : List[str] = tokenizer.convert_ids_to_tokens(a ) self.assertListEqual( a , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def _UpperCAmelCase ( self ) -> Optional[int]: return ReformerTokenizer.from_pretrained('google/reformer-crime-and-punishment' ) @slow def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : Union[str, Any] = 'Hello World!' lowercase__ : Optional[int] = [1_2_6, 3_2, 2_6_2, 1_5_2, 3_8, 7_2, 2_8_7] self.assertListEqual(a , self.big_tokenizer.encode(a ) ) @slow def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : List[str] = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) lowercase__ : int = [ 1_0_8, 2_6_5, 2_4, 1_1_1, 4, 2_5_8, 1_5_6, 3_5, 2_8, 2_7_5, 3, 2_5_9, 2_9_7, 2_6_0, 8_4, 4, 3_5, 1_1_0, 4_4, 8, 2_5_9, 9_1, 2_6_8, 2_1, 1_1, 2_0_9, 2_7_4, 1_0_9, 2_6_6, 2_7_7, 1_1_7, 8_6, 9_3, 3_1_5, 2_5_8, 2_7_8, 2_5_8, 2_7_7, 2_5_8, 0, 2_5_8, 2_8_8, 2_5_8, 3_1_9, 2_5_8, 0, 2_5_8, 0, 2_5_8, 0, 2_5_8, 0, 2_5_8, 2_8_7, 2_5_8, 3_1_5, 2_5_8, 2_8_9, 2_5_8, 2_7_8, 9_9, 2_6_9, 2_6_6, 2_6_2, 8, 2_5_9, 2_4_1, 4, 2_1_7, 2_3_0, 2_6_8, 2_6_6, 5_5, 1_6_8, 1_0_6, 7_5, 1_9_3, 2_6_6, 2_2_3, 2_7, 4_9, 2_6, 2_8_2, 2_5, 2_6_4, 2_9_9, 1_9, 2_6, 0, 2_5_8, 2_7_7, 1_1_7, 8_6, 9_3, 1_7_6, 1_8_3, 2_7_0, 1_1, 2_6_2, 4_2, 6_1, 2_6_5, ] self.assertListEqual(a , self.big_tokenizer.encode(a ) ) @require_torch @slow def _UpperCAmelCase ( self ) -> List[Any]: import torch from transformers import ReformerConfig, ReformerModel # Build sequence lowercase__ : str = list(self.big_tokenizer.get_vocab().keys() )[:1_0] lowercase__ : List[Any] = ' '.join(a ) lowercase__ : List[Any] = self.big_tokenizer.encode_plus(a , return_tensors='pt' ) lowercase__ : Optional[int] = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='pt' ) lowercase__ : Optional[int] = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) lowercase__ : Dict = encoded_sequence['input_ids'].shape lowercase__ : Dict = ReformerModel(a ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**a ) model(**a ) @slow def _UpperCAmelCase ( self ) -> Optional[Any]: # fmt: off lowercase__ : Optional[Any] = {'input_ids': [[1_0_8, 2_6_5, 2_4, 1_1_1, 4, 2_5_8, 1_5_6, 7, 5_1, 2_7_9, 5_8, 7, 7_6, 2_5, 6_9, 2_7_8], [1_4_0, 2_4_3, 2_6_4, 1_3_4, 1_7, 2_6_7, 7_7, 2_6_3, 2_2, 2_6_2, 2_9_7, 2_5_8, 3_0_4, 1_7_7, 2_7_9, 2_6_6, 1_4, 8_9, 1_3, 3_5, 2_6_1, 2_9_9, 2_7_2, 1_3_7, 2_7_5, 2_7_8]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 lowercase__ : List[Any] = [ 'This is a very simple sentence.', 'The quick brown fox jumps over the lazy dog.', ] self.tokenizer_integration_test_util( expected_encoding=a , model_name='google/reformer-crime-and-punishment' , revision='0e6c3decb8211d49bf881013425dc8b0448b3f5a' , padding=a , sequences=a , )
77
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[List[PIL.Image.Image], np.ndarray] lowerCamelCase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : np.ndarray lowerCamelCase__ : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
77
1
"""simple docstring""" import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() _UpperCamelCase : Optional[int] = logging.get_logger(__name__) _UpperCamelCase : str = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } _UpperCamelCase : Optional[Any] = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", ] def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Dict , _lowerCAmelCase : int , _lowerCAmelCase : List[str] ): '''simple docstring''' for attribute in key.split('.' ): lowercase__ : Any = getattr(_lowerCAmelCase , _lowerCAmelCase ) if weight_type is not None: lowercase__ : str = getattr(_lowerCAmelCase , _lowerCAmelCase ).shape else: lowercase__ : Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": lowercase__ : Optional[int] = value elif weight_type == "weight_g": lowercase__ : int = value elif weight_type == "weight_v": lowercase__ : Optional[Any] = value elif weight_type == "bias": lowercase__ : int = value else: lowercase__ : Dict = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : str = [] lowercase__ : int = fairseq_model.state_dict() lowercase__ : Union[str, Any] = hf_model.feature_extractor lowercase__ : Dict = hf_model.adapter for name, value in fairseq_dict.items(): lowercase__ : int = False if "conv_layers" in name: load_conv_layer( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) lowercase__ : Tuple = True elif any(x in name for x in ['adaptor', 'w2v_encoder.proj.', 'w2v_proj_ln.'] ): load_adapter(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Dict = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: lowercase__ : Optional[Any] = True if "*" in mapped_key: lowercase__ : str = name.split(_lowerCAmelCase )[0].split('.' )[-2] lowercase__ : Tuple = mapped_key.replace('*' , _lowerCAmelCase ) if "weight_g" in name: lowercase__ : Tuple = 'weight_g' elif "weight_v" in name: lowercase__ : Any = 'weight_v' elif "bias" in name: lowercase__ : Tuple = 'bias' elif "weight" in name: lowercase__ : List[str] = 'weight' else: lowercase__ : str = None set_recursively(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) continue if not is_used: unused_weights.append(_lowerCAmelCase ) logger.warning(f"""Unused weights: {unused_weights}""" ) def a_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Union[str, Any] = full_name.split('conv_layers.' )[-1] lowercase__ : str = name.split('.' ) lowercase__ : Union[str, Any] = int(items[0] ) lowercase__ : List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) lowercase__ : Tuple = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) lowercase__ : Optional[int] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) lowercase__ : Dict = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) lowercase__ : int = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] ): '''simple docstring''' lowercase__ : Tuple = full_name.split('adaptor.' )[-1] lowercase__ : Dict = name.split('.' ) if items[1].isdigit(): lowercase__ : Dict = int(items[1] ) else: lowercase__ : Union[str, Any] = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), f"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.""" lowercase__ : Any = value logger.info(f"""Adapter proj layer norm bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), f"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.""" lowercase__ : List[Any] = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), f"""{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.""" lowercase__ : Optional[Any] = value logger.info(f"""Adapter proj layer bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), f"""{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.""" lowercase__ : Union[str, Any] = value logger.info(f"""Adapter proj layer weight was initialized from {full_name}.""" ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), f"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.""" lowercase__ : int = value logger.info(f"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), f"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.""" lowercase__ : Dict = value logger.info(f"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' lowercase__ , lowercase__ : Union[str, Any] = emb.weight.shape lowercase__ : str = nn.Linear(_lowerCAmelCase , _lowerCAmelCase , bias=_lowerCAmelCase ) lowercase__ : Optional[int] = emb.weight.data return lin_layer @torch.no_grad() def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , ): '''simple docstring''' lowercase__ : Optional[int] = WavaVecaConfig.from_pretrained( _lowerCAmelCase , add_adapter=_lowerCAmelCase , adapter_stride=_lowerCAmelCase , adapter_kernel_size=_lowerCAmelCase , use_auth_token=_lowerCAmelCase , output_hidden_size=_lowerCAmelCase , ) lowercase__ : List[str] = MBartConfig.from_pretrained(_lowerCAmelCase ) # load model lowercase__ , lowercase__ , lowercase__ : Dict = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={ 'config_yaml': config_yaml_path, 'data': '/'.join(dict_path.split('/' )[:-1] ), 'w2v_path': checkpoint_path, 'load_pretrained_decoder_from': None, } , ) lowercase__ : List[str] = model[0].eval() # load feature extractor lowercase__ : Tuple = WavaVecaFeatureExtractor.from_pretrained(_lowerCAmelCase , use_auth_token=_lowerCAmelCase ) # set weights for wav2vec2 encoder lowercase__ : int = WavaVecaModel(_lowerCAmelCase ) recursively_load_weights_wavaveca(model.encoder , _lowerCAmelCase ) # load decoder weights lowercase__ : List[Any] = MBartForCausalLM(_lowerCAmelCase ) lowercase__ , lowercase__ : Union[str, Any] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=_lowerCAmelCase ) logger.warning(f"""The following keys are missing when loading the decoder weights: {missing_keys}""" ) logger.warning(f"""The following keys are unexpected when loading the decoder weights: {unexpected_keys}""" ) lowercase__ : Optional[int] = SpeechEncoderDecoderModel(encoder=_lowerCAmelCase , decoder=_lowerCAmelCase ) lowercase__ : Dict = False lowercase__ : Any = MBartaaTokenizer(_lowerCAmelCase ) tokenizer.save_pretrained(_lowerCAmelCase ) lowercase__ : str = hf_wavavec.config.to_dict() lowercase__ : Dict = tokenizer.pad_token_id lowercase__ : str = tokenizer.bos_token_id lowercase__ : Tuple = tokenizer.eos_token_id lowercase__ : Any = 'mbart50' lowercase__ : Union[str, Any] = 'wav2vec2' lowercase__ : Optional[int] = tokenizer.eos_token_id lowercase__ : Union[str, Any] = 25_0004 lowercase__ : int = tokenizer.eos_token_id lowercase__ : str = SpeechEncoderDecoderConfig.from_dict(_lowerCAmelCase ) hf_wavavec.save_pretrained(_lowerCAmelCase ) feature_extractor.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": _UpperCamelCase : Dict = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_yaml_path", default=None, type=str, help="Path to yaml file of fine-tuned model") parser.add_argument( "--encoder_config_path", default="facebook/wav2vec2-xls-r-1b", type=str, help="Path to hf encoder wav2vec2 checkpoint config", ) parser.add_argument( "--decoder_config_path", default="facebook/mbart-large-50-one-to-many-mmt", type=str, help="Path to hf decoder checkpoint config", ) parser.add_argument("--add_adapter", default=True, type=bool, help="whethere to add model adapter layers") parser.add_argument("--adapter_stride", default=2, type=int, help="stride of adapter layers") parser.add_argument("--adapter_kernel_size", default=3, type=int, help="kernel size of adapter layers") parser.add_argument("--encoder_output_dim", default=10_24, type=int, help="encoder output dim") parser.add_argument("--start_token_id", default=25_00_04, type=int, help="`decoder_start_token_id` of model config") _UpperCamelCase : Any = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
77
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class UpperCAmelCase_ : def __init__( self , a , a , a , a , a , a=0.2 , a=0.2 ) -> Dict: lowercase__ : Any = bp_numa lowercase__ : Optional[int] = bp_numa lowercase__ : Tuple = bp_numa lowercase__ : Optional[Any] = conva_get[:2] lowercase__ : Optional[int] = conva_get[2] lowercase__ : Optional[Any] = size_pa lowercase__ : Union[str, Any] = rate_w lowercase__ : Union[str, Any] = rate_t lowercase__ : List[Any] = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ : Optional[Any] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : int = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : Any = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 def _UpperCAmelCase ( self , a ) -> Union[str, Any]: # save model dict with pickle lowercase__ : Optional[Any] = { 'num_bp1': self.num_bpa, 'num_bp2': self.num_bpa, 'num_bp3': self.num_bpa, 'conv1': self.conva, 'step_conv1': self.step_conva, 'size_pooling1': self.size_poolinga, 'rate_weight': self.rate_weight, 'rate_thre': self.rate_thre, 'w_conv1': self.w_conva, 'wkj': self.wkj, 'vji': self.vji, 'thre_conv1': self.thre_conva, 'thre_bp2': self.thre_bpa, 'thre_bp3': self.thre_bpa, } with open(a , 'wb' ) as f: pickle.dump(a , a ) print(f"""Model saved: {save_path}""" ) @classmethod def _UpperCAmelCase ( cls , a ) -> Any: # read saved model with open(a , 'rb' ) as f: lowercase__ : Optional[int] = pickle.load(a ) # noqa: S301 lowercase__ : Optional[int] = model_dic.get('conv1' ) conv_get.append(model_dic.get('step_conv1' ) ) lowercase__ : List[Any] = model_dic.get('size_pooling1' ) lowercase__ : Tuple = model_dic.get('num_bp1' ) lowercase__ : int = model_dic.get('num_bp2' ) lowercase__ : int = model_dic.get('num_bp3' ) lowercase__ : Union[str, Any] = model_dic.get('rate_weight' ) lowercase__ : Tuple = model_dic.get('rate_thre' ) # create model instance lowercase__ : Tuple = CNN(a , a , a , a , a , a , a ) # modify model parameter lowercase__ : str = model_dic.get('w_conv1' ) lowercase__ : Optional[int] = model_dic.get('wkj' ) lowercase__ : Tuple = model_dic.get('vji' ) lowercase__ : str = model_dic.get('thre_conv1' ) lowercase__ : Union[str, Any] = model_dic.get('thre_bp2' ) lowercase__ : List[str] = model_dic.get('thre_bp3' ) return conv_ins def _UpperCAmelCase ( self , a ) -> str: return 1 / (1 + np.exp(-1 * x )) def _UpperCAmelCase ( self , a ) -> Any: return round(a , 3 ) def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[str]: # convolution process lowercase__ : int = convs[0] lowercase__ : Optional[Any] = convs[1] lowercase__ : int = np.shape(a )[0] # get the data slice of original image data, data_focus lowercase__ : Optional[Any] = [] for i_focus in range(0 , size_data - size_conv + 1 , a ): for j_focus in range(0 , size_data - size_conv + 1 , a ): lowercase__ : Optional[int] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(a ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ : Union[str, Any] = [] lowercase__ : Dict = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(a ): lowercase__ : Any = [] for i_focus in range(len(a ) ): lowercase__ : Tuple = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(a ) ) lowercase__ : Optional[Any] = np.asmatrix(a ).reshape( a , a ) data_featuremap.append(a ) # expanding the data slice to One dimenssion lowercase__ : str = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(a ) ) lowercase__ : int = np.asarray(a ) return focus_list, data_featuremap def _UpperCAmelCase ( self , a , a , a="average_pool" ) -> str: # pooling process lowercase__ : List[str] = len(featuremaps[0] ) lowercase__ : List[str] = int(size_map / size_pooling ) lowercase__ : str = [] for i_map in range(len(a ) ): lowercase__ : List[str] = featuremaps[i_map] lowercase__ : Optional[int] = [] for i_focus in range(0 , a , a ): for j_focus in range(0 , a , a ): lowercase__ : List[Any] = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(a ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(a ) ) lowercase__ : List[Any] = np.asmatrix(a ).reshape(a , a ) featuremap_pooled.append(a ) return featuremap_pooled def _UpperCAmelCase ( self , a ) -> List[str]: # expanding three dimension data to one dimension list lowercase__ : Any = [] for i in range(len(a ) ): lowercase__ : Optional[int] = np.shape(data[i] ) lowercase__ : int = data[i].reshape(1 , shapes[0] * shapes[1] ) lowercase__ : str = data_listed.getA().tolist()[0] data_expanded.extend(a ) lowercase__ : int = np.asarray(a ) return data_expanded def _UpperCAmelCase ( self , a ) -> Dict: # expanding matrix to one dimension list lowercase__ : Dict = np.asarray(a ) lowercase__ : Union[str, Any] = np.shape(a ) lowercase__ : Optional[Any] = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[Any]: lowercase__ : Dict = [] lowercase__ : int = 0 for i_map in range(a ): lowercase__ : str = np.ones((size_map, size_map) ) for i in range(0 , a , a ): for j in range(0 , a , a ): lowercase__ : Optional[Any] = pd_pool[ i_pool ] lowercase__ : Union[str, Any] = i_pool + 1 lowercase__ : List[Any] = np.multiply( a , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(a ) return pd_all def _UpperCAmelCase ( self , a , a , a , a , a , a=bool ) -> str: # model traning print('----------------------Start Training-------------------------' ) print((' - - Shape: Train_Data ', np.shape(a )) ) print((' - - Shape: Teach_Data ', np.shape(a )) ) lowercase__ : int = 0 lowercase__ : List[Any] = [] lowercase__ : Union[str, Any] = 1_0_0_0_0 while rp < n_repeat and mse >= error_accuracy: lowercase__ : List[Any] = 0 print(f"""-------------Learning Time {rp}--------------""" ) for p in range(len(a ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ : Optional[int] = np.asmatrix(datas_train[p] ) lowercase__ : int = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ : Union[str, Any] = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Optional[Any] = self.pooling(a , self.size_poolinga ) lowercase__ : Tuple = np.shape(a ) lowercase__ : List[str] = self._expand(a ) lowercase__ : Optional[int] = data_bp_input lowercase__ : Optional[Any] = np.dot(a , self.vji.T ) - self.thre_bpa lowercase__ : str = self.sig(a ) lowercase__ : Tuple = np.dot(a , self.wkj.T ) - self.thre_bpa lowercase__ : Any = self.sig(a ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ : int = np.multiply( (data_teach - bp_outa) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Any = np.multiply( np.dot(a , self.wkj ) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Optional[int] = np.dot(a , self.vji ) lowercase__ : Union[str, Any] = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ : Any = pd_conva_pooled.T.getA().tolist() lowercase__ : List[str] = self._calculate_gradient_from_pool( a , a , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ : Optional[int] = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ : Tuple = self.rate_weight * np.dot(a , a ) lowercase__ : Union[str, Any] = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ : Any = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ : Tuple = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ : Optional[Any] = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ : Dict = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ : str = rp + 1 lowercase__ : List[str] = error_count / patterns all_mse.append(a ) def draw_error(): lowercase__ : Any = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(a , '+-' ) plt.plot(a , 'r--' ) plt.xlabel('Learning Times' ) plt.ylabel('All_mse' ) plt.grid(a , alpha=0.5 ) plt.show() print('------------------Training Complished---------------------' ) print((' - - Training epoch: ', rp, f""" - - Mse: {mse:.6f}""") ) if draw_e: draw_error() return mse def _UpperCAmelCase ( self , a ) -> List[Any]: # model predict lowercase__ : Optional[int] = [] print('-------------------Start Testing-------------------------' ) print((' - - Shape: Test_Data ', np.shape(a )) ) for p in range(len(a ) ): lowercase__ : List[str] = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ : Tuple = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Any = self.pooling(a , self.size_poolinga ) lowercase__ : Union[str, Any] = self._expand(a ) lowercase__ : Optional[Any] = data_bp_input lowercase__ : str = bp_outa * self.vji.T - self.thre_bpa lowercase__ : Optional[Any] = self.sig(a ) lowercase__ : Dict = bp_outa * self.wkj.T - self.thre_bpa lowercase__ : List[str] = self.sig(a ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ : Optional[int] = [list(map(self.do_round , a ) ) for each in produce_out] return np.asarray(a ) def _UpperCAmelCase ( self , a ) -> List[str]: # return the data of image after convoluting process so we can check it out lowercase__ : Any = np.asmatrix(a ) lowercase__ , lowercase__ : str = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Tuple = self.pooling(a , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
77
1
"""simple docstring""" from collections.abc import Generator def a_ ( ): '''simple docstring''' lowercase__ , lowercase__ : List[str] = 0, 1 while True: lowercase__ , lowercase__ : Optional[int] = b, a + b yield b def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' lowercase__ : List[Any] = 1 lowercase__ : Any = fibonacci_generator() while len(str(next(_lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
77
"""simple docstring""" from collections.abc import Generator def a_ ( ): '''simple docstring''' lowercase__ , lowercase__ : List[str] = 0, 1 while True: lowercase__ , lowercase__ : Optional[int] = b, a + b yield b def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' lowercase__ : List[Any] = 1 lowercase__ : Any = fibonacci_generator() while len(str(next(_lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
77
1
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[List[PIL.Image.Image], np.ndarray] lowerCamelCase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : np.ndarray lowerCamelCase__ : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
77
"""simple docstring""" import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class UpperCAmelCase_ : def __init__( self , a ) -> List[str]: if isinstance(a , a ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden lowercase__ : Optional[Any] = deepcopy(a ) elif os.path.exists(a ): with io.open(a , 'r' , encoding='utf-8' ) as f: lowercase__ : List[Any] = json.load(a ) else: try: lowercase__ : Optional[int] = baseaa.urlsafe_baadecode(a ).decode('utf-8' ) lowercase__ : List[str] = json.loads(a ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""" ) lowercase__ : Any = config self.set_stage_and_offload() def _UpperCAmelCase ( self ) -> Dict: # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. lowercase__ : Tuple = self.get_value('zero_optimization.stage' , -1 ) # offload lowercase__ : int = False if self.is_zeroa() or self.is_zeroa(): lowercase__ : str = set(['cpu', 'nvme'] ) lowercase__ : Optional[Any] = set( [ self.get_value('zero_optimization.offload_optimizer.device' ), self.get_value('zero_optimization.offload_param.device' ), ] ) if len(offload_devices & offload_devices_valid ) > 0: lowercase__ : Optional[Any] = True def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Dict = self.config # find the config node of interest if it exists lowercase__ : int = ds_key_long.split('.' ) lowercase__ : Dict = nodes.pop() for node in nodes: lowercase__ : Optional[Any] = config.get(a ) if config is None: return None, ds_key return config, ds_key def _UpperCAmelCase ( self , a , a=None ) -> Union[str, Any]: lowercase__ , lowercase__ : Tuple = self.find_config_node(a ) if config is None: return default return config.get(a , a ) def _UpperCAmelCase ( self , a , a=False ) -> Any: lowercase__ : str = self.config # find the config node of interest if it exists lowercase__ : List[Any] = ds_key_long.split('.' ) for node in nodes: lowercase__ : str = config lowercase__ : str = config.get(a ) if config is None: if must_exist: raise ValueError(f"""Can't find {ds_key_long} entry in the config: {self.config}""" ) else: return # if found remove it if parent_config is not None: parent_config.pop(a ) def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : Union[str, Any] = self.get_value(a ) return False if value is None else bool(a ) def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Any = self.get_value(a ) return False if value is None else not bool(a ) def _UpperCAmelCase ( self ) -> Tuple: return self._stage == 2 def _UpperCAmelCase ( self ) -> List[Any]: return self._stage == 3 def _UpperCAmelCase ( self ) -> str: return self._offload class UpperCAmelCase_ : def __init__( self , a ) -> str: lowercase__ : Tuple = engine def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: # runs backpropagation and handles mixed precision self.engine.backward(a , **a ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class UpperCAmelCase_ ( _a): def __init__( self , a ) -> Dict: super().__init__(a , device_placement=a , scaler=a ) lowercase__ : Union[str, Any] = hasattr(self.optimizer , 'overflow' ) def _UpperCAmelCase ( self , a=None ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def _UpperCAmelCase ( self ) -> Optional[int]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def _UpperCAmelCase ( self ) -> Tuple: if self.__has_overflow__: return self.optimizer.overflow return False class UpperCAmelCase_ ( _a): def __init__( self , a , a ) -> Any: super().__init__(a , a ) def _UpperCAmelCase ( self ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class UpperCAmelCase_ : def __init__( self , a , a=0.001 , a=0 , **a ) -> Tuple: lowercase__ : List[Any] = params lowercase__ : int = lr lowercase__ : int = weight_decay lowercase__ : Union[str, Any] = kwargs class UpperCAmelCase_ : def __init__( self , a , a=None , a=0 , **a ) -> Tuple: lowercase__ : Dict = optimizer lowercase__ : List[str] = total_num_steps lowercase__ : Optional[int] = warmup_num_steps lowercase__ : List[Any] = kwargs
77
1
"""simple docstring""" import torch def a_ ( ): '''simple docstring''' if torch.cuda.is_available(): lowercase__ : Tuple = torch.cuda.device_count() else: lowercase__ : Optional[int] = 0 print(f"""Successfully ran on {num_gpus} GPUs""" ) if __name__ == "__main__": main()
77
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _UpperCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , *a , **a ) -> Union[str, Any]: super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _UpperCAmelCase ( self , a=None ) -> Dict: lowercase__ : Any = {} if top_k is not None: lowercase__ : List[str] = top_k return {}, {}, postprocess_params def __call__( self , a , **a ) -> Tuple: return super().__call__(a , **a ) def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : List[Any] = load_image(a ) lowercase__ : Union[str, Any] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _UpperCAmelCase ( self , a ) -> List[str]: lowercase__ : Dict = self.model(**a ) return model_outputs def _UpperCAmelCase ( self , a , a=5 ) -> Dict: if top_k > self.model.config.num_labels: lowercase__ : List[Any] = self.model.config.num_labels if self.framework == "pt": lowercase__ : Tuple = model_outputs.logits.softmax(-1 )[0] lowercase__ , lowercase__ : Optional[Any] = probs.topk(a ) elif self.framework == "tf": lowercase__ : Union[str, Any] = stable_softmax(model_outputs.logits , axis=-1 )[0] lowercase__ : str = tf.math.top_k(a , k=a ) lowercase__ , lowercase__ : Dict = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) lowercase__ : Dict = scores.tolist() lowercase__ : Dict = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
77
1
"""simple docstring""" import argparse import os import re import packaging.version _UpperCamelCase : Optional[Any] = "examples/" _UpperCamelCase : Tuple = { "examples": (re.compile(r"^check_min_version\(\"[^\"]+\"\)\s*$", re.MULTILINE), "check_min_version(\"VERSION\")\n"), "init": (re.compile(r"^__version__\s+=\s+\"([^\"]+)\"\s*$", re.MULTILINE), "__version__ = \"VERSION\"\n"), "setup": (re.compile(r"^(\s*)version\s*=\s*\"[^\"]+\",", re.MULTILINE), r"\1version=\"VERSION\","), "doc": (re.compile(r"^(\s*)release\s*=\s*\"[^\"]+\"$", re.MULTILINE), "release = \"VERSION\"\n"), } _UpperCamelCase : Any = { "init": "src/diffusers/__init__.py", "setup": "setup.py", } _UpperCamelCase : Union[str, Any] = "README.md" def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] ): '''simple docstring''' with open(_lowerCAmelCase , 'r' , encoding='utf-8' , newline='\n' ) as f: lowercase__ : Optional[Any] = f.read() lowercase__ , lowercase__ : Tuple = REPLACE_PATTERNS[pattern] lowercase__ : int = replace.replace('VERSION' , _lowerCAmelCase ) lowercase__ : Optional[int] = re_pattern.sub(_lowerCAmelCase , _lowerCAmelCase ) with open(_lowerCAmelCase , 'w' , encoding='utf-8' , newline='\n' ) as f: f.write(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' for folder, directories, fnames in os.walk(_lowerCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('research_projects' ) if "legacy" in directories: directories.remove('legacy' ) for fname in fnames: if fname.endswith('.py' ): update_version_in_file(os.path.join(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , pattern='examples' ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : List[Any]=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not patch: update_version_in_examples(_lowerCAmelCase ) def a_ ( ): '''simple docstring''' lowercase__ : List[str] = '🤗 Transformers currently provides the following architectures' lowercase__ : int = '1. Want to contribute a new model?' with open(_lowerCAmelCase , 'r' , encoding='utf-8' , newline='\n' ) as f: lowercase__ : List[Any] = f.readlines() # Find the start of the list. lowercase__ : Tuple = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowercase__ : List[str] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('1.' ): lowercase__ : Optional[Any] = lines[index].replace( 'https://huggingface.co/docs/diffusers/main/model_doc' , 'https://huggingface.co/docs/diffusers/model_doc' , ) index += 1 with open(_lowerCAmelCase , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(_lowerCAmelCase ) def a_ ( ): '''simple docstring''' with open(REPLACE_FILES['init'] , 'r' ) as f: lowercase__ : List[str] = f.read() lowercase__ : int = REPLACE_PATTERNS['init'][0].search(_lowerCAmelCase ).groups()[0] return packaging.version.parse(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : Optional[Any]=False ): '''simple docstring''' lowercase__ : Any = get_version() if patch and default_version.is_devrelease: raise ValueError('Can\'t create a patch version from the dev branch, checkout a released version!' ) if default_version.is_devrelease: lowercase__ : List[Any] = default_version.base_version elif patch: lowercase__ : Any = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: lowercase__ : List[Any] = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. lowercase__ : List[Any] = input(f"""Which version are you releasing? [{default_version}]""" ) if len(_lowerCAmelCase ) == 0: lowercase__ : Dict = default_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase , patch=_lowerCAmelCase ) def a_ ( ): '''simple docstring''' lowercase__ : List[str] = get_version() lowercase__ : Any = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" lowercase__ : int = current_version.base_version # Check with the user we got that right. lowercase__ : Dict = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(_lowerCAmelCase ) == 0: lowercase__ : Optional[int] = dev_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": _UpperCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument("--post_release", action="store_true", help="Whether this is pre or post release.") parser.add_argument("--patch", action="store_true", help="Whether or not this is a patch release.") _UpperCamelCase : Union[str, Any] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("Nothing to do after a patch :-)") else: post_release_work()
77
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : str = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all LED models at https://huggingface.co/models?filter=LED _UpperCamelCase : Optional[Any] = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } _UpperCamelCase : Optional[int] = { "allenai/led-base-16384": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def a_ ( ): '''simple docstring''' lowercase__ : int = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowercase__ : Union[str, Any] = bs[:] lowercase__ : str = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCAmelCase ) cs.append(2**8 + n ) n += 1 lowercase__ : str = [chr(_lowerCAmelCase ) for n in cs] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Dict = set() lowercase__ : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Optional[Any] = char return pairs class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = VOCAB_FILES_NAMES lowerCamelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , a , a , a="replace" , a="<s>" , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a=False , **a , ) -> Any: lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else bos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else eos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else sep_token lowercase__ : Dict = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else cls_token lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else unk_token lowercase__ : Tuple = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( errors=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , add_prefix_space=a , **a , ) with open(a , encoding='utf-8' ) as vocab_handle: lowercase__ : Tuple = json.load(a ) lowercase__ : Dict = {v: k for k, v in self.encoder.items()} lowercase__ : str = errors # how to handle errors in decoding lowercase__ : Optional[Any] = bytes_to_unicode() lowercase__ : Optional[Any] = {v: k for k, v in self.byte_encoder.items()} with open(a , encoding='utf-8' ) as merges_handle: lowercase__ : Optional[Any] = merges_handle.read().split('\n' )[1:-1] lowercase__ : Optional[int] = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ : Union[str, Any] = dict(zip(a , range(len(a ) ) ) ) lowercase__ : Tuple = {} lowercase__ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ : List[Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _UpperCAmelCase ( self ) -> List[Any]: return len(self.encoder ) def _UpperCAmelCase ( self ) -> str: return dict(self.encoder , **self.added_tokens_encoder ) def _UpperCAmelCase ( self , a ) -> List[str]: if token in self.cache: return self.cache[token] lowercase__ : Optional[Any] = tuple(a ) lowercase__ : int = get_pairs(a ) if not pairs: return token while True: lowercase__ : List[str] = min(a , key=lambda a : self.bpe_ranks.get(a , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ : List[str] = bigram lowercase__ : Union[str, Any] = [] lowercase__ : List[Any] = 0 while i < len(a ): try: lowercase__ : str = word.index(a , a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : Optional[int] = j if word[i] == first and i < len(a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : int = tuple(a ) lowercase__ : Dict = new_word if len(a ) == 1: break else: lowercase__ : Any = get_pairs(a ) lowercase__ : List[str] = ' '.join(a ) lowercase__ : Optional[Any] = word return word def _UpperCAmelCase ( self , a ) -> Union[str, Any]: lowercase__ : Tuple = [] for token in re.findall(self.pat , a ): lowercase__ : Union[str, Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(a ).split(' ' ) ) return bpe_tokens def _UpperCAmelCase ( self , a ) -> Optional[Any]: return self.encoder.get(a , self.encoder.get(self.unk_token ) ) def _UpperCAmelCase ( self , a ) -> Optional[int]: return self.decoder.get(a ) def _UpperCAmelCase ( self , a ) -> str: lowercase__ : Any = ''.join(a ) lowercase__ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Any = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : str = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(a , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a , ensure_ascii=a ) + '\n' ) lowercase__ : List[Any] = 0 with open(a , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) lowercase__ : Union[str, Any] = token_index writer.write(' '.join(a ) + '\n' ) index += 1 return vocab_file, merge_file def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ : Union[str, Any] = [self.cls_token_id] lowercase__ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) + [1] def _UpperCAmelCase ( self , a , a = None ) -> List[int]: lowercase__ : Dict = [self.sep_token_id] lowercase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCAmelCase ( self , a , a=False , **a ) -> Optional[int]: lowercase__ : Tuple = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(a ) > 0 and not text[0].isspace()): lowercase__ : List[str] = ' ' + text return (text, kwargs) def _UpperCAmelCase ( self , a , a = None , a = PaddingStrategy.DO_NOT_PAD , a = None , a = None , ) -> dict: lowercase__ : Dict = super()._pad( encoded_inputs=a , max_length=a , padding_strategy=a , pad_to_multiple_of=a , return_attention_mask=a , ) # Load from model defaults if return_attention_mask is None: lowercase__ : Union[str, Any] = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ : Any = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ : Tuple = len(encoded_inputs['global_attention_mask'] ) != len(a ) if needs_to_be_padded: lowercase__ : str = len(a ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ : Union[str, Any] = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowercase__ : List[str] = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
77
1
"""simple docstring""" import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' lowercase__ : str = SwinaSRConfig() if "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: lowercase__ : Any = 4 elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: lowercase__ : Union[str, Any] = 4 lowercase__ : int = 48 lowercase__ : int = 'pixelshuffle_aux' elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: lowercase__ : List[str] = [6, 6, 6, 6] lowercase__ : Union[str, Any] = 60 lowercase__ : Union[str, Any] = [6, 6, 6, 6] lowercase__ : Tuple = 'pixelshuffledirect' elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: lowercase__ : Dict = 4 lowercase__ : Union[str, Any] = 'nearest+conv' elif "Swin2SR_Jpeg_dynamic" in checkpoint_url: lowercase__ : Union[str, Any] = 1 lowercase__ : Optional[int] = 1 lowercase__ : Any = 126 lowercase__ : Tuple = 7 lowercase__ : Dict = 2_5_5.0 lowercase__ : Optional[Any] = '' return config def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple ): '''simple docstring''' if "patch_embed.proj" in name and "layers" not in name: lowercase__ : List[Any] = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: lowercase__ : Optional[int] = name.replace('patch_embed.norm' , 'embeddings.patch_embeddings.layernorm' ) if "layers" in name: lowercase__ : List[Any] = name.replace('layers' , 'encoder.stages' ) if "residual_group.blocks" in name: lowercase__ : List[str] = name.replace('residual_group.blocks' , 'layers' ) if "attn.proj" in name: lowercase__ : Optional[Any] = name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: lowercase__ : Optional[int] = name.replace('attn' , 'attention.self' ) if "norm1" in name: lowercase__ : Optional[Any] = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: lowercase__ : Dict = name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: lowercase__ : Optional[int] = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: lowercase__ : int = name.replace('mlp.fc2' , 'output.dense' ) if "q_bias" in name: lowercase__ : Optional[int] = name.replace('q_bias' , 'query.bias' ) if "k_bias" in name: lowercase__ : Dict = name.replace('k_bias' , 'key.bias' ) if "v_bias" in name: lowercase__ : int = name.replace('v_bias' , 'value.bias' ) if "cpb_mlp" in name: lowercase__ : List[Any] = name.replace('cpb_mlp' , 'continuous_position_bias_mlp' ) if "patch_embed.proj" in name: lowercase__ : Optional[Any] = name.replace('patch_embed.proj' , 'patch_embed.projection' ) if name == "norm.weight": lowercase__ : Optional[int] = 'layernorm.weight' if name == "norm.bias": lowercase__ : Union[str, Any] = 'layernorm.bias' if "conv_first" in name: lowercase__ : Any = name.replace('conv_first' , 'first_convolution' ) if ( "upsample" in name or "conv_before_upsample" in name or "conv_bicubic" in name or "conv_up" in name or "conv_hr" in name or "conv_last" in name or "aux" in name ): # heads if "conv_last" in name: lowercase__ : Optional[int] = name.replace('conv_last' , 'final_convolution' ) if config.upsampler in ["pixelshuffle", "pixelshuffle_aux", "nearest+conv"]: if "conv_before_upsample.0" in name: lowercase__ : List[str] = name.replace('conv_before_upsample.0' , 'conv_before_upsample' ) if "upsample.0" in name: lowercase__ : List[str] = name.replace('upsample.0' , 'upsample.convolution_0' ) if "upsample.2" in name: lowercase__ : Tuple = name.replace('upsample.2' , 'upsample.convolution_1' ) lowercase__ : Optional[int] = 'upsample.' + name elif config.upsampler == "pixelshuffledirect": lowercase__ : Any = name.replace('upsample.0.weight' , 'upsample.conv.weight' ) lowercase__ : Optional[Any] = name.replace('upsample.0.bias' , 'upsample.conv.bias' ) else: pass else: lowercase__ : Optional[int] = 'swin2sr.' + name return name def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[Any] ): '''simple docstring''' for key in orig_state_dict.copy().keys(): lowercase__ : List[str] = orig_state_dict.pop(_lowerCAmelCase ) if "qkv" in key: lowercase__ : Optional[int] = key.split('.' ) lowercase__ : str = int(key_split[1] ) lowercase__ : List[str] = int(key_split[4] ) lowercase__ : List[str] = config.embed_dim if "weight" in key: lowercase__ : List[Any] = val[:dim, :] lowercase__ : str = val[dim : dim * 2, :] lowercase__ : List[str] = val[-dim:, :] else: lowercase__ : int = val[:dim] lowercase__ : List[Any] = val[dim : dim * 2] lowercase__ : Union[str, Any] = val[-dim:] pass else: lowercase__ : Dict = val return orig_state_dict def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : List[str] = get_config(_lowerCAmelCase ) lowercase__ : Optional[int] = SwinaSRForImageSuperResolution(_lowerCAmelCase ) model.eval() lowercase__ : Tuple = torch.hub.load_state_dict_from_url(_lowerCAmelCase , map_location='cpu' ) lowercase__ : List[Any] = convert_state_dict(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ , lowercase__ : Dict = model.load_state_dict(_lowerCAmelCase , strict=_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: raise ValueError('Missing keys when converting: {}'.format(_lowerCAmelCase ) ) for key in unexpected_keys: if not ("relative_position_index" in key or "relative_coords_table" in key or "self_mask" in key): raise ValueError(f"""Unexpected key {key} in state_dict""" ) # verify values lowercase__ : int = 'https://github.com/mv-lab/swin2sr/blob/main/testsets/real-inputs/shanghai.jpg?raw=true' lowercase__ : Dict = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ).convert('RGB' ) lowercase__ : Optional[Any] = SwinaSRImageProcessor() # pixel_values = processor(image, return_tensors="pt").pixel_values lowercase__ : int = 126 if 'Jpeg' in checkpoint_url else 256 lowercase__ : List[str] = Compose( [ Resize((image_size, image_size) ), ToTensor(), Normalize(mean=[0.4_8_5, 0.4_5_6, 0.4_0_6] , std=[0.2_2_9, 0.2_2_4, 0.2_2_5] ), ] ) lowercase__ : str = transforms(_lowerCAmelCase ).unsqueeze(0 ) if config.num_channels == 1: lowercase__ : str = pixel_values[:, 0, :, :].unsqueeze(1 ) lowercase__ : str = model(_lowerCAmelCase ) # assert values if "Swin2SR_ClassicalSR_X2_64" in checkpoint_url: lowercase__ : Dict = torch.Size([1, 3, 512, 512] ) lowercase__ : Optional[int] = torch.tensor( [[-0.7_0_8_7, -0.7_1_3_8, -0.6_7_2_1], [-0.8_3_4_0, -0.8_0_9_5, -0.7_2_9_8], [-0.9_1_4_9, -0.8_4_1_4, -0.7_9_4_0]] ) elif "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: lowercase__ : Any = torch.Size([1, 3, 1024, 1024] ) lowercase__ : str = torch.tensor( [[-0.7_7_7_5, -0.8_1_0_5, -0.8_9_3_3], [-0.7_7_6_4, -0.8_3_5_6, -0.9_2_2_5], [-0.7_9_7_6, -0.8_6_8_6, -0.9_5_7_9]] ) elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: # TODO values didn't match exactly here lowercase__ : Tuple = torch.Size([1, 3, 1024, 1024] ) lowercase__ : str = torch.tensor( [[-0.8_0_3_5, -0.7_5_0_4, -0.7_4_9_1], [-0.8_5_3_8, -0.8_1_2_4, -0.7_7_8_2], [-0.8_8_0_4, -0.8_6_5_1, -0.8_4_9_3]] ) elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: lowercase__ : Union[str, Any] = torch.Size([1, 3, 512, 512] ) lowercase__ : str = torch.tensor( [[-0.7_6_6_9, -0.8_6_6_2, -0.8_7_6_7], [-0.8_8_1_0, -0.9_9_6_2, -0.9_8_2_0], [-0.9_3_4_0, -1.0_3_2_2, -1.1_1_4_9]] ) elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: lowercase__ : Dict = torch.Size([1, 3, 1024, 1024] ) lowercase__ : Any = torch.tensor( [[-0.5_2_3_8, -0.5_5_5_7, -0.6_3_2_1], [-0.6_0_1_6, -0.5_9_0_3, -0.6_3_9_1], [-0.6_2_4_4, -0.6_3_3_4, -0.6_8_8_9]] ) assert ( outputs.reconstruction.shape == expected_shape ), f"""Shape of reconstruction should be {expected_shape}, but is {outputs.reconstruction.shape}""" assert torch.allclose(outputs.reconstruction[0, 0, :3, :3] , _lowerCAmelCase , atol=1E-3 ) print('Looks ok!' ) lowercase__ : Optional[int] = { 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth': ( 'swin2SR-classical-sr-x2-64' ), 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X4_64.pth': ( 'swin2SR-classical-sr-x4-64' ), 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_CompressedSR_X4_48.pth': ( 'swin2SR-compressed-sr-x4-48' ), 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_Lightweight_X2_64.pth': ( 'swin2SR-lightweight-x2-64' ), 'https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR.pth': ( 'swin2SR-realworld-sr-x4-64-bsrgan-psnr' ), } lowercase__ : int = url_to_name[checkpoint_url] if pytorch_dump_folder_path is not None: print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(_lowerCAmelCase ) if push_to_hub: model.push_to_hub(f"""caidas/{model_name}""" ) processor.push_to_hub(f"""caidas/{model_name}""" ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth", type=str, help="URL of the original Swin2SR checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether to push the converted model to the hub.") _UpperCamelCase : List[Any] = parser.parse_args() convert_swinasr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
77
"""simple docstring""" _UpperCamelCase : Union[str, Any] = 8.3_1_4_4_5_9_8 def a_ ( _lowerCAmelCase : float , _lowerCAmelCase : float ): '''simple docstring''' if temperature < 0: raise Exception('Temperature cannot be less than 0 K' ) if molar_mass <= 0: raise Exception('Molar mass cannot be less than or equal to 0 kg/mol' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _UpperCamelCase : List[Any] = 3_00 _UpperCamelCase : Tuple = 28 _UpperCamelCase : Any = rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int ): '''simple docstring''' return [ { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], }, { 0: [6], 1: [9], 2: [4, 5], 3: [4], 4: [2, 3], 5: [2], 6: [0, 7], 7: [6], 8: [], 9: [1], }, { 0: [4], 1: [6], 2: [], 3: [5, 6, 7], 4: [0, 6], 5: [3, 8, 9], 6: [1, 3, 4, 7], 7: [3, 6, 8, 9], 8: [5, 7], 9: [5, 7], }, { 0: [1, 3], 1: [0, 2, 4], 2: [1, 3, 4], 3: [0, 2, 4], 4: [1, 2, 3], }, ][index] def a_ ( _lowerCAmelCase : dict[int, list[int]] ): '''simple docstring''' lowercase__ : Optional[Any] = 0 lowercase__ : Optional[int] = len(_lowerCAmelCase ) # No of vertices in graph lowercase__ : Optional[int] = [0] * n lowercase__ : str = [False] * n def dfs(_lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : List[str] ): lowercase__ : List[str] = True lowercase__ : Optional[int] = id_ id_ += 1 for to in graph[at]: if to == parent: pass elif not visited[to]: dfs(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , id_ ) lowercase__ : int = min(low[at] , low[to] ) if id_ <= low[to]: bridges.append((at, to) if at < to else (to, at) ) else: # This edge is a back edge and cannot be a bridge lowercase__ : Optional[int] = min(low[at] , low[to] ) lowercase__ : list[tuple[int, int]] = [] for i in range(_lowerCAmelCase ): if not visited[i]: dfs(_lowerCAmelCase , -1 , _lowerCAmelCase , id_ ) return bridges if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization lowerCamelCase__ : str = field(default="text-classification" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({"labels": ClassLabel}) lowerCamelCase__ : str = "text" lowerCamelCase__ : str = "labels" def _UpperCAmelCase ( self , a ) -> Tuple: if self.label_column not in features: raise ValueError(f"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , a ): raise ValueError(f"""Column {self.label_column} is not a ClassLabel.""" ) lowercase__ : Optional[Any] = copy.deepcopy(self ) lowercase__ : Optional[Any] = self.label_schema.copy() lowercase__ : Any = features[self.label_column] lowercase__ : Optional[Any] = label_schema return task_template @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return { self.text_column: "text", self.label_column: "labels", }
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : list[list[int]] , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : set ): '''simple docstring''' lowercase__ , lowercase__ : Tuple = len(_lowerCAmelCase ), len(grid[0] ) if ( min(_lowerCAmelCase , _lowerCAmelCase ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) lowercase__ : Optional[int] = 0 count += depth_first_search(_lowerCAmelCase , row + 1 , _lowerCAmelCase , _lowerCAmelCase ) count += depth_first_search(_lowerCAmelCase , row - 1 , _lowerCAmelCase , _lowerCAmelCase ) count += depth_first_search(_lowerCAmelCase , _lowerCAmelCase , col + 1 , _lowerCAmelCase ) count += depth_first_search(_lowerCAmelCase , _lowerCAmelCase , col - 1 , _lowerCAmelCase ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np _UpperCamelCase : Any = re.compile(r"\b(a|an|the)\b", re.UNICODE) _UpperCamelCase : Union[str, Any] = None def a_ ( ): '''simple docstring''' lowercase__ : Optional[int] = argparse.ArgumentParser('Official evaluation script for SQuAD version 2.0.' ) parser.add_argument('data_file' , metavar='data.json' , help='Input data JSON file.' ) parser.add_argument('pred_file' , metavar='pred.json' , help='Model predictions.' ) parser.add_argument( '--out-file' , '-o' , metavar='eval.json' , help='Write accuracy metrics to file (default is stdout).' ) parser.add_argument( '--na-prob-file' , '-n' , metavar='na_prob.json' , help='Model estimates of probability of no answer.' ) parser.add_argument( '--na-prob-thresh' , '-t' , type=_lowerCAmelCase , default=1.0 , help='Predict "" if no-answer probability exceeds this (default = 1.0).' , ) parser.add_argument( '--out-image-dir' , '-p' , metavar='out_images' , default=_lowerCAmelCase , help='Save precision-recall curves to directory.' ) parser.add_argument('--verbose' , '-v' , action='store_true' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Tuple = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Optional[int] = bool(qa['answers']['text'] ) return qid_to_has_ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' def remove_articles(_lowerCAmelCase : int ): return ARTICLES_REGEX.sub(' ' , _lowerCAmelCase ) def white_space_fix(_lowerCAmelCase : str ): return " ".join(text.split() ) def remove_punc(_lowerCAmelCase : List[Any] ): lowercase__ : int = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_lowerCAmelCase : List[str] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_lowerCAmelCase ) ) ) ) def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' if not s: return [] return normalize_answer(_lowerCAmelCase ).split() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ): '''simple docstring''' return int(normalize_answer(_lowerCAmelCase ) == normalize_answer(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ): '''simple docstring''' lowercase__ : Dict = get_tokens(_lowerCAmelCase ) lowercase__ : List[str] = get_tokens(_lowerCAmelCase ) lowercase__ : List[Any] = collections.Counter(_lowerCAmelCase ) & collections.Counter(_lowerCAmelCase ) lowercase__ : int = sum(common.values() ) if len(_lowerCAmelCase ) == 0 or len(_lowerCAmelCase ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 lowercase__ : Any = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Dict = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Any = (2 * precision * recall) / (precision + recall) return fa def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Optional[int] = {} lowercase__ : Union[str, Any] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Any = qa['id'] lowercase__ : Union[str, Any] = [t for t in qa['answers']['text'] if normalize_answer(_lowerCAmelCase )] if not gold_answers: # For unanswerable questions, only correct answer is empty string lowercase__ : Dict = [''] if qid not in preds: print(f"""Missing prediction for {qid}""" ) continue lowercase__ : Optional[int] = preds[qid] # Take max over all gold answers lowercase__ : int = max(compute_exact(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) lowercase__ : Optional[Any] = max(compute_fa(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) return exact_scores, fa_scores def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : str = {} for qid, s in scores.items(): lowercase__ : int = na_probs[qid] > na_prob_thresh if pred_na: lowercase__ : Optional[Any] = float(not qid_to_has_ans[qid] ) else: lowercase__ : Optional[Any] = s return new_scores def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str=None ): '''simple docstring''' if not qid_list: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores.values() ) / total), ('f1', 1_0_0.0 * sum(fa_scores.values() ) / total), ('total', total), ] ) else: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores[k] for k in qid_list ) / total), ('f1', 1_0_0.0 * sum(fa_scores[k] for k in qid_list ) / total), ('total', total), ] ) def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' for k in new_eval: lowercase__ : int = new_eval[k] def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple ): '''simple docstring''' plt.step(_lowerCAmelCase , _lowerCAmelCase , color='b' , alpha=0.2 , where='post' ) plt.fill_between(_lowerCAmelCase , _lowerCAmelCase , step='post' , alpha=0.2 , color='b' ) plt.xlabel('Recall' ) plt.ylabel('Precision' ) plt.xlim([0.0, 1.0_5] ) plt.ylim([0.0, 1.0_5] ) plt.title(_lowerCAmelCase ) plt.savefig(_lowerCAmelCase ) plt.clf() def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : str , _lowerCAmelCase : Any=None , _lowerCAmelCase : List[str]=None ): '''simple docstring''' lowercase__ : Optional[int] = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) lowercase__ : Tuple = 0.0 lowercase__ : List[str] = 1.0 lowercase__ : List[str] = 0.0 lowercase__ : Union[str, Any] = [1.0] lowercase__ : List[Any] = [0.0] lowercase__ : Optional[int] = 0.0 for i, qid in enumerate(_lowerCAmelCase ): if qid_to_has_ans[qid]: true_pos += scores[qid] lowercase__ : Tuple = true_pos / float(i + 1 ) lowercase__ : Union[str, Any] = true_pos / float(_lowerCAmelCase ) if i == len(_lowerCAmelCase ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_lowerCAmelCase ) recalls.append(_lowerCAmelCase ) if out_image: plot_pr_curve(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return {"ap": 1_0_0.0 * avg_prec} def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple ): '''simple docstring''' if out_image_dir and not os.path.exists(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) lowercase__ : List[str] = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return lowercase__ : Dict = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_exact.png' ) , title='Precision-Recall curve for Exact Match score' , ) lowercase__ : Tuple = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_f1.png' ) , title='Precision-Recall curve for F1 score' , ) lowercase__ : List[Any] = {k: float(_lowerCAmelCase ) for k, v in qid_to_has_ans.items()} lowercase__ : Any = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_oracle.png' ) , title='Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)' , ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_exact' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_f1' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_oracle' ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' if not qid_list: return lowercase__ : List[str] = [na_probs[k] for k in qid_list] lowercase__ : Tuple = np.ones_like(_lowerCAmelCase ) / float(len(_lowerCAmelCase ) ) plt.hist(_lowerCAmelCase , weights=_lowerCAmelCase , bins=20 , range=(0.0, 1.0) ) plt.xlabel('Model probability of no-answer' ) plt.ylabel('Proportion of dataset' ) plt.title(f"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(_lowerCAmelCase , f"""na_prob_hist_{name}.png""" ) ) plt.clf() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : Tuple = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) lowercase__ : int = num_no_ans lowercase__ : Optional[int] = cur_score lowercase__ : Tuple = 0.0 lowercase__ : Dict = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) for i, qid in enumerate(_lowerCAmelCase ): if qid not in scores: continue if qid_to_has_ans[qid]: lowercase__ : Optional[int] = scores[qid] else: if preds[qid]: lowercase__ : List[Any] = -1 else: lowercase__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: lowercase__ : Dict = cur_score lowercase__ : Optional[int] = na_probs[qid] return 1_0_0.0 * best_score / len(_lowerCAmelCase ), best_thresh def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str ): '''simple docstring''' lowercase__ , lowercase__ : List[Any] = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ , lowercase__ : Dict = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Any = best_exact lowercase__ : Tuple = exact_thresh lowercase__ : Optional[Any] = best_fa lowercase__ : Any = fa_thresh def a_ ( ): '''simple docstring''' with open(OPTS.data_file ) as f: lowercase__ : List[Any] = json.load(_lowerCAmelCase ) lowercase__ : Union[str, Any] = dataset_json['data'] with open(OPTS.pred_file ) as f: lowercase__ : str = json.load(_lowerCAmelCase ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: lowercase__ : Union[str, Any] = json.load(_lowerCAmelCase ) else: lowercase__ : str = {k: 0.0 for k in preds} lowercase__ : int = make_qid_to_has_ans(_lowerCAmelCase ) # maps qid to True/False lowercase__ : List[str] = [k for k, v in qid_to_has_ans.items() if v] lowercase__ : Any = [k for k, v in qid_to_has_ans.items() if not v] lowercase__ , lowercase__ : Any = get_raw_scores(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Optional[Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Union[str, Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Tuple = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase ) if has_ans_qids: lowercase__ : int = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'HasAns' ) if no_ans_qids: lowercase__ : Optional[Any] = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'NoAns' ) if OPTS.na_prob_file: find_all_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'hasAns' ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'noAns' ) if OPTS.out_file: with open(OPTS.out_file , 'w' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) else: print(json.dumps(_lowerCAmelCase , indent=2 ) ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
77
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCamelCase : int = { "configuration_longt5": ["LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "LongT5Config", "LongT5OnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = [ "LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST", "LongT5EncoderModel", "LongT5ForConditionalGeneration", "LongT5Model", "LongT5PreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Any = [ "FlaxLongT5ForConditionalGeneration", "FlaxLongT5Model", "FlaxLongT5PreTrainedModel", ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys _UpperCamelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
77
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = 0 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Tuple = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Any: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : str = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : int = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = CLIPConfig() # Create a dummy config file with image_proceesor_type lowercase__ : Optional[int] = Path(a ) / 'preprocessor_config.json' lowercase__ : Optional[int] = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowercase__ : int = AutoImageProcessor.from_pretrained(a ).to_dict() config_dict.pop('image_processor_type' ) lowercase__ : Tuple = CLIPImageProcessor(**a ) # save in new folder model_config.save_pretrained(a ) config.save_pretrained(a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) # make sure private variable is not incorrectly saved lowercase__ : Optional[int] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = Path(a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'clip-base is not a local folder and is not a valid model identifier' ): lowercase__ : Any = AutoImageProcessor.from_pretrained('clip-base' ) def _UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex( a , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowercase__ : Dict = AutoImageProcessor.from_pretrained(a , revision='aaaaaa' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def _UpperCAmelCase ( self ) -> Optional[int]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(a ): lowercase__ : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : str = AutoImageProcessor.from_pretrained(a , trust_remote_code=a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def _UpperCAmelCase ( self ) -> int: try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoImageProcessor.register(a , a ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Optional[Any] = Path(a ) / 'preprocessor_config.json' lowercase__ : List[Any] = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = CustomImageProcessor.from_pretrained(a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCAmelCase ( self ) -> Dict: class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = True try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # If remote code is not set, the default is to use local lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowercase__ : int = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : Dict = [int(_lowerCAmelCase ) for i in ip_va_address.split('.' ) if i.isdigit()] return len(_lowerCAmelCase ) == 4 and all(0 <= int(_lowerCAmelCase ) <= 254 for octet in octets ) if __name__ == "__main__": _UpperCamelCase : Optional[Any] = input().strip() _UpperCamelCase : Union[str, Any] = "valid" if is_ip_va_address_valid(ip) else "invalid" print(f'''{ip} is a {valid_or_invalid} IP v4 address.''')
77
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class UpperCAmelCase_ ( _a): lowerCamelCase__ : Dict = ["image_processor", "tokenizer"] lowerCamelCase__ : Dict = "BlipImageProcessor" lowerCamelCase__ : Union[str, Any] = "AutoTokenizer" def __init__( self , a , a , a ) -> Optional[int]: super().__init__(a , a ) # add QFormer tokenizer lowercase__ : Dict = qformer_tokenizer def __call__( self , a = None , a = None , a = True , a = False , a = None , a = None , a = 0 , a = None , a = None , a = False , a = False , a = False , a = False , a = False , a = True , a = None , **a , ) -> BatchFeature: if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) lowercase__ : List[Any] = BatchFeature() if text is not None: lowercase__ : Optional[int] = self.tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) encoding.update(a ) lowercase__ : Optional[int] = self.qformer_tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) lowercase__ : List[str] = qformer_text_encoding.pop('input_ids' ) lowercase__ : Any = qformer_text_encoding.pop('attention_mask' ) if images is not None: lowercase__ : List[Any] = self.image_processor(a , return_tensors=a ) encoding.update(a ) return encoding def _UpperCAmelCase ( self , *a , **a ) -> List[str]: return self.tokenizer.batch_decode(*a , **a ) def _UpperCAmelCase ( self , *a , **a ) -> Tuple: return self.tokenizer.decode(*a , **a ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : str = self.tokenizer.model_input_names lowercase__ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: if os.path.isfile(a ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(a , exist_ok=a ) lowercase__ : int = os.path.join(a , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(a ) return super().save_pretrained(a , **a ) @classmethod def _UpperCAmelCase ( cls , a , **a ) -> str: lowercase__ : str = AutoTokenizer.from_pretrained(a , subfolder='qformer_tokenizer' ) lowercase__ : int = cls._get_arguments_from_pretrained(a , **a ) args.append(a ) return cls(*a )
77
1
"""simple docstring""" import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand _UpperCamelCase : List[Any] = ( "4S 3H 2C 7S 5H", "9D 8H 2C 6S 7H", "2D 6D 9D TH 7D", "TC 8C 2S JH 6C", "JH 8S TH AH QH", "TS KS 5S 9S AC", "KD 6S 9D TH AD", "KS 8D 4D 9S 4S", # pair "8C 4S KH JS 4D", # pair "QH 8H KD JH 8S", # pair "KC 4H KS 2H 8D", # pair "KD 4S KC 3H 8S", # pair "AH 8S AS KC JH", # pair "3H 4C 4H 3S 2H", # 2 pairs "5S 5D 2C KH KH", # 2 pairs "3C KH 5D 5S KH", # 2 pairs "AS 3C KH AD KH", # 2 pairs "7C 7S 3S 7H 5S", # 3 of a kind "7C 7S KH 2H 7H", # 3 of a kind "AC KH QH AH AS", # 3 of a kind "2H 4D 3C AS 5S", # straight (low ace) "3C 5C 4C 2C 6H", # straight "6S 8S 7S 5H 9H", # straight "JS QS 9H TS KH", # straight "QC KH TS JS AH", # straight (high ace) "8C 9C 5C 3C TC", # flush "3S 8S 9S 5S KS", # flush "4C 5C 9C 8C KC", # flush "JH 8H AH KH QH", # flush "3D 2H 3H 2C 2D", # full house "2H 2C 3S 3H 3D", # full house "KH KC 3S 3H 3D", # full house "JC 6H JS JD JH", # 4 of a kind "JC 7H JS JD JH", # 4 of a kind "JC KH JS JD JH", # 4 of a kind "2S AS 4S 5S 3S", # straight flush (low ace) "2D 6D 3D 4D 5D", # straight flush "5C 6C 3C 7C 4C", # straight flush "JH 9H TH KH QH", # straight flush "JH AH TH KH QH", # royal flush (high ace straight flush) ) _UpperCamelCase : Tuple = ( ("2H 3H 4H 5H 6H", "KS AS TS QS JS", "Loss"), ("2H 3H 4H 5H 6H", "AS AD AC AH JD", "Win"), ("AS AH 2H AD AC", "JS JD JC JH 3D", "Win"), ("2S AH 2H AS AC", "JS JD JC JH AD", "Loss"), ("2S AH 2H AS AC", "2H 3H 5H 6H 7H", "Win"), ("AS 3S 4S 8S 2S", "2H 3H 5H 6H 7H", "Win"), ("2H 3H 5H 6H 7H", "2S 3H 4H 5S 6C", "Win"), ("2S 3H 4H 5S 6C", "3D 4C 5H 6H 2S", "Tie"), ("2S 3H 4H 5S 6C", "AH AC 5H 6H AS", "Win"), ("2S 2H 4H 5S 4C", "AH AC 5H 6H AS", "Loss"), ("2S 2H 4H 5S 4C", "AH AC 5H 6H 7S", "Win"), ("6S AD 7H 4S AS", "AH AC 5H 6H 7S", "Loss"), ("2S AH 4H 5S KC", "AH AC 5H 6H 7S", "Loss"), ("2S 3H 6H 7S 9C", "7H 3C TH 6H 9S", "Loss"), ("4S 5H 6H TS AC", "3S 5H 6H TS AC", "Win"), ("2S AH 4H 5S 6C", "AD 4C 5H 6H 2C", "Tie"), ("AS AH 3H AD AC", "AS AH 2H AD AC", "Win"), ("AH AC 5H 5C QS", "AH AC 5H 5C KS", "Loss"), ("AH AC 5H 5C QS", "KH KC 5H 5C QS", "Win"), ("7C 7S KH 2H 7H", "3C 3S AH 2H 3H", "Win"), ("3C 3S AH 2H 3H", "7C 7S KH 2H 7H", "Loss"), ("6H 5H 4H 3H 2H", "5H 4H 3H 2H AH", "Win"), ("5H 4H 3H 2H AH", "5H 4H 3H 2H AH", "Tie"), ("5H 4H 3H 2H AH", "6H 5H 4H 3H 2H", "Loss"), ("AH AD KS KC AC", "AH KD KH AC KC", "Win"), ("2H 4D 3C AS 5S", "2H 4D 3C 6S 5S", "Loss"), ("2H 3S 3C 3H 2S", "3S 3C 2S 2H 2D", "Win"), ("4D 6D 5D 2D JH", "3S 8S 3H TC KH", "Loss"), ("4S 6C 8S 3S 7S", "AD KS 2D 7D 7C", "Loss"), ("6S 4C 7H 8C 3H", "5H JC AH 9D 9C", "Loss"), ("9D 9H JH TC QH", "3C 2S JS 5C 7H", "Win"), ("2H TC 8S AD 9S", "4H TS 7H 2C 5C", "Win"), ("9D 3S 2C 7S 7C", "JC TD 3C TC 9H", "Loss"), ) _UpperCamelCase : List[str] = ( ("2H 3H 4H 5H 6H", True), ("AS AH 2H AD AC", False), ("2H 3H 5H 6H 7H", True), ("KS AS TS QS JS", True), ("8H 9H QS JS TH", False), ("AS 3S 4S 8S 2S", True), ) _UpperCamelCase : Any = ( ("2H 3H 4H 5H 6H", True), ("AS AH 2H AD AC", False), ("2H 3H 5H 6H 7H", False), ("KS AS TS QS JS", True), ("8H 9H QS JS TH", True), ) _UpperCamelCase : List[Any] = ( ("2H 4D 3C AS 5S", True, [5, 4, 3, 2, 14]), ("2H 5D 3C AS 5S", False, [14, 5, 5, 3, 2]), ("JH QD KC AS TS", False, [14, 13, 12, 11, 10]), ("9D 3S 2C 7S 7C", False, [9, 7, 7, 3, 2]), ) _UpperCamelCase : Any = ( ("JH AH TH KH QH", 0), ("JH 9H TH KH QH", 0), ("JC KH JS JD JH", 7), ("KH KC 3S 3H 3D", 6), ("8C 9C 5C 3C TC", 0), ("JS QS 9H TS KH", 0), ("7C 7S KH 2H 7H", 3), ("3C KH 5D 5S KH", 2), ("QH 8H KD JH 8S", 1), ("2D 6D 9D TH 7D", 0), ) _UpperCamelCase : List[Any] = ( ("JH AH TH KH QH", 23), ("JH 9H TH KH QH", 22), ("JC KH JS JD JH", 21), ("KH KC 3S 3H 3D", 20), ("8C 9C 5C 3C TC", 19), ("JS QS 9H TS KH", 18), ("7C 7S KH 2H 7H", 17), ("3C KH 5D 5S KH", 16), ("QH 8H KD JH 8S", 15), ("2D 6D 9D TH 7D", 14), ) def a_ ( ): '''simple docstring''' lowercase__ , lowercase__ : Dict = randrange(len(_lowerCAmelCase ) ), randrange(len(_lowerCAmelCase ) ) lowercase__ : str = ['Loss', 'Tie', 'Win'][(play >= oppo) + (play > oppo)] lowercase__ , lowercase__ : Dict = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def a_ ( _lowerCAmelCase : int = 100 ): '''simple docstring''' return (generate_random_hand() for _ in range(_lowerCAmelCase )) @pytest.mark.parametrize('hand, expected' , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] ): '''simple docstring''' assert PokerHand(_lowerCAmelCase )._is_flush() == expected @pytest.mark.parametrize('hand, expected' , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' assert PokerHand(_lowerCAmelCase )._is_straight() == expected @pytest.mark.parametrize('hand, expected, card_values' , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any ): '''simple docstring''' lowercase__ : Any = PokerHand(_lowerCAmelCase ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize('hand, expected' , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Tuple ): '''simple docstring''' assert PokerHand(_lowerCAmelCase )._is_same_kind() == expected @pytest.mark.parametrize('hand, expected' , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' assert PokerHand(_lowerCAmelCase )._hand_type == expected @pytest.mark.parametrize('hand, other, expected' , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' assert PokerHand(_lowerCAmelCase ).compare_with(PokerHand(_lowerCAmelCase ) ) == expected @pytest.mark.parametrize('hand, other, expected' , generate_random_hands() ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ): '''simple docstring''' assert PokerHand(_lowerCAmelCase ).compare_with(PokerHand(_lowerCAmelCase ) ) == expected def a_ ( ): '''simple docstring''' lowercase__ : List[str] = [PokerHand(_lowerCAmelCase ) for hand in SORTED_HANDS] lowercase__ : int = poker_hands.copy() shuffle(_lowerCAmelCase ) lowercase__ : Optional[int] = chain(sorted(_lowerCAmelCase ) ) for index, hand in enumerate(_lowerCAmelCase ): assert hand == poker_hands[index] def a_ ( ): '''simple docstring''' lowercase__ : List[str] = [PokerHand('2D AC 3H 4H 5S' ), PokerHand('2S 3H 4H 5S 6C' )] pokerhands.sort(reverse=_lowerCAmelCase ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def a_ ( ): '''simple docstring''' lowercase__ : int = PokerHand('2C 4S AS 3D 5C' ) lowercase__ : Optional[Any] = True lowercase__ : Optional[int] = [5, 4, 3, 2, 14] for _ in range(10 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def a_ ( ): '''simple docstring''' lowercase__ : Tuple = 0 lowercase__ : Tuple = os.path.abspath(os.path.dirname(_lowerCAmelCase ) ) lowercase__ : Any = os.path.join(_lowerCAmelCase , 'poker_hands.txt' ) with open(_lowerCAmelCase ) as file_hand: for line in file_hand: lowercase__ : List[str] = line[:14].strip() lowercase__ : Tuple = line[15:].strip() lowercase__ , lowercase__ : List[Any] = PokerHand(_lowerCAmelCase ), PokerHand(_lowerCAmelCase ) lowercase__ : Optional[int] = player.compare_with(_lowerCAmelCase ) if output == "Win": answer += 1 assert answer == 376
77
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({}) lowerCamelCase__ : str = "text" @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return {self.text_column: "text"}
77
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : int = UnCLIPImageVariationPipeline lowerCamelCase__ : int = IMAGE_VARIATION_PARAMS - {"height", "width", "guidance_scale"} lowerCamelCase__ : Any = IMAGE_VARIATION_BATCH_PARAMS lowerCamelCase__ : Dict = [ "generator", "return_dict", "decoder_num_inference_steps", "super_res_num_inference_steps", ] lowerCamelCase__ : Dict = False @property def _UpperCAmelCase ( self ) -> List[Any]: return 3_2 @property def _UpperCAmelCase ( self ) -> Optional[int]: return 3_2 @property def _UpperCAmelCase ( self ) -> Union[str, Any]: return self.time_input_dim @property def _UpperCAmelCase ( self ) -> Optional[Any]: return self.time_input_dim * 4 @property def _UpperCAmelCase ( self ) -> List[Any]: return 1_0_0 @property def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Optional[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def _UpperCAmelCase ( self ) -> Optional[int]: torch.manual_seed(0 ) lowercase__ : List[str] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(a ) @property def _UpperCAmelCase ( self ) -> List[Any]: torch.manual_seed(0 ) lowercase__ : List[Any] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=3_2 , intermediate_size=3_7 , patch_size=1 , ) return CLIPVisionModelWithProjection(a ) @property def _UpperCAmelCase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) lowercase__ : Dict = { 'clip_embeddings_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'cross_attention_dim': self.cross_attention_dim, } lowercase__ : Dict = UnCLIPTextProjModel(**a ) return model @property def _UpperCAmelCase ( self ) -> str: torch.manual_seed(0 ) lowercase__ : Union[str, Any] = { 'sample_size': 3_2, # RGB in channels 'in_channels': 3, # Out channels is double in channels because predicts mean and variance 'out_channels': 6, 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': 'identity', } lowercase__ : Optional[int] = UNetaDConditionModel(**a ) return model @property def _UpperCAmelCase ( self ) -> str: return { "sample_size": 6_4, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def _UpperCAmelCase ( self ) -> Any: torch.manual_seed(0 ) lowercase__ : Dict = UNetaDModel(**self.dummy_super_res_kwargs ) return model @property def _UpperCAmelCase ( self ) -> Tuple: # seeded differently to get different unet than `self.dummy_super_res_first` torch.manual_seed(1 ) lowercase__ : Tuple = UNetaDModel(**self.dummy_super_res_kwargs ) return model def _UpperCAmelCase ( self ) -> str: lowercase__ : Union[str, Any] = self.dummy_decoder lowercase__ : Dict = self.dummy_text_proj lowercase__ : List[Any] = self.dummy_text_encoder lowercase__ : Tuple = self.dummy_tokenizer lowercase__ : List[str] = self.dummy_super_res_first lowercase__ : List[Any] = self.dummy_super_res_last lowercase__ : Dict = UnCLIPScheduler( variance_type='learned_range' , prediction_type='epsilon' , num_train_timesteps=1_0_0_0 , ) lowercase__ : Tuple = UnCLIPScheduler( variance_type='fixed_small_log' , prediction_type='epsilon' , num_train_timesteps=1_0_0_0 , ) lowercase__ : Optional[Any] = CLIPImageProcessor(crop_size=3_2 , size=3_2 ) lowercase__ : Any = self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def _UpperCAmelCase ( self , a , a=0 , a=True ) -> Union[str, Any]: lowercase__ : Union[str, Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(a ) ).to(a ) if str(a ).startswith('mps' ): lowercase__ : List[Any] = torch.manual_seed(a ) else: lowercase__ : Any = torch.Generator(device=a ).manual_seed(a ) if pil_image: lowercase__ : Dict = input_image * 0.5 + 0.5 lowercase__ : Tuple = input_image.clamp(0 , 1 ) lowercase__ : Optional[Any] = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() lowercase__ : List[str] = DiffusionPipeline.numpy_to_pil(a )[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : Optional[int] = 'cpu' lowercase__ : List[str] = self.get_dummy_components() lowercase__ : Any = self.pipeline_class(**a ) lowercase__ : Optional[Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowercase__ : Union[str, Any] = self.get_dummy_inputs(a , pil_image=a ) lowercase__ : Optional[int] = pipe(**a ) lowercase__ : int = output.images lowercase__ : Tuple = self.get_dummy_inputs(a , pil_image=a ) lowercase__ : Union[str, Any] = pipe( **a , return_dict=a , )[0] lowercase__ : Dict = image[0, -3:, -3:, -1] lowercase__ : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) lowercase__ : Optional[int] = np.array( [ 0.9_997, 0.0_002, 0.9_997, 0.9_997, 0.9_969, 0.0_023, 0.9_997, 0.9_969, 0.9_970, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _UpperCAmelCase ( self ) -> str: lowercase__ : Any = 'cpu' lowercase__ : Optional[int] = self.get_dummy_components() lowercase__ : Union[str, Any] = self.pipeline_class(**a ) lowercase__ : Union[str, Any] = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowercase__ : List[str] = self.get_dummy_inputs(a , pil_image=a ) lowercase__ : List[Any] = pipe(**a ) lowercase__ : List[str] = output.images lowercase__ : Any = self.get_dummy_inputs(a , pil_image=a ) lowercase__ : Tuple = pipe( **a , return_dict=a , )[0] lowercase__ : str = image[0, -3:, -3:, -1] lowercase__ : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) lowercase__ : Any = np.array([0.9_997, 0.0_003, 0.9_997, 0.9_997, 0.9_970, 0.0_024, 0.9_997, 0.9_971, 0.9_971] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Union[str, Any] = 'cpu' lowercase__ : Tuple = self.get_dummy_components() lowercase__ : List[str] = self.pipeline_class(**a ) lowercase__ : Tuple = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowercase__ : List[Any] = self.get_dummy_inputs(a , pil_image=a ) lowercase__ : Optional[int] = [ pipeline_inputs['image'], pipeline_inputs['image'], ] lowercase__ : Any = pipe(**a ) lowercase__ : List[str] = output.images lowercase__ : List[Any] = self.get_dummy_inputs(a , pil_image=a ) lowercase__ : Optional[Any] = [ tuple_pipeline_inputs['image'], tuple_pipeline_inputs['image'], ] lowercase__ : List[str] = pipe( **a , return_dict=a , )[0] lowercase__ : Optional[int] = image[0, -3:, -3:, -1] lowercase__ : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 6_4, 6_4, 3) lowercase__ : Tuple = np.array( [ 0.9_997, 0.9_989, 0.0_008, 0.0_021, 0.9_960, 0.0_018, 0.0_014, 0.0_002, 0.9_933, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : int = torch.device('cpu' ) class UpperCAmelCase_ : lowerCamelCase__ : Union[str, Any] = 1 lowercase__ : str = self.get_dummy_components() lowercase__ : str = self.pipeline_class(**a ) lowercase__ : Any = pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowercase__ : Optional[Any] = torch.Generator(device=a ).manual_seed(0 ) lowercase__ : int = pipe.decoder.dtype lowercase__ : Optional[int] = 1 lowercase__ : Optional[int] = ( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) lowercase__ : Tuple = pipe.prepare_latents( a , dtype=a , device=a , generator=a , latents=a , scheduler=DummyScheduler() ) lowercase__ : List[Any] = ( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) lowercase__ : str = pipe.prepare_latents( a , dtype=a , device=a , generator=a , latents=a , scheduler=DummyScheduler() ) lowercase__ : Dict = self.get_dummy_inputs(a , pil_image=a ) lowercase__ : Optional[Any] = pipe( **a , decoder_latents=a , super_res_latents=a ).images lowercase__ : Union[str, Any] = self.get_dummy_inputs(a , pil_image=a ) # Don't pass image, instead pass embedding lowercase__ : Optional[Any] = pipeline_inputs.pop('image' ) lowercase__ : Dict = pipe.image_encoder(a ).image_embeds lowercase__ : List[str] = pipe( **a , decoder_latents=a , super_res_latents=a , image_embeddings=a , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a ).max() < 1e-4 @skip_mps def _UpperCAmelCase ( self ) -> str: lowercase__ : List[Any] = torch_device == 'cpu' # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor lowercase__ : Optional[int] = 1e-2 self._test_attention_slicing_forward_pass( test_max_difference=a , expected_max_diff=a ) @skip_mps def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : List[Any] = torch_device == 'cpu' lowercase__ : List[Any] = True lowercase__ : List[Any] = [ 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] self._test_inference_batch_single_identical( test_max_difference=a , relax_max_difference=a , additional_params_copy_to_batched_inputs=a , ) def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Optional[int] = [ 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes lowercase__ : Union[str, Any] = [2, 3] self._test_inference_batch_consistent( batch_sizes=a , additional_params_copy_to_batched_inputs=a , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=a ) @skip_mps def _UpperCAmelCase ( self ) -> Any: return super().test_dict_tuple_outputs_equivalent() @skip_mps def _UpperCAmelCase ( self ) -> List[Any]: return super().test_save_load_local() @skip_mps def _UpperCAmelCase ( self ) -> Tuple: return super().test_save_load_optional_components() @slow @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Union[str, Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png' ) lowercase__ : Tuple = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/unclip/karlo_v1_alpha_cat_variation_fp16.npy' ) lowercase__ : Optional[Any] = UnCLIPImageVariationPipeline.from_pretrained( 'kakaobrain/karlo-v1-alpha-image-variations' , torch_dtype=torch.floataa ) lowercase__ : int = pipeline.to(a ) pipeline.set_progress_bar_config(disable=a ) lowercase__ : Any = torch.Generator(device='cpu' ).manual_seed(0 ) lowercase__ : Dict = pipeline( a , generator=a , output_type='np' , ) lowercase__ : List[Any] = output.images[0] assert image.shape == (2_5_6, 2_5_6, 3) assert_mean_pixel_difference(a , a , 1_5 )
77
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
77
1
"""simple docstring""" import math import flax.linen as nn import jax.numpy as jnp def a_ ( _lowerCAmelCase : jnp.ndarray , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1.0E4 , _lowerCAmelCase : bool = False , _lowerCAmelCase : float = 1.0 , ): '''simple docstring''' assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, f"""Embedding dimension {embedding_dim} should be even""" lowercase__ : Optional[Any] = float(embedding_dim // 2 ) lowercase__ : Optional[Any] = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) lowercase__ : Any = min_timescale * jnp.exp(jnp.arange(_lowerCAmelCase , dtype=jnp.floataa ) * -log_timescale_increment ) lowercase__ : Dict = jnp.expand_dims(_lowerCAmelCase , 1 ) * jnp.expand_dims(_lowerCAmelCase , 0 ) # scale embeddings lowercase__ : List[str] = scale * emb if flip_sin_to_cos: lowercase__ : Dict = jnp.concatenate([jnp.cos(_lowerCAmelCase ), jnp.sin(_lowerCAmelCase )] , axis=1 ) else: lowercase__ : Optional[int] = jnp.concatenate([jnp.sin(_lowerCAmelCase ), jnp.cos(_lowerCAmelCase )] , axis=1 ) lowercase__ : List[Any] = jnp.reshape(_lowerCAmelCase , [jnp.shape(_lowerCAmelCase )[0], embedding_dim] ) return signal class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : jnp.dtype = jnp.floataa @nn.compact def __call__( self , a ) -> Any: lowercase__ : str = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_1' )(a ) lowercase__ : Union[str, Any] = nn.silu(a ) lowercase__ : Optional[Any] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_2' )(a ) return temb class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : bool = False lowerCamelCase__ : float = 1 @nn.compact def __call__( self , a ) -> str: return get_sinusoidal_embeddings( a , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
77
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING _UpperCamelCase : Dict = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , **a ) -> Dict: super().__init__(**a ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type(a ) def __call__( self , a , a = None , **a , ) -> List[str]: if "text_queries" in kwargs: lowercase__ : Optional[Any] = kwargs.pop('text_queries' ) if isinstance(a , (str, Image.Image) ): lowercase__ : Optional[Any] = {'image': image, 'candidate_labels': candidate_labels} else: lowercase__ : List[str] = image lowercase__ : Optional[Any] = super().__call__(a , **a ) return results def _UpperCAmelCase ( self , **a ) -> Dict: lowercase__ : Optional[Any] = {} if "threshold" in kwargs: lowercase__ : Tuple = kwargs['threshold'] if "top_k" in kwargs: lowercase__ : List[Any] = kwargs['top_k'] return {}, {}, postprocess_params def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : Any = load_image(inputs['image'] ) lowercase__ : Optional[int] = inputs['candidate_labels'] if isinstance(a , a ): lowercase__ : Optional[int] = candidate_labels.split(',' ) lowercase__ : Optional[int] = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(a ): lowercase__ : List[str] = self.tokenizer(a , return_tensors=self.framework ) lowercase__ : List[Any] = self.image_processor(a , return_tensors=self.framework ) yield { "is_last": i == len(a ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : List[Any] = model_inputs.pop('target_size' ) lowercase__ : Dict = model_inputs.pop('candidate_label' ) lowercase__ : Dict = model_inputs.pop('is_last' ) lowercase__ : Optional[int] = self.model(**a ) lowercase__ : Any = {'target_size': target_size, 'candidate_label': candidate_label, 'is_last': is_last, **outputs} return model_outputs def _UpperCAmelCase ( self , a , a=0.1 , a=None ) -> Union[str, Any]: lowercase__ : Dict = [] for model_output in model_outputs: lowercase__ : List[Any] = model_output['candidate_label'] lowercase__ : Optional[int] = BaseModelOutput(a ) lowercase__ : Any = self.image_processor.post_process_object_detection( outputs=a , threshold=a , target_sizes=model_output['target_size'] )[0] for index in outputs["scores"].nonzero(): lowercase__ : Union[str, Any] = outputs['scores'][index].item() lowercase__ : Tuple = self._get_bounding_box(outputs['boxes'][index][0] ) lowercase__ : Tuple = {'score': score, 'label': label, 'box': box} results.append(a ) lowercase__ : Dict = sorted(a , key=lambda a : x["score"] , reverse=a ) if top_k: lowercase__ : Dict = results[:top_k] return results def _UpperCAmelCase ( self , a ) -> Dict[str, int]: if self.framework != "pt": raise ValueError('The ZeroShotObjectDetectionPipeline is only available in PyTorch.' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ : Tuple = box.int().tolist() lowercase__ : Any = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
77
1
"""simple docstring""" import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _UpperCamelCase : Optional[Any] = logging.getLogger() @unittest.skip("Temporarily disable the doc tests.") @require_torch @require_tf @slow class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self , a , a = None , a = None , a = None , a = True , ) -> int: lowercase__ : Any = [file for file in os.listdir(a ) if os.path.isfile(os.path.join(a , a ) )] if identifier is not None: lowercase__ : Optional[Any] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(a , a ): for n_ in n_identifier: lowercase__ : Tuple = [file for file in files if n_ not in file] else: lowercase__ : int = [file for file in files if n_identifier not in file] lowercase__ : Optional[int] = ignore_files or [] ignore_files.append('__init__.py' ) lowercase__ : Any = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' , a ) if only_modules: lowercase__ : int = file.split('.' )[0] try: lowercase__ : Any = getattr(a , a ) lowercase__ : Optional[Any] = doctest.DocTestSuite(a ) lowercase__ : Tuple = unittest.TextTestRunner().run(a ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(f"""{module_identifier} is not a module.""" ) else: lowercase__ : Union[str, Any] = doctest.testfile(str('..' / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def _UpperCAmelCase ( self ) -> int: lowercase__ : List[str] = Path('src/transformers' ) lowercase__ : str = 'modeling' lowercase__ : int = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(a , identifier=a , ignore_files=a ) def _UpperCAmelCase ( self ) -> Dict: lowercase__ : Any = Path('src/transformers' ) lowercase__ : int = 'tokenization' self.analyze_directory(a , identifier=a ) def _UpperCAmelCase ( self ) -> str: lowercase__ : Dict = Path('src/transformers' ) lowercase__ : str = 'configuration' self.analyze_directory(a , identifier=a ) def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : List[Any] = Path('src/transformers' ) lowercase__ : Dict = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(a , n_identifier=a ) def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : List[str] = Path('docs/source' ) lowercase__ : str = ['favicon.ico'] self.analyze_directory(a , ignore_files=a , only_modules=a )
77
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ : def __init__( self , a , a=1_2 , a=7 , a=True , a=True , a=True , a=9_9 , a=3_2 , a=3_2 , a=2 , a=4 , a=3_7 , a=0.1 , a=0.1 , a=5_1_2 , a=0.02 , a=0 , a=None , ) -> Union[str, Any]: lowercase__ : Any = parent lowercase__ : str = batch_size lowercase__ : List[Any] = seq_length lowercase__ : Union[str, Any] = is_training lowercase__ : List[str] = use_input_mask lowercase__ : int = use_labels lowercase__ : List[Any] = vocab_size lowercase__ : str = hidden_size lowercase__ : int = projection_dim lowercase__ : Optional[int] = num_hidden_layers lowercase__ : Any = num_attention_heads lowercase__ : Optional[Any] = intermediate_size lowercase__ : Optional[Any] = dropout lowercase__ : Optional[int] = attention_dropout lowercase__ : Optional[int] = max_position_embeddings lowercase__ : str = initializer_range lowercase__ : Tuple = scope lowercase__ : int = bos_token_id def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : int = None if self.use_input_mask: lowercase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowercase__ : int = input_mask.numpy() lowercase__ , lowercase__ : Tuple = input_mask.shape lowercase__ : List[str] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a ): lowercase__ : Dict = 1 lowercase__ : Union[str, Any] = 0 lowercase__ : Tuple = self.get_config() return config, input_ids, tf.convert_to_tensor(a ) def _UpperCAmelCase ( self ) -> List[Any]: return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _UpperCAmelCase ( self , a , a , a ) -> Any: lowercase__ : List[Any] = TFBlipTextModel(config=a ) lowercase__ : Optional[int] = model(a , attention_mask=a , training=a ) lowercase__ : List[str] = model(a , training=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _UpperCAmelCase ( self ) -> Any: lowercase__ : Optional[Any] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ : Any = config_and_inputs lowercase__ : Any = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : Dict = (TFBlipTextModel,) if is_tf_available() else () lowerCamelCase__ : Optional[Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Optional[int] = BlipTextModelTester(self ) lowercase__ : int = ConfigTester(self , config_class=a , hidden_size=3_7 ) def _UpperCAmelCase ( self ) -> Tuple: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> int: lowercase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _UpperCAmelCase ( self ) -> Optional[Any]: pass def _UpperCAmelCase ( self ) -> Optional[int]: pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def _UpperCAmelCase ( self ) -> List[str]: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> Dict: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> str: pass @slow def _UpperCAmelCase ( self ) -> int: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Any = TFBlipTextModel.from_pretrained(a ) self.assertIsNotNone(a ) def _UpperCAmelCase ( self , a=True ) -> List[str]: super().test_pt_tf_model_equivalence(allow_missing_keys=a )
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : Optional[int] = [0] * len(_lowerCAmelCase ) lowercase__ : Tuple = [] lowercase__ : Tuple = [] lowercase__ : str = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(_lowerCAmelCase ) ): if indegree[i] == 0: queue.append(_lowerCAmelCase ) while queue: lowercase__ : Any = queue.pop(0 ) cnt += 1 topo.append(_lowerCAmelCase ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(_lowerCAmelCase ) if cnt != len(_lowerCAmelCase ): print('Cycle exists' ) else: print(_lowerCAmelCase ) # Adjacency List of Graph _UpperCamelCase : Optional[Any] = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
77
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase : Tuple = logging.get_logger(__name__) _UpperCamelCase : Union[str, Any] = { "alibaba-damo/mgp-str-base": "https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json", } class UpperCAmelCase_ ( _a): lowerCamelCase__ : List[Any] = "mgp-str" def __init__( self , a=[3_2, 1_2_8] , a=4 , a=3 , a=2_7 , a=3_8 , a=5_0_2_5_7 , a=3_0_5_2_2 , a=7_6_8 , a=1_2 , a=1_2 , a=4.0 , a=True , a=False , a=1e-5 , a=0.0 , a=0.0 , a=0.0 , a=False , a=0.02 , **a , ) -> Tuple: super().__init__(**a ) lowercase__ : int = image_size lowercase__ : List[Any] = patch_size lowercase__ : Optional[Any] = num_channels lowercase__ : Optional[Any] = max_token_length lowercase__ : Dict = num_character_labels lowercase__ : Optional[int] = num_bpe_labels lowercase__ : Dict = num_wordpiece_labels lowercase__ : Tuple = hidden_size lowercase__ : List[Any] = num_hidden_layers lowercase__ : Union[str, Any] = num_attention_heads lowercase__ : Dict = mlp_ratio lowercase__ : Optional[int] = distilled lowercase__ : Optional[int] = layer_norm_eps lowercase__ : Optional[int] = drop_rate lowercase__ : List[str] = qkv_bias lowercase__ : Optional[int] = attn_drop_rate lowercase__ : Any = drop_path_rate lowercase__ : List[Any] = output_aa_attentions lowercase__ : Tuple = initializer_range
77
1
"""simple docstring""" from collections.abc import Callable class UpperCAmelCase_ : def __init__( self , a = None ) -> None: # Stores actual heap items. lowercase__ : list = [] # Stores indexes of each item for supporting updates and deletion. lowercase__ : dict = {} # Stores current size of heap. lowercase__ : Any = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. lowercase__ : str = key or (lambda a : x) def _UpperCAmelCase ( self , a ) -> int | None: return int((i - 1) / 2 ) if i > 0 else None def _UpperCAmelCase ( self , a ) -> int | None: lowercase__ : List[Any] = int(2 * i + 1 ) return left if 0 < left < self.size else None def _UpperCAmelCase ( self , a ) -> int | None: lowercase__ : Union[str, Any] = int(2 * i + 2 ) return right if 0 < right < self.size else None def _UpperCAmelCase ( self , a , a ) -> None: lowercase__ , lowercase__ : str = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. lowercase__ , lowercase__ : Optional[int] = self.arr[j], self.arr[i] def _UpperCAmelCase ( self , a , a ) -> bool: return self.arr[i][1] < self.arr[j][1] def _UpperCAmelCase ( self , a ) -> int: lowercase__ : Any = self._left(a ) lowercase__ : Optional[Any] = self._right(a ) lowercase__ : Optional[int] = i if left is not None and not self._cmp(a , a ): lowercase__ : str = left if right is not None and not self._cmp(a , a ): lowercase__ : List[str] = right return valid_parent def _UpperCAmelCase ( self , a ) -> None: lowercase__ : List[Any] = self._parent(a ) while parent is not None and not self._cmp(a , a ): self._swap(a , a ) lowercase__ , lowercase__ : int = parent, self._parent(a ) def _UpperCAmelCase ( self , a ) -> None: lowercase__ : int = self._get_valid_parent(a ) while valid_parent != index: self._swap(a , a ) lowercase__ , lowercase__ : int = valid_parent, self._get_valid_parent(a ) def _UpperCAmelCase ( self , a , a ) -> None: if item not in self.pos_map: return lowercase__ : List[Any] = self.pos_map[item] lowercase__ : Optional[int] = [item, self.key(a )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(a ) self._heapify_down(a ) def _UpperCAmelCase ( self , a ) -> None: if item not in self.pos_map: return lowercase__ : int = self.pos_map[item] del self.pos_map[item] lowercase__ : List[Any] = self.arr[self.size - 1] lowercase__ : List[Any] = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(a ) self._heapify_down(a ) def _UpperCAmelCase ( self , a , a ) -> None: lowercase__ : Dict = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(a )] ) else: lowercase__ : Any = [item, self.key(a )] lowercase__ : List[str] = self.size self.size += 1 self._heapify_up(self.size - 1 ) def _UpperCAmelCase ( self ) -> tuple | None: return self.arr[0] if self.size else None def _UpperCAmelCase ( self ) -> tuple | None: lowercase__ : str = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def a_ ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : int = FunnelConfig.from_json_file(_lowerCAmelCase ) print(f"""Building PyTorch model from configuration: {config}""" ) lowercase__ : List[Any] = FunnelBaseModel(_lowerCAmelCase ) if base_model else FunnelModel(_lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_funnel(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _lowerCAmelCase ) if __name__ == "__main__": _UpperCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--base_model", action="store_true", help="Whether you want just the base model (no decoder) or not." ) _UpperCamelCase : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
77
1
"""simple docstring""" import argparse import os import re import tensorflow as tf import torch from transformers import BertConfig, BertModel from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase : Optional[Any] = logging.get_logger(__name__) def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : int , _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : List[str] = os.path.abspath(_lowerCAmelCase ) logger.info(f"""Converting TensorFlow checkpoint from {tf_path}""" ) # Load weights from TF model lowercase__ : int = tf.train.list_variables(_lowerCAmelCase ) lowercase__ : List[Any] = [] lowercase__ : List[Any] = [] lowercase__ : Dict = [] for full_name, shape in init_vars: # logger.info(f"Loading TF weight {name} with shape {shape}") lowercase__ : int = full_name.split('/' ) if full_name == "_CHECKPOINTABLE_OBJECT_GRAPH" or name[0] in ["global_step", "save_counter"]: logger.info(f"""Skipping non-model layer {full_name}""" ) continue if "optimizer" in full_name: logger.info(f"""Skipping optimization layer {full_name}""" ) continue if name[0] == "model": # ignore initial 'model' lowercase__ : List[Any] = name[1:] # figure out how many levels deep the name is lowercase__ : Dict = 0 for _name in name: if _name.startswith('layer_with_weights' ): depth += 1 else: break layer_depth.append(_lowerCAmelCase ) # read data lowercase__ : Any = tf.train.load_variable(_lowerCAmelCase , _lowerCAmelCase ) names.append('/'.join(_lowerCAmelCase ) ) arrays.append(_lowerCAmelCase ) logger.info(f"""Read a total of {len(_lowerCAmelCase ):,} layers""" ) # Sanity check if len(set(_lowerCAmelCase ) ) != 1: raise ValueError(f"""Found layer names with different depths (layer depth {list(set(_lowerCAmelCase ) )})""" ) lowercase__ : Optional[int] = list(set(_lowerCAmelCase ) )[0] if layer_depth != 1: raise ValueError( 'The model contains more than just the embedding/encoder layers. This script does not handle MLM/NSP' ' heads.' ) # convert layers logger.info('Converting weights...' ) for full_name, array in zip(_lowerCAmelCase , _lowerCAmelCase ): lowercase__ : Union[str, Any] = full_name.split('/' ) lowercase__ : Any = model lowercase__ : List[str] = [] for i, m_name in enumerate(_lowerCAmelCase ): if m_name == ".ATTRIBUTES": # variable names end with .ATTRIBUTES/VARIABLE_VALUE break if m_name.startswith('layer_with_weights' ): lowercase__ : Optional[int] = int(m_name.split('-' )[-1] ) if layer_num <= 2: # embedding layers # layer_num 0: word_embeddings # layer_num 1: position_embeddings # layer_num 2: token_type_embeddings continue elif layer_num == 3: # embedding LayerNorm trace.extend(['embeddings', 'LayerNorm'] ) lowercase__ : Union[str, Any] = getattr(_lowerCAmelCase , 'embeddings' ) lowercase__ : Dict = getattr(_lowerCAmelCase , 'LayerNorm' ) elif layer_num > 3 and layer_num < config.num_hidden_layers + 4: # encoder layers trace.extend(['encoder', 'layer', str(layer_num - 4 )] ) lowercase__ : int = getattr(_lowerCAmelCase , 'encoder' ) lowercase__ : List[str] = getattr(_lowerCAmelCase , 'layer' ) lowercase__ : List[str] = pointer[layer_num - 4] elif layer_num == config.num_hidden_layers + 4: # pooler layer trace.extend(['pooler', 'dense'] ) lowercase__ : List[str] = getattr(_lowerCAmelCase , 'pooler' ) lowercase__ : int = getattr(_lowerCAmelCase , 'dense' ) elif m_name == "embeddings": trace.append('embeddings' ) lowercase__ : int = getattr(_lowerCAmelCase , 'embeddings' ) if layer_num == 0: trace.append('word_embeddings' ) lowercase__ : List[Any] = getattr(_lowerCAmelCase , 'word_embeddings' ) elif layer_num == 1: trace.append('position_embeddings' ) lowercase__ : List[str] = getattr(_lowerCAmelCase , 'position_embeddings' ) elif layer_num == 2: trace.append('token_type_embeddings' ) lowercase__ : Optional[Any] = getattr(_lowerCAmelCase , 'token_type_embeddings' ) else: raise ValueError(f"""Unknown embedding layer with name {full_name}""" ) trace.append('weight' ) lowercase__ : List[str] = getattr(_lowerCAmelCase , 'weight' ) elif m_name == "_attention_layer": # self-attention layer trace.extend(['attention', 'self'] ) lowercase__ : int = getattr(_lowerCAmelCase , 'attention' ) lowercase__ : str = getattr(_lowerCAmelCase , 'self' ) elif m_name == "_attention_layer_norm": # output attention norm trace.extend(['attention', 'output', 'LayerNorm'] ) lowercase__ : Optional[int] = getattr(_lowerCAmelCase , 'attention' ) lowercase__ : Optional[Any] = getattr(_lowerCAmelCase , 'output' ) lowercase__ : List[str] = getattr(_lowerCAmelCase , 'LayerNorm' ) elif m_name == "_attention_output_dense": # output attention dense trace.extend(['attention', 'output', 'dense'] ) lowercase__ : List[Any] = getattr(_lowerCAmelCase , 'attention' ) lowercase__ : Union[str, Any] = getattr(_lowerCAmelCase , 'output' ) lowercase__ : List[Any] = getattr(_lowerCAmelCase , 'dense' ) elif m_name == "_output_dense": # output dense trace.extend(['output', 'dense'] ) lowercase__ : Optional[int] = getattr(_lowerCAmelCase , 'output' ) lowercase__ : List[Any] = getattr(_lowerCAmelCase , 'dense' ) elif m_name == "_output_layer_norm": # output dense trace.extend(['output', 'LayerNorm'] ) lowercase__ : Tuple = getattr(_lowerCAmelCase , 'output' ) lowercase__ : List[str] = getattr(_lowerCAmelCase , 'LayerNorm' ) elif m_name == "_key_dense": # attention key trace.append('key' ) lowercase__ : Optional[int] = getattr(_lowerCAmelCase , 'key' ) elif m_name == "_query_dense": # attention query trace.append('query' ) lowercase__ : List[str] = getattr(_lowerCAmelCase , 'query' ) elif m_name == "_value_dense": # attention value trace.append('value' ) lowercase__ : Dict = getattr(_lowerCAmelCase , 'value' ) elif m_name == "_intermediate_dense": # attention intermediate dense trace.extend(['intermediate', 'dense'] ) lowercase__ : List[Any] = getattr(_lowerCAmelCase , 'intermediate' ) lowercase__ : Any = getattr(_lowerCAmelCase , 'dense' ) elif m_name == "_output_layer_norm": # output layer norm trace.append('output' ) lowercase__ : Optional[Any] = getattr(_lowerCAmelCase , 'output' ) # weights & biases elif m_name in ["bias", "beta"]: trace.append('bias' ) lowercase__ : Optional[Any] = getattr(_lowerCAmelCase , 'bias' ) elif m_name in ["kernel", "gamma"]: trace.append('weight' ) lowercase__ : str = getattr(_lowerCAmelCase , 'weight' ) else: logger.warning(f"""Ignored {m_name}""" ) # for certain layers reshape is necessary lowercase__ : Dict = '.'.join(_lowerCAmelCase ) if re.match(R'(\S+)\.attention\.self\.(key|value|query)\.(bias|weight)' , _lowerCAmelCase ) or re.match( R'(\S+)\.attention\.output\.dense\.weight' , _lowerCAmelCase ): lowercase__ : Dict = array.reshape(pointer.data.shape ) if "kernel" in full_name: lowercase__ : Optional[int] = array.transpose() if pointer.shape == array.shape: lowercase__ : Dict = torch.from_numpy(_lowerCAmelCase ) else: raise ValueError( f"""Shape mismatch in layer {full_name}: Model expects shape {pointer.shape} but layer contains shape:""" f""" {array.shape}""" ) logger.info(f"""Successfully set variable {full_name} to PyTorch layer {trace}""" ) return model def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : List[Any] ): '''simple docstring''' logger.info(f"""Loading model based on config from {config_path}...""" ) lowercase__ : List[str] = BertConfig.from_json_file(_lowerCAmelCase ) lowercase__ : Any = BertModel(_lowerCAmelCase ) # Load weights from checkpoint logger.info(f"""Loading weights from checkpoint {tf_checkpoint_path}...""" ) load_tfa_weights_in_bert(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model logger.info(f"""Saving PyTorch model to {pytorch_dump_path}...""" ) torch.save(model.state_dict() , _lowerCAmelCase ) if __name__ == "__main__": _UpperCamelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( "--tf_checkpoint_path", type=str, required=True, help="Path to the TensorFlow 2.x checkpoint path." ) parser.add_argument( "--bert_config_file", type=str, required=True, help="The config json file corresponding to the BERT model. This specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", type=str, required=True, help="Path to the output PyTorch model (must include filename).", ) _UpperCamelCase : Tuple = parser.parse_args() convert_tfa_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
77
"""simple docstring""" import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _UpperCamelCase : Any = logging.getLogger(__name__) _UpperCamelCase : Optional[Any] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _UpperCamelCase : Any = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The model checkpoint for weights initialization. Leave None if you want to train a model from" " scratch." ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(_a)} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained config name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "The input training data file (a text file)."}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The input training data files (multiple files in glob format). " "Very often splitting large files to smaller files can prevent tokenizer going out of memory" ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input train ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input eval ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Whether distinct lines of text in the dataset are to be handled as distinct sequences."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Train with masked-language modeling loss instead of language modeling."}) lowerCamelCase__ : bool = field(default=_a , metadata={"help": "Whether ot not to use whole word mask."}) lowerCamelCase__ : float = field( default=0.15 , metadata={"help": "Ratio of tokens to mask for masked language modeling loss"}) lowerCamelCase__ : float = field( default=1 / 6 , metadata={ "help": ( "Ratio of length of a span of masked tokens to surrounding context length for permutation language" " modeling." ) } , ) lowerCamelCase__ : int = field( default=5 , metadata={"help": "Maximum length of a span of masked tokens for permutation language modeling."}) lowerCamelCase__ : int = field( default=-1 , metadata={ "help": ( "Optional input sequence length after tokenization." "The training dataset will be truncated in block of this size for training." "Default to the model max input length for single sentence inputs (take into account special tokens)." ) } , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Overwrite the cached training and evaluation sets"}) def a_ ( _lowerCAmelCase : DataTrainingArguments , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[str] = None , ): '''simple docstring''' def _dataset(_lowerCAmelCase : Any , _lowerCAmelCase : Any=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('You need to set world whole masking and mlm to True for Chinese Whole Word Mask' ) return LineByLineWithRefDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , ref_path=_lowerCAmelCase , ) return LineByLineTextDataset(tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size ) else: return TextDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=_lowerCAmelCase , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(_lowerCAmelCase ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def a_ ( ): '''simple docstring''' lowercase__ : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ : List[Any] = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( 'Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ' 'or remove the --do_eval argument.' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s' , _lowerCAmelCase ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: lowercase__ : List[Any] = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : Any = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: lowercase__ : Union[str, Any] = CONFIG_MAPPING[model_args.model_type]() logger.warning('You are instantiating a new config instance from scratch.' ) if model_args.tokenizer_name: lowercase__ : Dict = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : int = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another' ' script, save it,and load it from here, using --tokenizer_name' ) if model_args.model_name_or_path: lowercase__ : Optional[Any] = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=_lowerCAmelCase , cache_dir=model_args.cache_dir , ) else: logger.info('Training new model from scratch' ) lowercase__ : int = AutoModelWithLMHead.from_config(_lowerCAmelCase ) model.resize_token_embeddings(len(_lowerCAmelCase ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( 'BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the' '--mlm flag (masked language modeling).' ) if data_args.block_size <= 0: lowercase__ : List[str] = tokenizer.max_len # Our input block size will be the max possible for the model else: lowercase__ : int = min(data_args.block_size , tokenizer.max_len ) # Get datasets lowercase__ : Tuple = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) lowercase__ : Optional[Any] = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , evaluate=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": lowercase__ : List[Any] = DataCollatorForPermutationLanguageModeling( tokenizer=_lowerCAmelCase , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: lowercase__ : List[str] = DataCollatorForWholeWordMask( tokenizer=_lowerCAmelCase , mlm_probability=data_args.mlm_probability ) else: lowercase__ : str = DataCollatorForLanguageModeling( tokenizer=_lowerCAmelCase , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer lowercase__ : Optional[int] = Trainer( model=_lowerCAmelCase , args=_lowerCAmelCase , data_collator=_lowerCAmelCase , train_dataset=_lowerCAmelCase , eval_dataset=_lowerCAmelCase , prediction_loss_only=_lowerCAmelCase , ) # Training if training_args.do_train: lowercase__ : List[str] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=_lowerCAmelCase ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase__ : List[Any] = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) lowercase__ : Dict = trainer.evaluate() lowercase__ : List[Any] = math.exp(eval_output['eval_loss'] ) lowercase__ : int = {'perplexity': perplexity} lowercase__ : int = os.path.join(training_args.output_dir , 'eval_results_lm.txt' ) if trainer.is_world_master(): with open(_lowerCAmelCase , 'w' ) as writer: logger.info('***** Eval results *****' ) for key in sorted(result.keys() ): logger.info(' %s = %s' , _lowerCAmelCase , str(result[key] ) ) writer.write('%s = %s\n' % (key, str(result[key] )) ) results.update(_lowerCAmelCase ) return results def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' main() if __name__ == "__main__": main()
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int ): '''simple docstring''' if b == 0: return 1 if (b % 2) == 0: return actual_power(_lowerCAmelCase , int(b / 2 ) ) * actual_power(_lowerCAmelCase , int(b / 2 ) ) else: return a * actual_power(_lowerCAmelCase , int(b / 2 ) ) * actual_power(_lowerCAmelCase , int(b / 2 ) ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int ): '''simple docstring''' if b < 0: return 1 / actual_power(_lowerCAmelCase , _lowerCAmelCase ) return actual_power(_lowerCAmelCase , _lowerCAmelCase ) if __name__ == "__main__": print(power(-2, -3))
77
"""simple docstring""" import math import flax.linen as nn import jax.numpy as jnp def a_ ( _lowerCAmelCase : jnp.ndarray , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1.0E4 , _lowerCAmelCase : bool = False , _lowerCAmelCase : float = 1.0 , ): '''simple docstring''' assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, f"""Embedding dimension {embedding_dim} should be even""" lowercase__ : Optional[Any] = float(embedding_dim // 2 ) lowercase__ : Optional[Any] = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) lowercase__ : Any = min_timescale * jnp.exp(jnp.arange(_lowerCAmelCase , dtype=jnp.floataa ) * -log_timescale_increment ) lowercase__ : Dict = jnp.expand_dims(_lowerCAmelCase , 1 ) * jnp.expand_dims(_lowerCAmelCase , 0 ) # scale embeddings lowercase__ : List[str] = scale * emb if flip_sin_to_cos: lowercase__ : Dict = jnp.concatenate([jnp.cos(_lowerCAmelCase ), jnp.sin(_lowerCAmelCase )] , axis=1 ) else: lowercase__ : Optional[int] = jnp.concatenate([jnp.sin(_lowerCAmelCase ), jnp.cos(_lowerCAmelCase )] , axis=1 ) lowercase__ : List[Any] = jnp.reshape(_lowerCAmelCase , [jnp.shape(_lowerCAmelCase )[0], embedding_dim] ) return signal class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : jnp.dtype = jnp.floataa @nn.compact def __call__( self , a ) -> Any: lowercase__ : str = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_1' )(a ) lowercase__ : Union[str, Any] = nn.silu(a ) lowercase__ : Optional[Any] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_2' )(a ) return temb class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : bool = False lowerCamelCase__ : float = 1 @nn.compact def __call__( self , a ) -> str: return get_sinusoidal_embeddings( a , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
77
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging _UpperCamelCase : Optional[Any] = logging.get_logger(__name__) class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = ["pixel_values"] def __init__( self , a = True , a = None , a = PILImageResampling.BILINEAR , a = True , a = None , a = True , a = 1 / 2_5_5 , a = True , a = None , a = None , **a , ) -> None: super().__init__(**a ) lowercase__ : Dict = size if size is not None else {'shortest_edge': 2_5_6} lowercase__ : Union[str, Any] = get_size_dict(a , default_to_square=a ) lowercase__ : str = crop_size if crop_size is not None else {'height': 2_2_4, 'width': 2_2_4} lowercase__ : Tuple = get_size_dict(a ) lowercase__ : Union[str, Any] = do_resize lowercase__ : List[Any] = size lowercase__ : List[Any] = resample lowercase__ : Any = do_center_crop lowercase__ : Optional[Any] = crop_size lowercase__ : int = do_rescale lowercase__ : int = rescale_factor lowercase__ : Dict = do_normalize lowercase__ : Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowercase__ : Tuple = image_std if image_std is not None else IMAGENET_STANDARD_STD def _UpperCAmelCase ( self , a , a , a = PILImageResampling.BICUBIC , a = None , **a , ) -> np.ndarray: lowercase__ : int = get_size_dict(a , default_to_square=a ) if "shortest_edge" not in size: raise ValueError(f"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) lowercase__ : Optional[int] = get_resize_output_image_size(a , size=size['shortest_edge'] , default_to_square=a ) return resize(a , size=a , resample=a , data_format=a , **a ) def _UpperCAmelCase ( self , a , a , a = None , **a , ) -> np.ndarray: lowercase__ : List[str] = get_size_dict(a ) return center_crop(a , size=(size['height'], size['width']) , data_format=a , **a ) def _UpperCAmelCase ( self , a , a , a = None , **a ) -> np.ndarray: return rescale(a , scale=a , data_format=a , **a ) def _UpperCAmelCase ( self , a , a , a , a = None , **a , ) -> np.ndarray: return normalize(a , mean=a , std=a , data_format=a , **a ) def _UpperCAmelCase ( self , a , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = ChannelDimension.FIRST , **a , ) -> Tuple: lowercase__ : Optional[Any] = do_resize if do_resize is not None else self.do_resize lowercase__ : Tuple = size if size is not None else self.size lowercase__ : Tuple = get_size_dict(a , default_to_square=a ) lowercase__ : Optional[int] = resample if resample is not None else self.resample lowercase__ : List[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop lowercase__ : Union[str, Any] = crop_size if crop_size is not None else self.crop_size lowercase__ : Union[str, Any] = get_size_dict(a ) lowercase__ : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale lowercase__ : Optional[int] = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase__ : Dict = do_normalize if do_normalize is not None else self.do_normalize lowercase__ : List[Any] = image_mean if image_mean is not None else self.image_mean lowercase__ : Optional[Any] = image_std if image_std is not None else self.image_std lowercase__ : Tuple = make_list_of_images(a ) if not valid_images(a ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. lowercase__ : int = [to_numpy_array(a ) for image in images] if do_resize: lowercase__ : List[Any] = [self.resize(image=a , size=a , resample=a ) for image in images] if do_center_crop: lowercase__ : Any = [self.center_crop(image=a , size=a ) for image in images] if do_rescale: lowercase__ : Any = [self.rescale(image=a , scale=a ) for image in images] if do_normalize: lowercase__ : Any = [self.normalize(image=a , mean=a , std=a ) for image in images] lowercase__ : str = [to_channel_dimension_format(a , a ) for image in images] lowercase__ : int = {'pixel_values': images} return BatchFeature(data=a , tensor_type=a )
77
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def a_ ( _lowerCAmelCase : Callable[[int | float], int | float] , _lowerCAmelCase : int | float , _lowerCAmelCase : int | float , _lowerCAmelCase : int = 100 , ): '''simple docstring''' lowercase__ : Dict = x_start lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) lowercase__ : Optional[Any] = 0.0 for _ in range(_lowerCAmelCase ): # Approximates curve as a sequence of linear lines and sums their length lowercase__ : Union[str, Any] = (x_end - x_start) / steps + xa lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step lowercase__ : Union[str, Any] = xa lowercase__ : int = fxa return length if __name__ == "__main__": def a_ ( _lowerCAmelCase : List[Any] ): '''simple docstring''' return math.sin(10 * x ) print("f(x) = sin(10 * x)") print("The length of the curve from x = -10 to x = 10 is:") _UpperCamelCase : str = 10 while i <= 10_00_00: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
77
1
"""simple docstring""" import torch from transformers import AutoModel class UpperCAmelCase_ ( torch.nn.Module): def __init__( self , a="sayef/fsner-bert-base-uncased" ) -> Tuple: super(a , self ).__init__() lowercase__ : str = AutoModel.from_pretrained(a , return_dict=a ) lowercase__ : Any = torch.nn.CosineSimilarity(3 , 1e-08 ) lowercase__ : Union[str, Any] = torch.nn.Softmax(dim=1 ) def _UpperCAmelCase ( self , **a ) -> List[str]: return self.bert(**a ).last_hidden_state def _UpperCAmelCase ( self , a ) -> Optional[Any]: return token_embeddings.sum(2 , keepdim=a ) def _UpperCAmelCase ( self , a , a , a=1 ) -> str: return self.softmax(T * self.cos(a , a ) ) def _UpperCAmelCase ( self , a , a ) -> str: lowercase__ : Union[str, Any] = W_supports['sizes'].tolist() lowercase__ : str = W_supports['start_token_id'].item() lowercase__ : str = W_supports['end_token_id'].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] lowercase__ : Any = self.BERT(**a ) lowercase__ : Any = self.BERT(**a ) lowercase__ : Tuple = None lowercase__ : Dict = None lowercase__ : str = W_supports['input_ids'] == start_token_id lowercase__ : List[str] = W_supports['input_ids'] == end_token_id for i, size in enumerate(a ): if i == 0: lowercase__ : int = 0 else: lowercase__ : Dict = support_sizes[i - 1] lowercase__ : Dict = S[s : s + size][start_token_masks[s : s + size]] lowercase__ : int = S[s : s + size][end_token_masks[s : s + size]] lowercase__ : Any = torch.matmul(q[i] , s_start.T ).sum(1 ).softmax(0 ) lowercase__ : int = torch.matmul(q[i] , s_end.T ).sum(1 ).softmax(0 ) if p_starts is not None: lowercase__ : Tuple = torch.vstack((p_starts, p_start) ) lowercase__ : Tuple = torch.vstack((p_ends, p_end) ) else: lowercase__ : Optional[int] = p_start lowercase__ : Optional[int] = p_end return p_starts, p_ends
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase : Tuple = { "configuration_whisper": ["WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP", "WhisperConfig", "WhisperOnnxConfig"], "feature_extraction_whisper": ["WhisperFeatureExtractor"], "processing_whisper": ["WhisperProcessor"], "tokenization_whisper": ["WhisperTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = ["WhisperTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : str = [ "WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "WhisperForConditionalGeneration", "WhisperModel", "WhisperPreTrainedModel", "WhisperForAudioClassification", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Any = [ "TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFWhisperForConditionalGeneration", "TFWhisperModel", "TFWhisperPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Tuple = [ "FlaxWhisperForConditionalGeneration", "FlaxWhisperModel", "FlaxWhisperPreTrainedModel", "FlaxWhisperForAudioClassification", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _UpperCamelCase : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def a_ ( ): '''simple docstring''' lowercase__ : Any = ArgumentParser( description=( 'PyTorch TPU distributed training launch ' 'helper utility that will spawn up ' 'multiple distributed processes' ) ) # Optional arguments for the launch helper parser.add_argument('--num_cores' , type=_lowerCAmelCase , default=1 , help='Number of TPU cores to use (1 or 8).' ) # positional parser.add_argument( 'training_script' , type=_lowerCAmelCase , help=( 'The full path to the single TPU training ' 'program/script to be launched in parallel, ' 'followed by all the arguments for the ' 'training script' ) , ) # rest from the training program parser.add_argument('training_script_args' , nargs=_lowerCAmelCase ) return parser.parse_args() def a_ ( ): '''simple docstring''' lowercase__ : Union[str, Any] = parse_args() # Import training_script as a module. lowercase__ : List[Any] = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) lowercase__ : Dict = script_fpath.stem lowercase__ : str = importlib.import_module(_lowerCAmelCase ) # Patch sys.argv lowercase__ : Union[str, Any] = [args.training_script] + args.training_script_args + ['--tpu_num_cores', str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
77
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a_ ( _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : int = args.pruning_method lowercase__ : Tuple = args.threshold lowercase__ : str = args.model_name_or_path.rstrip('/' ) lowercase__ : List[Any] = args.target_model_path print(f"""Load fine-pruned model from {model_name_or_path}""" ) lowercase__ : Optional[Any] = torch.load(os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) lowercase__ : List[str] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowercase__ : Tuple = tensor print(f"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowercase__ : List[str] = tensor print(f"""Copied layer {name}""" ) elif "bias" in name: lowercase__ : Optional[Any] = tensor print(f"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowercase__ : Optional[Any] = MagnitudeBinarizer.apply(inputs=_lowerCAmelCase , threshold=_lowerCAmelCase ) lowercase__ : Optional[int] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowercase__ : Optional[Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ : Any = TopKBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[Any] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowercase__ : Any = name[:-6] lowercase__ : Optional[Any] = model[f"""{prefix_}mask_scores"""] lowercase__ : Tuple = ThresholdBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[str] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowercase__ : Union[str, Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ , lowercase__ : Tuple = -0.1, 1.1 lowercase__ : Optional[Any] = torch.sigmoid(_lowerCAmelCase ) lowercase__ : Optional[Any] = s * (r - l) + l lowercase__ : Optional[Any] = s_bar.clamp(min=0.0 , max=1.0 ) lowercase__ : Union[str, Any] = tensor * mask print(f"""Pruned layer {name}""" ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: lowercase__ : Union[str, Any] = os.path.join( os.path.dirname(_lowerCAmelCase ) , f"""bertarized_{os.path.basename(_lowerCAmelCase )}""" ) if not os.path.isdir(_lowerCAmelCase ): shutil.copytree(_lowerCAmelCase , _lowerCAmelCase ) print(f"""\nCreated folder {target_model_path}""" ) torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": _UpperCamelCase : int = argparse.ArgumentParser() parser.add_argument( "--pruning_method", choices=["l0", "magnitude", "topK", "sigmoied_threshold"], type=str, required=True, help=( "Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning," " sigmoied_threshold = Soft movement pruning)" ), ) parser.add_argument( "--threshold", type=float, required=False, help=( "For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model." "For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared." "Not needed for `l0`" ), ) parser.add_argument( "--model_name_or_path", type=str, required=True, help="Folder containing the model that was previously fine-pruned", ) parser.add_argument( "--target_model_path", default=None, type=str, required=False, help="Folder containing the model that was previously fine-pruned", ) _UpperCamelCase : Dict = parser.parse_args() main(args)
77
1
"""simple docstring""" from scipy.stats import pearsonr import datasets _UpperCamelCase : List[Any] = "\nPearson correlation coefficient and p-value for testing non-correlation.\nThe Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases.\nThe p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets.\n" _UpperCamelCase : List[Any] = "\nArgs:\n predictions (`list` of `int`): Predicted class labels, as returned by a model.\n references (`list` of `int`): Ground truth labels.\n return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`.\n\nReturns:\n pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation.\n p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities.\n\nExamples:\n\n Example 1-A simple example using only predictions and references.\n >>> pearsonr_metric = datasets.load_metric(\"pearsonr\")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5])\n >>> print(round(results['pearsonr'], 2))\n -0.74\n\n Example 2-The same as Example 1, but that also returns the `p-value`.\n >>> pearsonr_metric = datasets.load_metric(\"pearsonr\")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True)\n >>> print(sorted(list(results.keys())))\n ['p-value', 'pearsonr']\n >>> print(round(results['pearsonr'], 2))\n -0.74\n >>> print(round(results['p-value'], 2))\n 0.15\n" _UpperCamelCase : Optional[int] = "\n@article{2020SciPy-NMeth,\nauthor = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, Ilhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Antonio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\ntitle = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\njournal = {Nature Methods},\nyear = {2020},\nvolume = {17},\npages = {261--272},\nadsurl = {https://rdcu.be/b08Wh},\ndoi = {10.1038/s41592-019-0686-2},\n}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class UpperCAmelCase_ ( datasets.Metric): def _UpperCAmelCase ( self ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('float' ), 'references': datasets.Value('float' ), } ) , reference_urls=['https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html'] , ) def _UpperCAmelCase ( self , a , a , a=False ) -> Any: if return_pvalue: lowercase__ : List[Any] = pearsonr(a , a ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(a , a )[0] )}
77
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[List[PIL.Image.Image], np.ndarray] lowerCamelCase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : np.ndarray lowerCamelCase__ : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
77
1
"""simple docstring""" import math _UpperCamelCase : Optional[int] = 10 _UpperCamelCase : List[Any] = 7 _UpperCamelCase : str = BALLS_PER_COLOUR * NUM_COLOURS def a_ ( _lowerCAmelCase : int = 20 ): '''simple docstring''' lowercase__ : List[Any] = math.comb(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Union[str, Any] = math.comb(NUM_BALLS - BALLS_PER_COLOUR , _lowerCAmelCase ) lowercase__ : List[str] = NUM_COLOURS * (1 - missing_colour / total) return f"""{result:.9f}""" if __name__ == "__main__": print(solution(20))
77
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class UpperCAmelCase_ : def __init__( self , a , a , a , a , a , a=0.2 , a=0.2 ) -> Dict: lowercase__ : Any = bp_numa lowercase__ : Optional[int] = bp_numa lowercase__ : Tuple = bp_numa lowercase__ : Optional[Any] = conva_get[:2] lowercase__ : Optional[int] = conva_get[2] lowercase__ : Optional[Any] = size_pa lowercase__ : Union[str, Any] = rate_w lowercase__ : Union[str, Any] = rate_t lowercase__ : List[Any] = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ : Optional[Any] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : int = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : Any = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 def _UpperCAmelCase ( self , a ) -> Union[str, Any]: # save model dict with pickle lowercase__ : Optional[Any] = { 'num_bp1': self.num_bpa, 'num_bp2': self.num_bpa, 'num_bp3': self.num_bpa, 'conv1': self.conva, 'step_conv1': self.step_conva, 'size_pooling1': self.size_poolinga, 'rate_weight': self.rate_weight, 'rate_thre': self.rate_thre, 'w_conv1': self.w_conva, 'wkj': self.wkj, 'vji': self.vji, 'thre_conv1': self.thre_conva, 'thre_bp2': self.thre_bpa, 'thre_bp3': self.thre_bpa, } with open(a , 'wb' ) as f: pickle.dump(a , a ) print(f"""Model saved: {save_path}""" ) @classmethod def _UpperCAmelCase ( cls , a ) -> Any: # read saved model with open(a , 'rb' ) as f: lowercase__ : Optional[int] = pickle.load(a ) # noqa: S301 lowercase__ : Optional[int] = model_dic.get('conv1' ) conv_get.append(model_dic.get('step_conv1' ) ) lowercase__ : List[Any] = model_dic.get('size_pooling1' ) lowercase__ : Tuple = model_dic.get('num_bp1' ) lowercase__ : int = model_dic.get('num_bp2' ) lowercase__ : int = model_dic.get('num_bp3' ) lowercase__ : Union[str, Any] = model_dic.get('rate_weight' ) lowercase__ : Tuple = model_dic.get('rate_thre' ) # create model instance lowercase__ : Tuple = CNN(a , a , a , a , a , a , a ) # modify model parameter lowercase__ : str = model_dic.get('w_conv1' ) lowercase__ : Optional[int] = model_dic.get('wkj' ) lowercase__ : Tuple = model_dic.get('vji' ) lowercase__ : str = model_dic.get('thre_conv1' ) lowercase__ : Union[str, Any] = model_dic.get('thre_bp2' ) lowercase__ : List[str] = model_dic.get('thre_bp3' ) return conv_ins def _UpperCAmelCase ( self , a ) -> str: return 1 / (1 + np.exp(-1 * x )) def _UpperCAmelCase ( self , a ) -> Any: return round(a , 3 ) def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[str]: # convolution process lowercase__ : int = convs[0] lowercase__ : Optional[Any] = convs[1] lowercase__ : int = np.shape(a )[0] # get the data slice of original image data, data_focus lowercase__ : Optional[Any] = [] for i_focus in range(0 , size_data - size_conv + 1 , a ): for j_focus in range(0 , size_data - size_conv + 1 , a ): lowercase__ : Optional[int] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(a ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ : Union[str, Any] = [] lowercase__ : Dict = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(a ): lowercase__ : Any = [] for i_focus in range(len(a ) ): lowercase__ : Tuple = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(a ) ) lowercase__ : Optional[Any] = np.asmatrix(a ).reshape( a , a ) data_featuremap.append(a ) # expanding the data slice to One dimenssion lowercase__ : str = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(a ) ) lowercase__ : int = np.asarray(a ) return focus_list, data_featuremap def _UpperCAmelCase ( self , a , a , a="average_pool" ) -> str: # pooling process lowercase__ : List[str] = len(featuremaps[0] ) lowercase__ : List[str] = int(size_map / size_pooling ) lowercase__ : str = [] for i_map in range(len(a ) ): lowercase__ : List[str] = featuremaps[i_map] lowercase__ : Optional[int] = [] for i_focus in range(0 , a , a ): for j_focus in range(0 , a , a ): lowercase__ : List[Any] = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(a ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(a ) ) lowercase__ : List[Any] = np.asmatrix(a ).reshape(a , a ) featuremap_pooled.append(a ) return featuremap_pooled def _UpperCAmelCase ( self , a ) -> List[str]: # expanding three dimension data to one dimension list lowercase__ : Any = [] for i in range(len(a ) ): lowercase__ : Optional[int] = np.shape(data[i] ) lowercase__ : int = data[i].reshape(1 , shapes[0] * shapes[1] ) lowercase__ : str = data_listed.getA().tolist()[0] data_expanded.extend(a ) lowercase__ : int = np.asarray(a ) return data_expanded def _UpperCAmelCase ( self , a ) -> Dict: # expanding matrix to one dimension list lowercase__ : Dict = np.asarray(a ) lowercase__ : Union[str, Any] = np.shape(a ) lowercase__ : Optional[Any] = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[Any]: lowercase__ : Dict = [] lowercase__ : int = 0 for i_map in range(a ): lowercase__ : str = np.ones((size_map, size_map) ) for i in range(0 , a , a ): for j in range(0 , a , a ): lowercase__ : Optional[Any] = pd_pool[ i_pool ] lowercase__ : Union[str, Any] = i_pool + 1 lowercase__ : List[Any] = np.multiply( a , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(a ) return pd_all def _UpperCAmelCase ( self , a , a , a , a , a , a=bool ) -> str: # model traning print('----------------------Start Training-------------------------' ) print((' - - Shape: Train_Data ', np.shape(a )) ) print((' - - Shape: Teach_Data ', np.shape(a )) ) lowercase__ : int = 0 lowercase__ : List[Any] = [] lowercase__ : Union[str, Any] = 1_0_0_0_0 while rp < n_repeat and mse >= error_accuracy: lowercase__ : List[Any] = 0 print(f"""-------------Learning Time {rp}--------------""" ) for p in range(len(a ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ : Optional[int] = np.asmatrix(datas_train[p] ) lowercase__ : int = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ : Union[str, Any] = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Optional[Any] = self.pooling(a , self.size_poolinga ) lowercase__ : Tuple = np.shape(a ) lowercase__ : List[str] = self._expand(a ) lowercase__ : Optional[int] = data_bp_input lowercase__ : Optional[Any] = np.dot(a , self.vji.T ) - self.thre_bpa lowercase__ : str = self.sig(a ) lowercase__ : Tuple = np.dot(a , self.wkj.T ) - self.thre_bpa lowercase__ : Any = self.sig(a ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ : int = np.multiply( (data_teach - bp_outa) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Any = np.multiply( np.dot(a , self.wkj ) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Optional[int] = np.dot(a , self.vji ) lowercase__ : Union[str, Any] = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ : Any = pd_conva_pooled.T.getA().tolist() lowercase__ : List[str] = self._calculate_gradient_from_pool( a , a , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ : Optional[int] = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ : Tuple = self.rate_weight * np.dot(a , a ) lowercase__ : Union[str, Any] = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ : Any = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ : Tuple = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ : Optional[Any] = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ : Dict = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ : str = rp + 1 lowercase__ : List[str] = error_count / patterns all_mse.append(a ) def draw_error(): lowercase__ : Any = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(a , '+-' ) plt.plot(a , 'r--' ) plt.xlabel('Learning Times' ) plt.ylabel('All_mse' ) plt.grid(a , alpha=0.5 ) plt.show() print('------------------Training Complished---------------------' ) print((' - - Training epoch: ', rp, f""" - - Mse: {mse:.6f}""") ) if draw_e: draw_error() return mse def _UpperCAmelCase ( self , a ) -> List[Any]: # model predict lowercase__ : Optional[int] = [] print('-------------------Start Testing-------------------------' ) print((' - - Shape: Test_Data ', np.shape(a )) ) for p in range(len(a ) ): lowercase__ : List[str] = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ : Tuple = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Any = self.pooling(a , self.size_poolinga ) lowercase__ : Union[str, Any] = self._expand(a ) lowercase__ : Optional[Any] = data_bp_input lowercase__ : str = bp_outa * self.vji.T - self.thre_bpa lowercase__ : Optional[Any] = self.sig(a ) lowercase__ : Dict = bp_outa * self.wkj.T - self.thre_bpa lowercase__ : List[str] = self.sig(a ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ : Optional[int] = [list(map(self.do_round , a ) ) for each in produce_out] return np.asarray(a ) def _UpperCAmelCase ( self , a ) -> List[str]: # return the data of image after convoluting process so we can check it out lowercase__ : Any = np.asmatrix(a ) lowercase__ , lowercase__ : str = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Tuple = self.pooling(a , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
77
1
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class UpperCAmelCase_ ( unittest.TestCase): def __init__( self , a , a=7 , a=3 , a=3_0 , a=4_0_0 , a=True , a=None , a=True , a=[0.5, 0.5, 0.5] , a=[0.5, 0.5, 0.5] , a=True , a=1 / 2_5_5 , a=True , ) -> Any: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowercase__ : Dict = size if size is not None else {'shortest_edge': 1_8, 'longest_edge': 1_3_3_3} lowercase__ : Optional[int] = parent lowercase__ : Tuple = batch_size lowercase__ : List[str] = num_channels lowercase__ : List[str] = min_resolution lowercase__ : Tuple = max_resolution lowercase__ : Union[str, Any] = do_resize lowercase__ : Dict = size lowercase__ : str = do_normalize lowercase__ : List[Any] = image_mean lowercase__ : int = image_std lowercase__ : List[Any] = do_rescale lowercase__ : int = rescale_factor lowercase__ : int = do_pad def _UpperCAmelCase ( self ) -> Optional[Any]: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _UpperCAmelCase ( self , a , a=False ) -> Dict: if not batched: lowercase__ : str = image_inputs[0] if isinstance(a , Image.Image ): lowercase__ , lowercase__ : List[str] = image.size else: lowercase__ , lowercase__ : int = image.shape[1], image.shape[2] if w < h: lowercase__ : Any = int(self.size['shortest_edge'] * h / w ) lowercase__ : Dict = self.size['shortest_edge'] elif w > h: lowercase__ : int = self.size['shortest_edge'] lowercase__ : Tuple = int(self.size['shortest_edge'] * w / h ) else: lowercase__ : Optional[Any] = self.size['shortest_edge'] lowercase__ : List[Any] = self.size['shortest_edge'] else: lowercase__ : Union[str, Any] = [] for image in image_inputs: lowercase__ , lowercase__ : int = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowercase__ : Optional[Any] = max(a , key=lambda a : item[0] )[0] lowercase__ : Optional[Any] = max(a , key=lambda a : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : List[str] = DetaImageProcessor if is_vision_available() else None def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Optional[Any] = DetaImageProcessingTester(self ) @property def _UpperCAmelCase ( self ) -> List[str]: return self.image_processor_tester.prepare_image_processor_dict() def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a , 'image_mean' ) ) self.assertTrue(hasattr(a , 'image_std' ) ) self.assertTrue(hasattr(a , 'do_normalize' ) ) self.assertTrue(hasattr(a , 'do_resize' ) ) self.assertTrue(hasattr(a , 'do_rescale' ) ) self.assertTrue(hasattr(a , 'do_pad' ) ) self.assertTrue(hasattr(a , 'size' ) ) def _UpperCAmelCase ( self ) -> str: lowercase__ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 1_8, 'longest_edge': 1_3_3_3} ) self.assertEqual(image_processor.do_pad , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: pass def _UpperCAmelCase ( self ) -> Tuple: # Initialize image_processing lowercase__ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase__ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=a ) for image in image_inputs: self.assertIsInstance(a , Image.Image ) # Test not batched input lowercase__ : List[str] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowercase__ , lowercase__ : Optional[int] = self.image_processor_tester.get_expected_values(a ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase__ , lowercase__ : Tuple = self.image_processor_tester.get_expected_values(a , batched=a ) lowercase__ : List[str] = image_processing(a , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _UpperCAmelCase ( self ) -> str: # Initialize image_processing lowercase__ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase__ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=a , numpify=a ) for image in image_inputs: self.assertIsInstance(a , np.ndarray ) # Test not batched input lowercase__ : int = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowercase__ , lowercase__ : List[Any] = self.image_processor_tester.get_expected_values(a ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase__ : Union[str, Any] = image_processing(a , return_tensors='pt' ).pixel_values lowercase__ , lowercase__ : int = self.image_processor_tester.get_expected_values(a , batched=a ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _UpperCAmelCase ( self ) -> Tuple: # Initialize image_processing lowercase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase__ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=a , torchify=a ) for image in image_inputs: self.assertIsInstance(a , torch.Tensor ) # Test not batched input lowercase__ : Dict = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values lowercase__ , lowercase__ : Optional[Any] = self.image_processor_tester.get_expected_values(a ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowercase__ : Optional[int] = image_processing(a , return_tensors='pt' ).pixel_values lowercase__ , lowercase__ : Optional[Any] = self.image_processor_tester.get_expected_values(a , batched=a ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def _UpperCAmelCase ( self ) -> Dict: # prepare image and target lowercase__ : Optional[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: lowercase__ : Tuple = json.loads(f.read() ) lowercase__ : Optional[int] = {'image_id': 3_9_7_6_9, 'annotations': target} # encode them lowercase__ : Union[str, Any] = DetaImageProcessor() lowercase__ : List[str] = image_processing(images=a , annotations=a , return_tensors='pt' ) # verify pixel values lowercase__ : int = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['pixel_values'].shape , a ) lowercase__ : List[str] = torch.tensor([0.2_796, 0.3_138, 0.3_481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , a , atol=1e-4 ) ) # verify area lowercase__ : List[str] = torch.tensor([5_887.9_600, 11_250.2_061, 489_353.8_438, 837_122.7_500, 147_967.5_156, 165_732.3_438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , a ) ) # verify boxes lowercase__ : Dict = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , a ) lowercase__ : str = torch.tensor([0.5_503, 0.2_765, 0.0_604, 0.2_215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , a , atol=1e-3 ) ) # verify image_id lowercase__ : Tuple = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , a ) ) # verify is_crowd lowercase__ : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , a ) ) # verify class_labels lowercase__ : List[Any] = torch.tensor([7_5, 7_5, 6_3, 6_5, 1_7, 1_7] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , a ) ) # verify orig_size lowercase__ : Tuple = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , a ) ) # verify size lowercase__ : Optional[int] = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , a ) ) @slow def _UpperCAmelCase ( self ) -> List[str]: # prepare image, target and masks_path lowercase__ : str = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: lowercase__ : int = json.loads(f.read() ) lowercase__ : Optional[Any] = {'file_name': '000000039769.png', 'image_id': 3_9_7_6_9, 'segments_info': target} lowercase__ : Any = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them lowercase__ : List[Any] = DetaImageProcessor(format='coco_panoptic' ) lowercase__ : int = image_processing(images=a , annotations=a , masks_path=a , return_tensors='pt' ) # verify pixel values lowercase__ : List[Any] = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['pixel_values'].shape , a ) lowercase__ : Optional[int] = torch.tensor([0.2_796, 0.3_138, 0.3_481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , a , atol=1e-4 ) ) # verify area lowercase__ : List[str] = torch.tensor([147_979.6_875, 165_527.0_469, 484_638.5_938, 11_292.9_375, 5_879.6_562, 7_634.1_147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , a ) ) # verify boxes lowercase__ : int = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , a ) lowercase__ : List[Any] = torch.tensor([0.2_625, 0.5_437, 0.4_688, 0.8_625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , a , atol=1e-3 ) ) # verify image_id lowercase__ : Union[str, Any] = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , a ) ) # verify is_crowd lowercase__ : Union[str, Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , a ) ) # verify class_labels lowercase__ : Tuple = torch.tensor([1_7, 1_7, 6_3, 7_5, 7_5, 9_3] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , a ) ) # verify masks lowercase__ : Dict = 8_2_2_8_7_3 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , a ) # verify orig_size lowercase__ : Any = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , a ) ) # verify size lowercase__ : List[str] = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , a ) )
77
"""simple docstring""" from collections.abc import Generator def a_ ( ): '''simple docstring''' lowercase__ , lowercase__ : List[str] = 0, 1 while True: lowercase__ , lowercase__ : Optional[int] = b, a + b yield b def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' lowercase__ : List[Any] = 1 lowercase__ : Any = fibonacci_generator() while len(str(next(_lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
77
1
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a_ ( _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : int = args.pruning_method lowercase__ : Tuple = args.threshold lowercase__ : str = args.model_name_or_path.rstrip('/' ) lowercase__ : List[Any] = args.target_model_path print(f"""Load fine-pruned model from {model_name_or_path}""" ) lowercase__ : Optional[Any] = torch.load(os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) lowercase__ : List[str] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowercase__ : Tuple = tensor print(f"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowercase__ : List[str] = tensor print(f"""Copied layer {name}""" ) elif "bias" in name: lowercase__ : Optional[Any] = tensor print(f"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowercase__ : Optional[Any] = MagnitudeBinarizer.apply(inputs=_lowerCAmelCase , threshold=_lowerCAmelCase ) lowercase__ : Optional[int] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowercase__ : Optional[Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ : Any = TopKBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[Any] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowercase__ : Any = name[:-6] lowercase__ : Optional[Any] = model[f"""{prefix_}mask_scores"""] lowercase__ : Tuple = ThresholdBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[str] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowercase__ : Union[str, Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ , lowercase__ : Tuple = -0.1, 1.1 lowercase__ : Optional[Any] = torch.sigmoid(_lowerCAmelCase ) lowercase__ : Optional[Any] = s * (r - l) + l lowercase__ : Optional[Any] = s_bar.clamp(min=0.0 , max=1.0 ) lowercase__ : Union[str, Any] = tensor * mask print(f"""Pruned layer {name}""" ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: lowercase__ : Union[str, Any] = os.path.join( os.path.dirname(_lowerCAmelCase ) , f"""bertarized_{os.path.basename(_lowerCAmelCase )}""" ) if not os.path.isdir(_lowerCAmelCase ): shutil.copytree(_lowerCAmelCase , _lowerCAmelCase ) print(f"""\nCreated folder {target_model_path}""" ) torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": _UpperCamelCase : int = argparse.ArgumentParser() parser.add_argument( "--pruning_method", choices=["l0", "magnitude", "topK", "sigmoied_threshold"], type=str, required=True, help=( "Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning," " sigmoied_threshold = Soft movement pruning)" ), ) parser.add_argument( "--threshold", type=float, required=False, help=( "For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model." "For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared." "Not needed for `l0`" ), ) parser.add_argument( "--model_name_or_path", type=str, required=True, help="Folder containing the model that was previously fine-pruned", ) parser.add_argument( "--target_model_path", default=None, type=str, required=False, help="Folder containing the model that was previously fine-pruned", ) _UpperCamelCase : Dict = parser.parse_args() main(args)
77
"""simple docstring""" import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class UpperCAmelCase_ : def __init__( self , a ) -> List[str]: if isinstance(a , a ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden lowercase__ : Optional[Any] = deepcopy(a ) elif os.path.exists(a ): with io.open(a , 'r' , encoding='utf-8' ) as f: lowercase__ : List[Any] = json.load(a ) else: try: lowercase__ : Optional[int] = baseaa.urlsafe_baadecode(a ).decode('utf-8' ) lowercase__ : List[str] = json.loads(a ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""" ) lowercase__ : Any = config self.set_stage_and_offload() def _UpperCAmelCase ( self ) -> Dict: # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. lowercase__ : Tuple = self.get_value('zero_optimization.stage' , -1 ) # offload lowercase__ : int = False if self.is_zeroa() or self.is_zeroa(): lowercase__ : str = set(['cpu', 'nvme'] ) lowercase__ : Optional[Any] = set( [ self.get_value('zero_optimization.offload_optimizer.device' ), self.get_value('zero_optimization.offload_param.device' ), ] ) if len(offload_devices & offload_devices_valid ) > 0: lowercase__ : Optional[Any] = True def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Dict = self.config # find the config node of interest if it exists lowercase__ : int = ds_key_long.split('.' ) lowercase__ : Dict = nodes.pop() for node in nodes: lowercase__ : Optional[Any] = config.get(a ) if config is None: return None, ds_key return config, ds_key def _UpperCAmelCase ( self , a , a=None ) -> Union[str, Any]: lowercase__ , lowercase__ : Tuple = self.find_config_node(a ) if config is None: return default return config.get(a , a ) def _UpperCAmelCase ( self , a , a=False ) -> Any: lowercase__ : str = self.config # find the config node of interest if it exists lowercase__ : List[Any] = ds_key_long.split('.' ) for node in nodes: lowercase__ : str = config lowercase__ : str = config.get(a ) if config is None: if must_exist: raise ValueError(f"""Can't find {ds_key_long} entry in the config: {self.config}""" ) else: return # if found remove it if parent_config is not None: parent_config.pop(a ) def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : Union[str, Any] = self.get_value(a ) return False if value is None else bool(a ) def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Any = self.get_value(a ) return False if value is None else not bool(a ) def _UpperCAmelCase ( self ) -> Tuple: return self._stage == 2 def _UpperCAmelCase ( self ) -> List[Any]: return self._stage == 3 def _UpperCAmelCase ( self ) -> str: return self._offload class UpperCAmelCase_ : def __init__( self , a ) -> str: lowercase__ : Tuple = engine def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: # runs backpropagation and handles mixed precision self.engine.backward(a , **a ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class UpperCAmelCase_ ( _a): def __init__( self , a ) -> Dict: super().__init__(a , device_placement=a , scaler=a ) lowercase__ : Union[str, Any] = hasattr(self.optimizer , 'overflow' ) def _UpperCAmelCase ( self , a=None ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def _UpperCAmelCase ( self ) -> Optional[int]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def _UpperCAmelCase ( self ) -> Tuple: if self.__has_overflow__: return self.optimizer.overflow return False class UpperCAmelCase_ ( _a): def __init__( self , a , a ) -> Any: super().__init__(a , a ) def _UpperCAmelCase ( self ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class UpperCAmelCase_ : def __init__( self , a , a=0.001 , a=0 , **a ) -> Tuple: lowercase__ : List[Any] = params lowercase__ : int = lr lowercase__ : int = weight_decay lowercase__ : Union[str, Any] = kwargs class UpperCAmelCase_ : def __init__( self , a , a=None , a=0 , **a ) -> Tuple: lowercase__ : Dict = optimizer lowercase__ : List[str] = total_num_steps lowercase__ : Optional[int] = warmup_num_steps lowercase__ : List[Any] = kwargs
77
1
"""simple docstring""" import pytest _UpperCamelCase : Dict = "__dummy_dataset1__" _UpperCamelCase : int = "\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\"\nURLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn-validation.jsonl\"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\n \"ner_tags\": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n \"O\",\n \"B-PER\",\n \"I-PER\",\n \"B-ORG\",\n \"I-ORG\",\n \"B-LOC\",\n \"I-LOC\",\n ]\n )\n ),\n \"langs\": datasets.Sequence(datasets.Value(\"string\")),\n \"spans\": datasets.Sequence(datasets.Value(\"string\")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={\"filepath\": dl_path[\"train\"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={\"filepath\": dl_path[\"validation\"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n" @pytest.fixture def a_ ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def a_ ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] ): '''simple docstring''' lowercase__ : Optional[Any] = dataset_loading_script_name lowercase__ : Optional[int] = tmp_path / 'datasets' / script_name script_dir.mkdir(parents=_lowerCAmelCase ) lowercase__ : Optional[int] = script_dir / f"""{script_name}.py""" with open(_lowerCAmelCase , 'w' ) as f: f.write(_lowerCAmelCase ) return str(_lowerCAmelCase )
77
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _UpperCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , *a , **a ) -> Union[str, Any]: super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _UpperCAmelCase ( self , a=None ) -> Dict: lowercase__ : Any = {} if top_k is not None: lowercase__ : List[str] = top_k return {}, {}, postprocess_params def __call__( self , a , **a ) -> Tuple: return super().__call__(a , **a ) def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : List[Any] = load_image(a ) lowercase__ : Union[str, Any] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _UpperCAmelCase ( self , a ) -> List[str]: lowercase__ : Dict = self.model(**a ) return model_outputs def _UpperCAmelCase ( self , a , a=5 ) -> Dict: if top_k > self.model.config.num_labels: lowercase__ : List[Any] = self.model.config.num_labels if self.framework == "pt": lowercase__ : Tuple = model_outputs.logits.softmax(-1 )[0] lowercase__ , lowercase__ : Optional[Any] = probs.topk(a ) elif self.framework == "tf": lowercase__ : Union[str, Any] = stable_softmax(model_outputs.logits , axis=-1 )[0] lowercase__ : str = tf.math.top_k(a , k=a ) lowercase__ , lowercase__ : Dict = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) lowercase__ : Dict = scores.tolist() lowercase__ : Dict = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
77
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase : Optional[int] = { "configuration_xlm_roberta": [ "XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMRobertaConfig", "XLMRobertaOnnxConfig", ], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = ["XLMRobertaTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Tuple = ["XLMRobertaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Optional[int] = [ "XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMRobertaForCausalLM", "XLMRobertaForMaskedLM", "XLMRobertaForMultipleChoice", "XLMRobertaForQuestionAnswering", "XLMRobertaForSequenceClassification", "XLMRobertaForTokenClassification", "XLMRobertaModel", "XLMRobertaPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Optional[Any] = [ "TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMRobertaForCausalLM", "TFXLMRobertaForMaskedLM", "TFXLMRobertaForMultipleChoice", "TFXLMRobertaForQuestionAnswering", "TFXLMRobertaForSequenceClassification", "TFXLMRobertaForTokenClassification", "TFXLMRobertaModel", "TFXLMRobertaPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : str = [ "FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "FlaxXLMRobertaForMaskedLM", "FlaxXLMRobertaForCausalLM", "FlaxXLMRobertaForMultipleChoice", "FlaxXLMRobertaForQuestionAnswering", "FlaxXLMRobertaForSequenceClassification", "FlaxXLMRobertaForTokenClassification", "FlaxXLMRobertaModel", "FlaxXLMRobertaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaConfig, XLMRobertaOnnxConfig, ) try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta import XLMRobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta_fast import XLMRobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaForCausalLM, XLMRobertaForMaskedLM, XLMRobertaForMultipleChoice, XLMRobertaForQuestionAnswering, XLMRobertaForSequenceClassification, XLMRobertaForTokenClassification, XLMRobertaModel, XLMRobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm_roberta import ( TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMRobertaForCausalLM, TFXLMRobertaForMaskedLM, TFXLMRobertaForMultipleChoice, TFXLMRobertaForQuestionAnswering, TFXLMRobertaForSequenceClassification, TFXLMRobertaForTokenClassification, TFXLMRobertaModel, TFXLMRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xlm_roberta import ( FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxXLMRobertaForCausalLM, FlaxXLMRobertaForMaskedLM, FlaxXLMRobertaForMultipleChoice, FlaxXLMRobertaForQuestionAnswering, FlaxXLMRobertaForSequenceClassification, FlaxXLMRobertaForTokenClassification, FlaxXLMRobertaModel, FlaxXLMRobertaPreTrainedModel, ) else: import sys _UpperCamelCase : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
77
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : str = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all LED models at https://huggingface.co/models?filter=LED _UpperCamelCase : Optional[Any] = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } _UpperCamelCase : Optional[int] = { "allenai/led-base-16384": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def a_ ( ): '''simple docstring''' lowercase__ : int = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowercase__ : Union[str, Any] = bs[:] lowercase__ : str = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCAmelCase ) cs.append(2**8 + n ) n += 1 lowercase__ : str = [chr(_lowerCAmelCase ) for n in cs] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Dict = set() lowercase__ : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Optional[Any] = char return pairs class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = VOCAB_FILES_NAMES lowerCamelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , a , a , a="replace" , a="<s>" , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a=False , **a , ) -> Any: lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else bos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else eos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else sep_token lowercase__ : Dict = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else cls_token lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else unk_token lowercase__ : Tuple = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( errors=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , add_prefix_space=a , **a , ) with open(a , encoding='utf-8' ) as vocab_handle: lowercase__ : Tuple = json.load(a ) lowercase__ : Dict = {v: k for k, v in self.encoder.items()} lowercase__ : str = errors # how to handle errors in decoding lowercase__ : Optional[Any] = bytes_to_unicode() lowercase__ : Optional[Any] = {v: k for k, v in self.byte_encoder.items()} with open(a , encoding='utf-8' ) as merges_handle: lowercase__ : Optional[Any] = merges_handle.read().split('\n' )[1:-1] lowercase__ : Optional[int] = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ : Union[str, Any] = dict(zip(a , range(len(a ) ) ) ) lowercase__ : Tuple = {} lowercase__ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ : List[Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _UpperCAmelCase ( self ) -> List[Any]: return len(self.encoder ) def _UpperCAmelCase ( self ) -> str: return dict(self.encoder , **self.added_tokens_encoder ) def _UpperCAmelCase ( self , a ) -> List[str]: if token in self.cache: return self.cache[token] lowercase__ : Optional[Any] = tuple(a ) lowercase__ : int = get_pairs(a ) if not pairs: return token while True: lowercase__ : List[str] = min(a , key=lambda a : self.bpe_ranks.get(a , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ : List[str] = bigram lowercase__ : Union[str, Any] = [] lowercase__ : List[Any] = 0 while i < len(a ): try: lowercase__ : str = word.index(a , a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : Optional[int] = j if word[i] == first and i < len(a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : int = tuple(a ) lowercase__ : Dict = new_word if len(a ) == 1: break else: lowercase__ : Any = get_pairs(a ) lowercase__ : List[str] = ' '.join(a ) lowercase__ : Optional[Any] = word return word def _UpperCAmelCase ( self , a ) -> Union[str, Any]: lowercase__ : Tuple = [] for token in re.findall(self.pat , a ): lowercase__ : Union[str, Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(a ).split(' ' ) ) return bpe_tokens def _UpperCAmelCase ( self , a ) -> Optional[Any]: return self.encoder.get(a , self.encoder.get(self.unk_token ) ) def _UpperCAmelCase ( self , a ) -> Optional[int]: return self.decoder.get(a ) def _UpperCAmelCase ( self , a ) -> str: lowercase__ : Any = ''.join(a ) lowercase__ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Any = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : str = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(a , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a , ensure_ascii=a ) + '\n' ) lowercase__ : List[Any] = 0 with open(a , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) lowercase__ : Union[str, Any] = token_index writer.write(' '.join(a ) + '\n' ) index += 1 return vocab_file, merge_file def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ : Union[str, Any] = [self.cls_token_id] lowercase__ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) + [1] def _UpperCAmelCase ( self , a , a = None ) -> List[int]: lowercase__ : Dict = [self.sep_token_id] lowercase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCAmelCase ( self , a , a=False , **a ) -> Optional[int]: lowercase__ : Tuple = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(a ) > 0 and not text[0].isspace()): lowercase__ : List[str] = ' ' + text return (text, kwargs) def _UpperCAmelCase ( self , a , a = None , a = PaddingStrategy.DO_NOT_PAD , a = None , a = None , ) -> dict: lowercase__ : Dict = super()._pad( encoded_inputs=a , max_length=a , padding_strategy=a , pad_to_multiple_of=a , return_attention_mask=a , ) # Load from model defaults if return_attention_mask is None: lowercase__ : Union[str, Any] = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ : Any = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ : Tuple = len(encoded_inputs['global_attention_mask'] ) != len(a ) if needs_to_be_padded: lowercase__ : str = len(a ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ : Union[str, Any] = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowercase__ : List[str] = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
77
1
"""simple docstring""" from __future__ import annotations from random import choice def a_ ( _lowerCAmelCase : Tuple ): '''simple docstring''' return choice(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : list[int] , _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Optional[Any] = random_pivot(_lowerCAmelCase ) # partition based on pivot # linear time lowercase__ : str = [e for e in lst if e < pivot] lowercase__ : List[Any] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(_lowerCAmelCase ) == k - 1: return pivot # pivot is in elements bigger than k elif len(_lowerCAmelCase ) < k - 1: return kth_number(_lowerCAmelCase , k - len(_lowerCAmelCase ) - 1 ) # pivot is in elements smaller than k else: return kth_number(_lowerCAmelCase , _lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" _UpperCamelCase : Union[str, Any] = 8.3_1_4_4_5_9_8 def a_ ( _lowerCAmelCase : float , _lowerCAmelCase : float ): '''simple docstring''' if temperature < 0: raise Exception('Temperature cannot be less than 0 K' ) if molar_mass <= 0: raise Exception('Molar mass cannot be less than or equal to 0 kg/mol' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _UpperCamelCase : List[Any] = 3_00 _UpperCamelCase : Tuple = 28 _UpperCamelCase : Any = rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
77
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class UpperCAmelCase_ ( metaclass=_a): lowerCamelCase__ : Dict = ["flax", "transformers"] def __init__( self , *a , **a ) -> int: requires_backends(self , ['flax', 'transformers'] ) @classmethod def _UpperCAmelCase ( cls , *a , **a ) -> Optional[int]: requires_backends(cls , ['flax', 'transformers'] ) @classmethod def _UpperCAmelCase ( cls , *a , **a ) -> Optional[int]: requires_backends(cls , ['flax', 'transformers'] ) class UpperCAmelCase_ ( metaclass=_a): lowerCamelCase__ : str = ["flax", "transformers"] def __init__( self , *a , **a ) -> Tuple: requires_backends(self , ['flax', 'transformers'] ) @classmethod def _UpperCAmelCase ( cls , *a , **a ) -> str: requires_backends(cls , ['flax', 'transformers'] ) @classmethod def _UpperCAmelCase ( cls , *a , **a ) -> Optional[Any]: requires_backends(cls , ['flax', 'transformers'] ) class UpperCAmelCase_ ( metaclass=_a): lowerCamelCase__ : Union[str, Any] = ["flax", "transformers"] def __init__( self , *a , **a ) -> str: requires_backends(self , ['flax', 'transformers'] ) @classmethod def _UpperCAmelCase ( cls , *a , **a ) -> int: requires_backends(cls , ['flax', 'transformers'] ) @classmethod def _UpperCAmelCase ( cls , *a , **a ) -> List[Any]: requires_backends(cls , ['flax', 'transformers'] ) class UpperCAmelCase_ ( metaclass=_a): lowerCamelCase__ : Union[str, Any] = ["flax", "transformers"] def __init__( self , *a , **a ) -> Union[str, Any]: requires_backends(self , ['flax', 'transformers'] ) @classmethod def _UpperCAmelCase ( cls , *a , **a ) -> List[Any]: requires_backends(cls , ['flax', 'transformers'] ) @classmethod def _UpperCAmelCase ( cls , *a , **a ) -> Optional[int]: requires_backends(cls , ['flax', 'transformers'] )
77
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization lowerCamelCase__ : str = field(default="text-classification" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({"labels": ClassLabel}) lowerCamelCase__ : str = "text" lowerCamelCase__ : str = "labels" def _UpperCAmelCase ( self , a ) -> Tuple: if self.label_column not in features: raise ValueError(f"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , a ): raise ValueError(f"""Column {self.label_column} is not a ClassLabel.""" ) lowercase__ : Optional[Any] = copy.deepcopy(self ) lowercase__ : Optional[Any] = self.label_schema.copy() lowercase__ : Any = features[self.label_column] lowercase__ : Optional[Any] = label_schema return task_template @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return { self.text_column: "text", self.label_column: "labels", }
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int ): '''simple docstring''' return x if y == 0 else greatest_common_divisor(_lowerCAmelCase , x % y ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int ): '''simple docstring''' return (x * y) // greatest_common_divisor(_lowerCAmelCase , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : int = 20 ): '''simple docstring''' lowercase__ : Dict = 1 for i in range(1 , n + 1 ): lowercase__ : List[Any] = lcm(_lowerCAmelCase , _lowerCAmelCase ) return g if __name__ == "__main__": print(f'''{solution() = }''')
77
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np _UpperCamelCase : Any = re.compile(r"\b(a|an|the)\b", re.UNICODE) _UpperCamelCase : Union[str, Any] = None def a_ ( ): '''simple docstring''' lowercase__ : Optional[int] = argparse.ArgumentParser('Official evaluation script for SQuAD version 2.0.' ) parser.add_argument('data_file' , metavar='data.json' , help='Input data JSON file.' ) parser.add_argument('pred_file' , metavar='pred.json' , help='Model predictions.' ) parser.add_argument( '--out-file' , '-o' , metavar='eval.json' , help='Write accuracy metrics to file (default is stdout).' ) parser.add_argument( '--na-prob-file' , '-n' , metavar='na_prob.json' , help='Model estimates of probability of no answer.' ) parser.add_argument( '--na-prob-thresh' , '-t' , type=_lowerCAmelCase , default=1.0 , help='Predict "" if no-answer probability exceeds this (default = 1.0).' , ) parser.add_argument( '--out-image-dir' , '-p' , metavar='out_images' , default=_lowerCAmelCase , help='Save precision-recall curves to directory.' ) parser.add_argument('--verbose' , '-v' , action='store_true' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Tuple = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Optional[int] = bool(qa['answers']['text'] ) return qid_to_has_ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' def remove_articles(_lowerCAmelCase : int ): return ARTICLES_REGEX.sub(' ' , _lowerCAmelCase ) def white_space_fix(_lowerCAmelCase : str ): return " ".join(text.split() ) def remove_punc(_lowerCAmelCase : List[Any] ): lowercase__ : int = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_lowerCAmelCase : List[str] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_lowerCAmelCase ) ) ) ) def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' if not s: return [] return normalize_answer(_lowerCAmelCase ).split() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ): '''simple docstring''' return int(normalize_answer(_lowerCAmelCase ) == normalize_answer(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ): '''simple docstring''' lowercase__ : Dict = get_tokens(_lowerCAmelCase ) lowercase__ : List[str] = get_tokens(_lowerCAmelCase ) lowercase__ : List[Any] = collections.Counter(_lowerCAmelCase ) & collections.Counter(_lowerCAmelCase ) lowercase__ : int = sum(common.values() ) if len(_lowerCAmelCase ) == 0 or len(_lowerCAmelCase ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 lowercase__ : Any = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Dict = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Any = (2 * precision * recall) / (precision + recall) return fa def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Optional[int] = {} lowercase__ : Union[str, Any] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Any = qa['id'] lowercase__ : Union[str, Any] = [t for t in qa['answers']['text'] if normalize_answer(_lowerCAmelCase )] if not gold_answers: # For unanswerable questions, only correct answer is empty string lowercase__ : Dict = [''] if qid not in preds: print(f"""Missing prediction for {qid}""" ) continue lowercase__ : Optional[int] = preds[qid] # Take max over all gold answers lowercase__ : int = max(compute_exact(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) lowercase__ : Optional[Any] = max(compute_fa(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) return exact_scores, fa_scores def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : str = {} for qid, s in scores.items(): lowercase__ : int = na_probs[qid] > na_prob_thresh if pred_na: lowercase__ : Optional[Any] = float(not qid_to_has_ans[qid] ) else: lowercase__ : Optional[Any] = s return new_scores def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str=None ): '''simple docstring''' if not qid_list: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores.values() ) / total), ('f1', 1_0_0.0 * sum(fa_scores.values() ) / total), ('total', total), ] ) else: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores[k] for k in qid_list ) / total), ('f1', 1_0_0.0 * sum(fa_scores[k] for k in qid_list ) / total), ('total', total), ] ) def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' for k in new_eval: lowercase__ : int = new_eval[k] def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple ): '''simple docstring''' plt.step(_lowerCAmelCase , _lowerCAmelCase , color='b' , alpha=0.2 , where='post' ) plt.fill_between(_lowerCAmelCase , _lowerCAmelCase , step='post' , alpha=0.2 , color='b' ) plt.xlabel('Recall' ) plt.ylabel('Precision' ) plt.xlim([0.0, 1.0_5] ) plt.ylim([0.0, 1.0_5] ) plt.title(_lowerCAmelCase ) plt.savefig(_lowerCAmelCase ) plt.clf() def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : str , _lowerCAmelCase : Any=None , _lowerCAmelCase : List[str]=None ): '''simple docstring''' lowercase__ : Optional[int] = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) lowercase__ : Tuple = 0.0 lowercase__ : List[str] = 1.0 lowercase__ : List[str] = 0.0 lowercase__ : Union[str, Any] = [1.0] lowercase__ : List[Any] = [0.0] lowercase__ : Optional[int] = 0.0 for i, qid in enumerate(_lowerCAmelCase ): if qid_to_has_ans[qid]: true_pos += scores[qid] lowercase__ : Tuple = true_pos / float(i + 1 ) lowercase__ : Union[str, Any] = true_pos / float(_lowerCAmelCase ) if i == len(_lowerCAmelCase ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_lowerCAmelCase ) recalls.append(_lowerCAmelCase ) if out_image: plot_pr_curve(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return {"ap": 1_0_0.0 * avg_prec} def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple ): '''simple docstring''' if out_image_dir and not os.path.exists(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) lowercase__ : List[str] = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return lowercase__ : Dict = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_exact.png' ) , title='Precision-Recall curve for Exact Match score' , ) lowercase__ : Tuple = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_f1.png' ) , title='Precision-Recall curve for F1 score' , ) lowercase__ : List[Any] = {k: float(_lowerCAmelCase ) for k, v in qid_to_has_ans.items()} lowercase__ : Any = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_oracle.png' ) , title='Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)' , ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_exact' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_f1' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_oracle' ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' if not qid_list: return lowercase__ : List[str] = [na_probs[k] for k in qid_list] lowercase__ : Tuple = np.ones_like(_lowerCAmelCase ) / float(len(_lowerCAmelCase ) ) plt.hist(_lowerCAmelCase , weights=_lowerCAmelCase , bins=20 , range=(0.0, 1.0) ) plt.xlabel('Model probability of no-answer' ) plt.ylabel('Proportion of dataset' ) plt.title(f"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(_lowerCAmelCase , f"""na_prob_hist_{name}.png""" ) ) plt.clf() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : Tuple = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) lowercase__ : int = num_no_ans lowercase__ : Optional[int] = cur_score lowercase__ : Tuple = 0.0 lowercase__ : Dict = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) for i, qid in enumerate(_lowerCAmelCase ): if qid not in scores: continue if qid_to_has_ans[qid]: lowercase__ : Optional[int] = scores[qid] else: if preds[qid]: lowercase__ : List[Any] = -1 else: lowercase__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: lowercase__ : Dict = cur_score lowercase__ : Optional[int] = na_probs[qid] return 1_0_0.0 * best_score / len(_lowerCAmelCase ), best_thresh def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str ): '''simple docstring''' lowercase__ , lowercase__ : List[Any] = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ , lowercase__ : Dict = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Any = best_exact lowercase__ : Tuple = exact_thresh lowercase__ : Optional[Any] = best_fa lowercase__ : Any = fa_thresh def a_ ( ): '''simple docstring''' with open(OPTS.data_file ) as f: lowercase__ : List[Any] = json.load(_lowerCAmelCase ) lowercase__ : Union[str, Any] = dataset_json['data'] with open(OPTS.pred_file ) as f: lowercase__ : str = json.load(_lowerCAmelCase ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: lowercase__ : Union[str, Any] = json.load(_lowerCAmelCase ) else: lowercase__ : str = {k: 0.0 for k in preds} lowercase__ : int = make_qid_to_has_ans(_lowerCAmelCase ) # maps qid to True/False lowercase__ : List[str] = [k for k, v in qid_to_has_ans.items() if v] lowercase__ : Any = [k for k, v in qid_to_has_ans.items() if not v] lowercase__ , lowercase__ : Any = get_raw_scores(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Optional[Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Union[str, Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Tuple = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase ) if has_ans_qids: lowercase__ : int = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'HasAns' ) if no_ans_qids: lowercase__ : Optional[Any] = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'NoAns' ) if OPTS.na_prob_file: find_all_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'hasAns' ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'noAns' ) if OPTS.out_file: with open(OPTS.out_file , 'w' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) else: print(json.dumps(_lowerCAmelCase , indent=2 ) ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
77
1
"""simple docstring""" _UpperCamelCase : int = { "meter": "m", "kilometer": "km", "megametre": "Mm", "gigametre": "Gm", "terametre": "Tm", "petametre": "Pm", "exametre": "Em", "zettametre": "Zm", "yottametre": "Ym", } # Exponent of the factor(meter) _UpperCamelCase : int = { "m": 0, "km": 3, "Mm": 6, "Gm": 9, "Tm": 12, "Pm": 15, "Em": 18, "Zm": 21, "Ym": 24, } def a_ ( _lowerCAmelCase : float , _lowerCAmelCase : str , _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : Optional[Any] = from_type.lower().strip('s' ) lowercase__ : Optional[Any] = to_type.lower().strip('s' ) lowercase__ : Union[str, Any] = UNIT_SYMBOL.get(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Tuple = UNIT_SYMBOL.get(_lowerCAmelCase , _lowerCAmelCase ) if from_sanitized not in METRIC_CONVERSION: lowercase__ : Tuple = ( f"""Invalid 'from_type' value: {from_type!r}.\n""" f"""Conversion abbreviations are: {", ".join(_lowerCAmelCase )}""" ) raise ValueError(_lowerCAmelCase ) if to_sanitized not in METRIC_CONVERSION: lowercase__ : Dict = ( f"""Invalid 'to_type' value: {to_type!r}.\n""" f"""Conversion abbreviations are: {", ".join(_lowerCAmelCase )}""" ) raise ValueError(_lowerCAmelCase ) lowercase__ : Dict = METRIC_CONVERSION[from_sanitized] lowercase__ : Optional[Any] = METRIC_CONVERSION[to_sanitized] lowercase__ : Union[str, Any] = 1 if from_exponent > to_exponent: lowercase__ : List[str] = from_exponent - to_exponent else: lowercase__ : Union[str, Any] = -(to_exponent - from_exponent) return value * pow(10 , _lowerCAmelCase ) if __name__ == "__main__": from doctest import testmod testmod()
77
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = 0 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Tuple = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Any: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : str = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : int = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = CLIPConfig() # Create a dummy config file with image_proceesor_type lowercase__ : Optional[int] = Path(a ) / 'preprocessor_config.json' lowercase__ : Optional[int] = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowercase__ : int = AutoImageProcessor.from_pretrained(a ).to_dict() config_dict.pop('image_processor_type' ) lowercase__ : Tuple = CLIPImageProcessor(**a ) # save in new folder model_config.save_pretrained(a ) config.save_pretrained(a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) # make sure private variable is not incorrectly saved lowercase__ : Optional[int] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = Path(a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'clip-base is not a local folder and is not a valid model identifier' ): lowercase__ : Any = AutoImageProcessor.from_pretrained('clip-base' ) def _UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex( a , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowercase__ : Dict = AutoImageProcessor.from_pretrained(a , revision='aaaaaa' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def _UpperCAmelCase ( self ) -> Optional[int]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(a ): lowercase__ : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : str = AutoImageProcessor.from_pretrained(a , trust_remote_code=a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def _UpperCAmelCase ( self ) -> int: try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoImageProcessor.register(a , a ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Optional[Any] = Path(a ) / 'preprocessor_config.json' lowercase__ : List[Any] = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = CustomImageProcessor.from_pretrained(a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCAmelCase ( self ) -> Dict: class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = True try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # If remote code is not set, the default is to use local lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowercase__ : int = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
77
1
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : Optional[Any] = 3 lowercase__ : Union[str, Any] = 2_5_0 lowercase__ : Union[str, Any] = ids_tensor((batch_size, length) , a ) lowercase__ : List[Any] = torch.ones((batch_size, length) , device=a , dtype=torch.float ) / length return input_ids, scores def _UpperCAmelCase ( self ) -> str: lowercase__ , lowercase__ : Optional[Any] = self._get_tensors(5 ) lowercase__ : int = StoppingCriteriaList( [ MaxLengthCriteria(max_length=1_0 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(a , a ) ) lowercase__ , lowercase__ : Optional[Any] = self._get_tensors(9 ) self.assertFalse(criteria(a , a ) ) lowercase__ , lowercase__ : Any = self._get_tensors(1_0 ) self.assertTrue(criteria(a , a ) ) def _UpperCAmelCase ( self ) -> int: lowercase__ : Optional[int] = MaxLengthCriteria(max_length=1_0 ) lowercase__ , lowercase__ : Dict = self._get_tensors(5 ) self.assertFalse(criteria(a , a ) ) lowercase__ , lowercase__ : Tuple = self._get_tensors(9 ) self.assertFalse(criteria(a , a ) ) lowercase__ , lowercase__ : str = self._get_tensors(1_0 ) self.assertTrue(criteria(a , a ) ) def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : Any = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) lowercase__ , lowercase__ : Optional[int] = self._get_tensors(5 ) self.assertFalse(criteria(a , a ) ) lowercase__ , lowercase__ : Dict = self._get_tensors(9 ) self.assertFalse(criteria(a , a ) ) lowercase__ , lowercase__ : Tuple = self._get_tensors(1_0 ) self.assertTrue(criteria(a , a ) ) lowercase__ : Optional[int] = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 1_0 ) def _UpperCAmelCase ( self ) -> int: lowercase__ , lowercase__ : List[Any] = self._get_tensors(5 ) lowercase__ : Dict = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(a , a ) ) lowercase__ : Dict = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(a , a ) ) def _UpperCAmelCase ( self ) -> Union[str, Any]: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(1_0 )] ) , 1_0 ) with self.assertWarns(a ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(1_0 )] ) , 1_1 ) lowercase__ : Tuple = validate_stopping_criteria(StoppingCriteriaList() , 1_1 ) self.assertEqual(len(a ) , 1 )
77
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class UpperCAmelCase_ ( _a): lowerCamelCase__ : Dict = ["image_processor", "tokenizer"] lowerCamelCase__ : Dict = "BlipImageProcessor" lowerCamelCase__ : Union[str, Any] = "AutoTokenizer" def __init__( self , a , a , a ) -> Optional[int]: super().__init__(a , a ) # add QFormer tokenizer lowercase__ : Dict = qformer_tokenizer def __call__( self , a = None , a = None , a = True , a = False , a = None , a = None , a = 0 , a = None , a = None , a = False , a = False , a = False , a = False , a = False , a = True , a = None , **a , ) -> BatchFeature: if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) lowercase__ : List[Any] = BatchFeature() if text is not None: lowercase__ : Optional[int] = self.tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) encoding.update(a ) lowercase__ : Optional[int] = self.qformer_tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) lowercase__ : List[str] = qformer_text_encoding.pop('input_ids' ) lowercase__ : Any = qformer_text_encoding.pop('attention_mask' ) if images is not None: lowercase__ : List[Any] = self.image_processor(a , return_tensors=a ) encoding.update(a ) return encoding def _UpperCAmelCase ( self , *a , **a ) -> List[str]: return self.tokenizer.batch_decode(*a , **a ) def _UpperCAmelCase ( self , *a , **a ) -> Tuple: return self.tokenizer.decode(*a , **a ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : str = self.tokenizer.model_input_names lowercase__ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: if os.path.isfile(a ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(a , exist_ok=a ) lowercase__ : int = os.path.join(a , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(a ) return super().save_pretrained(a , **a ) @classmethod def _UpperCAmelCase ( cls , a , **a ) -> str: lowercase__ : str = AutoTokenizer.from_pretrained(a , subfolder='qformer_tokenizer' ) lowercase__ : int = cls._get_arguments_from_pretrained(a , **a ) args.append(a ) return cls(*a )
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int | float | str ): '''simple docstring''' try: lowercase__ : Optional[Any] = float(_lowerCAmelCase ) except ValueError: raise ValueError('Please enter a valid number' ) lowercase__ : Tuple = decimal - int(_lowerCAmelCase ) if fractional_part == 0: return int(_lowerCAmelCase ), 1 else: lowercase__ : List[Any] = len(str(_lowerCAmelCase ).split('.' )[1] ) lowercase__ : int = int(decimal * (10**number_of_frac_digits) ) lowercase__ : str = 10**number_of_frac_digits lowercase__ , lowercase__ : Optional[int] = denominator, numerator while True: lowercase__ : Optional[int] = dividend % divisor if remainder == 0: break lowercase__ , lowercase__ : Tuple = divisor, remainder lowercase__ , lowercase__ : List[str] = numerator / divisor, denominator / divisor return int(_lowerCAmelCase ), int(_lowerCAmelCase ) if __name__ == "__main__": print(f'''{decimal_to_fraction(2) = }''') print(f'''{decimal_to_fraction(8_9.0) = }''') print(f'''{decimal_to_fraction("67") = }''') print(f'''{decimal_to_fraction("45.0") = }''') print(f'''{decimal_to_fraction(1.5) = }''') print(f'''{decimal_to_fraction("6.25") = }''') print(f'''{decimal_to_fraction("78td") = }''')
77
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({}) lowerCamelCase__ : str = "text" @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return {self.text_column: "text"}
77
1
"""simple docstring""" _UpperCamelCase : Union[str, Any] = 8.3_1_4_4_5_9_8 def a_ ( _lowerCAmelCase : float , _lowerCAmelCase : float ): '''simple docstring''' if temperature < 0: raise Exception('Temperature cannot be less than 0 K' ) if molar_mass <= 0: raise Exception('Molar mass cannot be less than or equal to 0 kg/mol' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _UpperCamelCase : List[Any] = 3_00 _UpperCamelCase : Tuple = 28 _UpperCamelCase : Any = rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
77
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Optional[int] = [] lowercase__ : Optional[Any] = set({'(', '[', '{'} ) lowercase__ : List[str] = set({')', ']', '}'} ) lowercase__ : Optional[int] = {'{': '}', '[': ']', '(': ')'} for i in range(len(_lowerCAmelCase ) ): if s[i] in open_brackets: stack.append(s[i] ) elif s[i] in closed_brackets and ( len(_lowerCAmelCase ) == 0 or (len(_lowerCAmelCase ) > 0 and open_to_closed[stack.pop()] != s[i]) ): return False return len(_lowerCAmelCase ) == 0 def a_ ( ): '''simple docstring''' lowercase__ : List[Any] = input('Enter sequence of brackets: ' ) if is_balanced(_lowerCAmelCase ): print(_lowerCAmelCase , 'is balanced' ) else: print(_lowerCAmelCase , 'is not balanced' ) if __name__ == "__main__": main()
77
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING _UpperCamelCase : Dict = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , **a ) -> Dict: super().__init__(**a ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type(a ) def __call__( self , a , a = None , **a , ) -> List[str]: if "text_queries" in kwargs: lowercase__ : Optional[Any] = kwargs.pop('text_queries' ) if isinstance(a , (str, Image.Image) ): lowercase__ : Optional[Any] = {'image': image, 'candidate_labels': candidate_labels} else: lowercase__ : List[str] = image lowercase__ : Optional[Any] = super().__call__(a , **a ) return results def _UpperCAmelCase ( self , **a ) -> Dict: lowercase__ : Optional[Any] = {} if "threshold" in kwargs: lowercase__ : Tuple = kwargs['threshold'] if "top_k" in kwargs: lowercase__ : List[Any] = kwargs['top_k'] return {}, {}, postprocess_params def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : Any = load_image(inputs['image'] ) lowercase__ : Optional[int] = inputs['candidate_labels'] if isinstance(a , a ): lowercase__ : Optional[int] = candidate_labels.split(',' ) lowercase__ : Optional[int] = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(a ): lowercase__ : List[str] = self.tokenizer(a , return_tensors=self.framework ) lowercase__ : List[Any] = self.image_processor(a , return_tensors=self.framework ) yield { "is_last": i == len(a ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : List[Any] = model_inputs.pop('target_size' ) lowercase__ : Dict = model_inputs.pop('candidate_label' ) lowercase__ : Dict = model_inputs.pop('is_last' ) lowercase__ : Optional[int] = self.model(**a ) lowercase__ : Any = {'target_size': target_size, 'candidate_label': candidate_label, 'is_last': is_last, **outputs} return model_outputs def _UpperCAmelCase ( self , a , a=0.1 , a=None ) -> Union[str, Any]: lowercase__ : Dict = [] for model_output in model_outputs: lowercase__ : List[Any] = model_output['candidate_label'] lowercase__ : Optional[int] = BaseModelOutput(a ) lowercase__ : Any = self.image_processor.post_process_object_detection( outputs=a , threshold=a , target_sizes=model_output['target_size'] )[0] for index in outputs["scores"].nonzero(): lowercase__ : Union[str, Any] = outputs['scores'][index].item() lowercase__ : Tuple = self._get_bounding_box(outputs['boxes'][index][0] ) lowercase__ : Tuple = {'score': score, 'label': label, 'box': box} results.append(a ) lowercase__ : Dict = sorted(a , key=lambda a : x["score"] , reverse=a ) if top_k: lowercase__ : Dict = results[:top_k] return results def _UpperCAmelCase ( self , a ) -> Dict[str, int]: if self.framework != "pt": raise ValueError('The ZeroShotObjectDetectionPipeline is only available in PyTorch.' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ : Tuple = box.int().tolist() lowercase__ : Any = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
77
1
"""simple docstring""" from __future__ import annotations from collections import namedtuple def a_ ( _lowerCAmelCase : float , _lowerCAmelCase : float , _lowerCAmelCase : float ): '''simple docstring''' lowercase__ : Union[str, Any] = namedtuple('result' , 'name value' ) if (voltage, current, power).count(0 ) != 1: raise ValueError('Only one argument must be 0' ) elif power < 0: raise ValueError( 'Power cannot be negative in any electrical/electronics system' ) elif voltage == 0: return result('voltage' , power / current ) elif current == 0: return result('current' , power / voltage ) elif power == 0: return result('power' , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError('Exactly one argument must be 0' ) if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ : def __init__( self , a , a=1_2 , a=7 , a=True , a=True , a=True , a=9_9 , a=3_2 , a=3_2 , a=2 , a=4 , a=3_7 , a=0.1 , a=0.1 , a=5_1_2 , a=0.02 , a=0 , a=None , ) -> Union[str, Any]: lowercase__ : Any = parent lowercase__ : str = batch_size lowercase__ : List[Any] = seq_length lowercase__ : Union[str, Any] = is_training lowercase__ : List[str] = use_input_mask lowercase__ : int = use_labels lowercase__ : List[Any] = vocab_size lowercase__ : str = hidden_size lowercase__ : int = projection_dim lowercase__ : Optional[int] = num_hidden_layers lowercase__ : Any = num_attention_heads lowercase__ : Optional[Any] = intermediate_size lowercase__ : Optional[Any] = dropout lowercase__ : Optional[int] = attention_dropout lowercase__ : Optional[int] = max_position_embeddings lowercase__ : str = initializer_range lowercase__ : Tuple = scope lowercase__ : int = bos_token_id def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : int = None if self.use_input_mask: lowercase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowercase__ : int = input_mask.numpy() lowercase__ , lowercase__ : Tuple = input_mask.shape lowercase__ : List[str] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a ): lowercase__ : Dict = 1 lowercase__ : Union[str, Any] = 0 lowercase__ : Tuple = self.get_config() return config, input_ids, tf.convert_to_tensor(a ) def _UpperCAmelCase ( self ) -> List[Any]: return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _UpperCAmelCase ( self , a , a , a ) -> Any: lowercase__ : List[Any] = TFBlipTextModel(config=a ) lowercase__ : Optional[int] = model(a , attention_mask=a , training=a ) lowercase__ : List[str] = model(a , training=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _UpperCAmelCase ( self ) -> Any: lowercase__ : Optional[Any] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ : Any = config_and_inputs lowercase__ : Any = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : Dict = (TFBlipTextModel,) if is_tf_available() else () lowerCamelCase__ : Optional[Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Optional[int] = BlipTextModelTester(self ) lowercase__ : int = ConfigTester(self , config_class=a , hidden_size=3_7 ) def _UpperCAmelCase ( self ) -> Tuple: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> int: lowercase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _UpperCAmelCase ( self ) -> Optional[Any]: pass def _UpperCAmelCase ( self ) -> Optional[int]: pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def _UpperCAmelCase ( self ) -> List[str]: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> Dict: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> str: pass @slow def _UpperCAmelCase ( self ) -> int: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Any = TFBlipTextModel.from_pretrained(a ) self.assertIsNotNone(a ) def _UpperCAmelCase ( self , a=True ) -> List[str]: super().test_pt_tf_model_equivalence(allow_missing_keys=a )
77
1
"""simple docstring""" from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class UpperCAmelCase_ ( _a): def __init__( self , a , a = None , a = None , a = None , a = False , a = False , a = None , **a , ) -> Union[str, Any]: super().__init__( a , split=a , features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowercase__ : List[Any] = path_or_paths if isinstance(a , a ) else {self.split: path_or_paths} lowercase__ : Union[str, Any] = Text( cache_dir=a , data_files=a , features=a , **a , ) def _UpperCAmelCase ( self ) -> Dict: # Build iterable dataset if self.streaming: lowercase__ : List[Any] = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: lowercase__ : Optional[Any] = None lowercase__ : Union[str, Any] = None lowercase__ : List[str] = None lowercase__ : List[Any] = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowercase__ : Dict = self.builder.as_dataset( split=self.split , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
77
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase : Tuple = logging.get_logger(__name__) _UpperCamelCase : Union[str, Any] = { "alibaba-damo/mgp-str-base": "https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json", } class UpperCAmelCase_ ( _a): lowerCamelCase__ : List[Any] = "mgp-str" def __init__( self , a=[3_2, 1_2_8] , a=4 , a=3 , a=2_7 , a=3_8 , a=5_0_2_5_7 , a=3_0_5_2_2 , a=7_6_8 , a=1_2 , a=1_2 , a=4.0 , a=True , a=False , a=1e-5 , a=0.0 , a=0.0 , a=0.0 , a=False , a=0.02 , **a , ) -> Tuple: super().__init__(**a ) lowercase__ : int = image_size lowercase__ : List[Any] = patch_size lowercase__ : Optional[Any] = num_channels lowercase__ : Optional[Any] = max_token_length lowercase__ : Dict = num_character_labels lowercase__ : Optional[int] = num_bpe_labels lowercase__ : Dict = num_wordpiece_labels lowercase__ : Tuple = hidden_size lowercase__ : List[Any] = num_hidden_layers lowercase__ : Union[str, Any] = num_attention_heads lowercase__ : Dict = mlp_ratio lowercase__ : Optional[int] = distilled lowercase__ : Optional[int] = layer_norm_eps lowercase__ : Optional[int] = drop_rate lowercase__ : List[str] = qkv_bias lowercase__ : Optional[int] = attn_drop_rate lowercase__ : Any = drop_path_rate lowercase__ : List[Any] = output_aa_attentions lowercase__ : Tuple = initializer_range
77
1
"""simple docstring""" from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def a_ ( ): '''simple docstring''' lowercase__ : Dict = [randint(-1000 , 1000 ) for i in range(10 )] lowercase__ : List[Any] = randint(-5000 , 5000 ) return (arr, r) _UpperCamelCase : List[Any] = make_dataset() def a_ ( _lowerCAmelCase : list[int] , _lowerCAmelCase : int ): '''simple docstring''' for triplet in permutations(_lowerCAmelCase , 3 ): if sum(_lowerCAmelCase ) == target: return tuple(sorted(_lowerCAmelCase ) ) return (0, 0, 0) def a_ ( _lowerCAmelCase : list[int] , _lowerCAmelCase : int ): '''simple docstring''' arr.sort() lowercase__ : Optional[int] = len(_lowerCAmelCase ) for i in range(n - 1 ): lowercase__ , lowercase__ : Tuple = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def a_ ( ): '''simple docstring''' lowercase__ : Any = '\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n' lowercase__ : Union[str, Any] = '\ntriplet_sum1(*dataset)\n' lowercase__ : Any = '\ntriplet_sum2(*dataset)\n' lowercase__ : List[str] = repeat(setup=_lowerCAmelCase , stmt=_lowerCAmelCase , repeat=5 , number=1_0000 ) lowercase__ : Dict = repeat(setup=_lowerCAmelCase , stmt=_lowerCAmelCase , repeat=5 , number=1_0000 ) return (min(_lowerCAmelCase ), min(_lowerCAmelCase )) if __name__ == "__main__": from doctest import testmod testmod() _UpperCamelCase : List[Any] = solution_times() print(f'''The time for naive implementation is {times[0]}.''') print(f'''The time for optimized implementation is {times[1]}.''')
77
"""simple docstring""" import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : int = FunnelConfig.from_json_file(_lowerCAmelCase ) print(f"""Building PyTorch model from configuration: {config}""" ) lowercase__ : List[Any] = FunnelBaseModel(_lowerCAmelCase ) if base_model else FunnelModel(_lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_funnel(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _lowerCAmelCase ) if __name__ == "__main__": _UpperCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--base_model", action="store_true", help="Whether you want just the base model (no decoder) or not." ) _UpperCamelCase : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : Dict = { "microsoft/focalnet-tiny": "https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json", } class UpperCAmelCase_ ( _a , _a): lowerCamelCase__ : str = "focalnet" def __init__( self , a=2_2_4 , a=4 , a=3 , a=9_6 , a=False , a=[1_9_2, 3_8_4, 7_6_8, 7_6_8] , a=[2, 2, 6, 2] , a=[2, 2, 2, 2] , a=[3, 3, 3, 3] , a="gelu" , a=4.0 , a=0.0 , a=0.1 , a=False , a=1e-4 , a=False , a=False , a=False , a=0.02 , a=1e-5 , a=3_2 , a=None , a=None , **a , ) -> Any: super().__init__(**a ) lowercase__ : List[str] = image_size lowercase__ : List[str] = patch_size lowercase__ : Optional[int] = num_channels lowercase__ : Any = embed_dim lowercase__ : str = use_conv_embed lowercase__ : str = hidden_sizes lowercase__ : Optional[int] = depths lowercase__ : Dict = focal_levels lowercase__ : List[Any] = focal_windows lowercase__ : List[Any] = hidden_act lowercase__ : List[Any] = mlp_ratio lowercase__ : int = hidden_dropout_prob lowercase__ : str = drop_path_rate lowercase__ : str = use_layerscale lowercase__ : Optional[int] = layerscale_value lowercase__ : Any = use_post_layernorm lowercase__ : Optional[Any] = use_post_layernorm_in_modulation lowercase__ : List[str] = normalize_modulator lowercase__ : int = initializer_range lowercase__ : List[str] = layer_norm_eps lowercase__ : int = encoder_stride lowercase__ : Dict = ['stem'] + [f"""stage{idx}""" for idx in range(1 , len(self.depths ) + 1 )] lowercase__ , lowercase__ : Tuple = get_aligned_output_features_output_indices( out_features=a , out_indices=a , stage_names=self.stage_names )
77
"""simple docstring""" import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _UpperCamelCase : Any = logging.getLogger(__name__) _UpperCamelCase : Optional[Any] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _UpperCamelCase : Any = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The model checkpoint for weights initialization. Leave None if you want to train a model from" " scratch." ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(_a)} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained config name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "The input training data file (a text file)."}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The input training data files (multiple files in glob format). " "Very often splitting large files to smaller files can prevent tokenizer going out of memory" ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input train ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input eval ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Whether distinct lines of text in the dataset are to be handled as distinct sequences."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Train with masked-language modeling loss instead of language modeling."}) lowerCamelCase__ : bool = field(default=_a , metadata={"help": "Whether ot not to use whole word mask."}) lowerCamelCase__ : float = field( default=0.15 , metadata={"help": "Ratio of tokens to mask for masked language modeling loss"}) lowerCamelCase__ : float = field( default=1 / 6 , metadata={ "help": ( "Ratio of length of a span of masked tokens to surrounding context length for permutation language" " modeling." ) } , ) lowerCamelCase__ : int = field( default=5 , metadata={"help": "Maximum length of a span of masked tokens for permutation language modeling."}) lowerCamelCase__ : int = field( default=-1 , metadata={ "help": ( "Optional input sequence length after tokenization." "The training dataset will be truncated in block of this size for training." "Default to the model max input length for single sentence inputs (take into account special tokens)." ) } , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Overwrite the cached training and evaluation sets"}) def a_ ( _lowerCAmelCase : DataTrainingArguments , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[str] = None , ): '''simple docstring''' def _dataset(_lowerCAmelCase : Any , _lowerCAmelCase : Any=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('You need to set world whole masking and mlm to True for Chinese Whole Word Mask' ) return LineByLineWithRefDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , ref_path=_lowerCAmelCase , ) return LineByLineTextDataset(tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size ) else: return TextDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=_lowerCAmelCase , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(_lowerCAmelCase ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def a_ ( ): '''simple docstring''' lowercase__ : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ : List[Any] = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( 'Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ' 'or remove the --do_eval argument.' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s' , _lowerCAmelCase ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: lowercase__ : List[Any] = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : Any = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: lowercase__ : Union[str, Any] = CONFIG_MAPPING[model_args.model_type]() logger.warning('You are instantiating a new config instance from scratch.' ) if model_args.tokenizer_name: lowercase__ : Dict = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : int = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another' ' script, save it,and load it from here, using --tokenizer_name' ) if model_args.model_name_or_path: lowercase__ : Optional[Any] = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=_lowerCAmelCase , cache_dir=model_args.cache_dir , ) else: logger.info('Training new model from scratch' ) lowercase__ : int = AutoModelWithLMHead.from_config(_lowerCAmelCase ) model.resize_token_embeddings(len(_lowerCAmelCase ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( 'BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the' '--mlm flag (masked language modeling).' ) if data_args.block_size <= 0: lowercase__ : List[str] = tokenizer.max_len # Our input block size will be the max possible for the model else: lowercase__ : int = min(data_args.block_size , tokenizer.max_len ) # Get datasets lowercase__ : Tuple = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) lowercase__ : Optional[Any] = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , evaluate=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": lowercase__ : List[Any] = DataCollatorForPermutationLanguageModeling( tokenizer=_lowerCAmelCase , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: lowercase__ : List[str] = DataCollatorForWholeWordMask( tokenizer=_lowerCAmelCase , mlm_probability=data_args.mlm_probability ) else: lowercase__ : str = DataCollatorForLanguageModeling( tokenizer=_lowerCAmelCase , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer lowercase__ : Optional[int] = Trainer( model=_lowerCAmelCase , args=_lowerCAmelCase , data_collator=_lowerCAmelCase , train_dataset=_lowerCAmelCase , eval_dataset=_lowerCAmelCase , prediction_loss_only=_lowerCAmelCase , ) # Training if training_args.do_train: lowercase__ : List[str] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=_lowerCAmelCase ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase__ : List[Any] = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) lowercase__ : Dict = trainer.evaluate() lowercase__ : List[Any] = math.exp(eval_output['eval_loss'] ) lowercase__ : int = {'perplexity': perplexity} lowercase__ : int = os.path.join(training_args.output_dir , 'eval_results_lm.txt' ) if trainer.is_world_master(): with open(_lowerCAmelCase , 'w' ) as writer: logger.info('***** Eval results *****' ) for key in sorted(result.keys() ): logger.info(' %s = %s' , _lowerCAmelCase , str(result[key] ) ) writer.write('%s = %s\n' % (key, str(result[key] )) ) results.update(_lowerCAmelCase ) return results def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' main() if __name__ == "__main__": main()
77
1
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin _UpperCamelCase : Optional[int] = "\nHugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenagers.[2] After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning.\n\nIn March 2021, Hugging Face raised $40 million in a Series B funding round.[3]\n\nOn April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model.[4] In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.[5]\n" class UpperCAmelCase_ ( unittest.TestCase , _a): def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : List[str] = load_tool('text-question-answering' ) self.tool.setup() lowercase__ : int = load_tool('text-question-answering' , remote=a ) def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : List[str] = self.tool(a , 'What did Hugging Face do in April 2021?' ) self.assertEqual(a , 'launched the BigScience Research Workshop' ) def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : Tuple = self.remote_tool(a , 'What did Hugging Face do in April 2021?' ) self.assertEqual(a , 'launched the BigScience Research Workshop' ) def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Any = self.tool(text=a , question='What did Hugging Face do in April 2021?' ) self.assertEqual(a , 'launched the BigScience Research Workshop' ) def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Optional[int] = self.remote_tool(text=a , question='What did Hugging Face do in April 2021?' ) self.assertEqual(a , 'launched the BigScience Research Workshop' )
77
"""simple docstring""" import math import flax.linen as nn import jax.numpy as jnp def a_ ( _lowerCAmelCase : jnp.ndarray , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1.0E4 , _lowerCAmelCase : bool = False , _lowerCAmelCase : float = 1.0 , ): '''simple docstring''' assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, f"""Embedding dimension {embedding_dim} should be even""" lowercase__ : Optional[Any] = float(embedding_dim // 2 ) lowercase__ : Optional[Any] = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) lowercase__ : Any = min_timescale * jnp.exp(jnp.arange(_lowerCAmelCase , dtype=jnp.floataa ) * -log_timescale_increment ) lowercase__ : Dict = jnp.expand_dims(_lowerCAmelCase , 1 ) * jnp.expand_dims(_lowerCAmelCase , 0 ) # scale embeddings lowercase__ : List[str] = scale * emb if flip_sin_to_cos: lowercase__ : Dict = jnp.concatenate([jnp.cos(_lowerCAmelCase ), jnp.sin(_lowerCAmelCase )] , axis=1 ) else: lowercase__ : Optional[int] = jnp.concatenate([jnp.sin(_lowerCAmelCase ), jnp.cos(_lowerCAmelCase )] , axis=1 ) lowercase__ : List[Any] = jnp.reshape(_lowerCAmelCase , [jnp.shape(_lowerCAmelCase )[0], embedding_dim] ) return signal class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : jnp.dtype = jnp.floataa @nn.compact def __call__( self , a ) -> Any: lowercase__ : str = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_1' )(a ) lowercase__ : Union[str, Any] = nn.silu(a ) lowercase__ : Optional[Any] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_2' )(a ) return temb class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : bool = False lowerCamelCase__ : float = 1 @nn.compact def __call__( self , a ) -> str: return get_sinusoidal_embeddings( a , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
77
1
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = 0 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Tuple = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Any: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : str = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : int = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = CLIPConfig() # Create a dummy config file with image_proceesor_type lowercase__ : Optional[int] = Path(a ) / 'preprocessor_config.json' lowercase__ : Optional[int] = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowercase__ : int = AutoImageProcessor.from_pretrained(a ).to_dict() config_dict.pop('image_processor_type' ) lowercase__ : Tuple = CLIPImageProcessor(**a ) # save in new folder model_config.save_pretrained(a ) config.save_pretrained(a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) # make sure private variable is not incorrectly saved lowercase__ : Optional[int] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = Path(a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'clip-base is not a local folder and is not a valid model identifier' ): lowercase__ : Any = AutoImageProcessor.from_pretrained('clip-base' ) def _UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex( a , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowercase__ : Dict = AutoImageProcessor.from_pretrained(a , revision='aaaaaa' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def _UpperCAmelCase ( self ) -> Optional[int]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(a ): lowercase__ : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : str = AutoImageProcessor.from_pretrained(a , trust_remote_code=a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def _UpperCAmelCase ( self ) -> int: try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoImageProcessor.register(a , a ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Optional[Any] = Path(a ) / 'preprocessor_config.json' lowercase__ : List[Any] = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = CustomImageProcessor.from_pretrained(a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCAmelCase ( self ) -> Dict: class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = True try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # If remote code is not set, the default is to use local lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowercase__ : int = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
77
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def a_ ( _lowerCAmelCase : Callable[[int | float], int | float] , _lowerCAmelCase : int | float , _lowerCAmelCase : int | float , _lowerCAmelCase : int = 100 , ): '''simple docstring''' lowercase__ : Dict = x_start lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) lowercase__ : Optional[Any] = 0.0 for _ in range(_lowerCAmelCase ): # Approximates curve as a sequence of linear lines and sums their length lowercase__ : Union[str, Any] = (x_end - x_start) / steps + xa lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step lowercase__ : Union[str, Any] = xa lowercase__ : int = fxa return length if __name__ == "__main__": def a_ ( _lowerCAmelCase : List[Any] ): '''simple docstring''' return math.sin(10 * x ) print("f(x) = sin(10 * x)") print("The length of the curve from x = -10 to x = 10 is:") _UpperCamelCase : str = 10 while i <= 10_00_00: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
77
1
"""simple docstring""" import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class UpperCAmelCase_ ( _a , _a , unittest.TestCase): lowerCamelCase__ : Optional[Any] = VQModel lowerCamelCase__ : Union[str, Any] = "sample" @property def _UpperCAmelCase ( self , a=(3_2, 3_2) ) -> int: lowercase__ : List[str] = 4 lowercase__ : Optional[Any] = 3 lowercase__ : Optional[Any] = floats_tensor((batch_size, num_channels) + sizes ).to(a ) return {"sample": image} @property def _UpperCAmelCase ( self ) -> Union[str, Any]: return (3, 3_2, 3_2) @property def _UpperCAmelCase ( self ) -> Union[str, Any]: return (3, 3_2, 3_2) def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : Optional[Any] = { 'block_out_channels': [3_2, 6_4], 'in_channels': 3, 'out_channels': 3, 'down_block_types': ['DownEncoderBlock2D', 'DownEncoderBlock2D'], 'up_block_types': ['UpDecoderBlock2D', 'UpDecoderBlock2D'], 'latent_channels': 3, } lowercase__ : Tuple = self.dummy_input return init_dict, inputs_dict def _UpperCAmelCase ( self ) -> Optional[int]: pass def _UpperCAmelCase ( self ) -> List[Any]: pass def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ , lowercase__ : Dict = VQModel.from_pretrained('fusing/vqgan-dummy' , output_loading_info=a ) self.assertIsNotNone(a ) self.assertEqual(len(loading_info['missing_keys'] ) , 0 ) model.to(a ) lowercase__ : Optional[int] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def _UpperCAmelCase ( self ) -> int: lowercase__ : str = VQModel.from_pretrained('fusing/vqgan-dummy' ) model.to(a ).eval() torch.manual_seed(0 ) if torch.cuda.is_available(): torch.cuda.manual_seed_all(0 ) lowercase__ : Optional[int] = torch.randn(1 , model.config.in_channels , model.config.sample_size , model.config.sample_size ) lowercase__ : Union[str, Any] = image.to(a ) with torch.no_grad(): lowercase__ : Optional[int] = model(a ).sample lowercase__ : Optional[Any] = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off lowercase__ : Optional[int] = torch.tensor([-0.0_153, -0.4_044, -0.1_880, -0.5_161, -0.2_418, -0.4_072, -0.1_612, -0.0_633, -0.0_143] ) # fmt: on self.assertTrue(torch.allclose(a , a , atol=1e-3 ) )
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase : Tuple = { "configuration_whisper": ["WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP", "WhisperConfig", "WhisperOnnxConfig"], "feature_extraction_whisper": ["WhisperFeatureExtractor"], "processing_whisper": ["WhisperProcessor"], "tokenization_whisper": ["WhisperTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = ["WhisperTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : str = [ "WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "WhisperForConditionalGeneration", "WhisperModel", "WhisperPreTrainedModel", "WhisperForAudioClassification", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Any = [ "TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFWhisperForConditionalGeneration", "TFWhisperModel", "TFWhisperPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Tuple = [ "FlaxWhisperForConditionalGeneration", "FlaxWhisperModel", "FlaxWhisperPreTrainedModel", "FlaxWhisperForAudioClassification", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _UpperCamelCase : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase : Any = logging.get_logger(__name__) _UpperCamelCase : str = { "microsoft/biogpt": "https://huggingface.co/microsoft/biogpt/resolve/main/config.json", # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class UpperCAmelCase_ ( _a): lowerCamelCase__ : Dict = "biogpt" def __init__( self , a=4_2_3_8_4 , a=1_0_2_4 , a=2_4 , a=1_6 , a=4_0_9_6 , a="gelu" , a=0.1 , a=0.1 , a=1_0_2_4 , a=0.02 , a=1e-12 , a=True , a=True , a=0.0 , a=0.0 , a=1 , a=0 , a=2 , **a , ) -> int: lowercase__ : Tuple = vocab_size lowercase__ : Optional[int] = max_position_embeddings lowercase__ : Dict = hidden_size lowercase__ : Optional[Any] = num_hidden_layers lowercase__ : str = num_attention_heads lowercase__ : Tuple = intermediate_size lowercase__ : Optional[int] = hidden_act lowercase__ : str = hidden_dropout_prob lowercase__ : Optional[Any] = attention_probs_dropout_prob lowercase__ : Optional[int] = initializer_range lowercase__ : List[str] = layer_norm_eps lowercase__ : Dict = scale_embedding lowercase__ : Tuple = use_cache lowercase__ : List[str] = layerdrop lowercase__ : Union[str, Any] = activation_dropout super().__init__(pad_token_id=a , bos_token_id=a , eos_token_id=a , **a )
77
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a_ ( _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : int = args.pruning_method lowercase__ : Tuple = args.threshold lowercase__ : str = args.model_name_or_path.rstrip('/' ) lowercase__ : List[Any] = args.target_model_path print(f"""Load fine-pruned model from {model_name_or_path}""" ) lowercase__ : Optional[Any] = torch.load(os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) lowercase__ : List[str] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowercase__ : Tuple = tensor print(f"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: lowercase__ : List[str] = tensor print(f"""Copied layer {name}""" ) elif "bias" in name: lowercase__ : Optional[Any] = tensor print(f"""Copied layer {name}""" ) else: if pruning_method == "magnitude": lowercase__ : Optional[Any] = MagnitudeBinarizer.apply(inputs=_lowerCAmelCase , threshold=_lowerCAmelCase ) lowercase__ : Optional[int] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue lowercase__ : Optional[Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ : Any = TopKBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[Any] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowercase__ : Any = name[:-6] lowercase__ : Optional[Any] = model[f"""{prefix_}mask_scores"""] lowercase__ : Tuple = ThresholdBinarizer.apply(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[str] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue lowercase__ : Union[str, Any] = name[:-6] lowercase__ : Optional[int] = model[f"""{prefix_}mask_scores"""] lowercase__ , lowercase__ : Tuple = -0.1, 1.1 lowercase__ : Optional[Any] = torch.sigmoid(_lowerCAmelCase ) lowercase__ : Optional[Any] = s * (r - l) + l lowercase__ : Optional[Any] = s_bar.clamp(min=0.0 , max=1.0 ) lowercase__ : Union[str, Any] = tensor * mask print(f"""Pruned layer {name}""" ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: lowercase__ : Union[str, Any] = os.path.join( os.path.dirname(_lowerCAmelCase ) , f"""bertarized_{os.path.basename(_lowerCAmelCase )}""" ) if not os.path.isdir(_lowerCAmelCase ): shutil.copytree(_lowerCAmelCase , _lowerCAmelCase ) print(f"""\nCreated folder {target_model_path}""" ) torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": _UpperCamelCase : int = argparse.ArgumentParser() parser.add_argument( "--pruning_method", choices=["l0", "magnitude", "topK", "sigmoied_threshold"], type=str, required=True, help=( "Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning," " sigmoied_threshold = Soft movement pruning)" ), ) parser.add_argument( "--threshold", type=float, required=False, help=( "For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model." "For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared." "Not needed for `l0`" ), ) parser.add_argument( "--model_name_or_path", type=str, required=True, help="Folder containing the model that was previously fine-pruned", ) parser.add_argument( "--target_model_path", default=None, type=str, required=False, help="Folder containing the model that was previously fine-pruned", ) _UpperCamelCase : Dict = parser.parse_args() main(args)
77
1
"""simple docstring""" import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex _UpperCamelCase : Any = logging.getLogger(__name__) class UpperCAmelCase_ : def __init__( self ) -> Dict: lowercase__ : Union[str, Any] = False def _UpperCAmelCase ( self , a , a , a , a ) -> Any: if not self.initialized: lowercase__ : Dict = RagRetriever( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) lowercase__ : List[str] = True def _UpperCAmelCase ( self ) -> List[Any]: self.retriever.index.init_index() def _UpperCAmelCase ( self , a , a ) -> Tuple: lowercase__ , lowercase__ : Optional[Any] = self.retriever._main_retrieve(a , a ) return doc_ids, retrieved_doc_embeds class UpperCAmelCase_ ( _a): def __init__( self , a , a , a , a , a=None ) -> Optional[int]: if index is not None and index.is_initialized() and len(a ) > 0: raise ValueError( 'When using Ray for distributed fine-tuning, ' 'you\'ll need to provide the paths instead, ' 'as the dataset and the index are loaded ' 'separately. More info in examples/rag/use_own_knowledge_dataset.py ' ) super().__init__( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) lowercase__ : Dict = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(a , a , a , a ) for worker in self.retrieval_workers ] ) def _UpperCAmelCase ( self ) -> int: logger.info('initializing retrieval' ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def _UpperCAmelCase ( self , a , a ) -> List[Any]: if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. lowercase__ : Union[str, Any] = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers ) - 1 )] lowercase__ , lowercase__ : Optional[Any] = ray.get(random_worker.retrieve.remote(a , a ) ) else: lowercase__ , lowercase__ : Optional[Any] = self._main_retrieve(a , a ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(a ) @classmethod def _UpperCAmelCase ( cls , a , a=None , **a ) -> Optional[int]: return super(a , cls ).get_tokenizers(a , a , **a ) @classmethod def _UpperCAmelCase ( cls , a , a , a=None , **a ) -> Dict: lowercase__ : List[Any] = kwargs.pop('config' , a ) or RagConfig.from_pretrained(a , **a ) lowercase__ : str = RagTokenizer.from_pretrained(a , config=a ) lowercase__ : List[str] = rag_tokenizer.question_encoder lowercase__ : Optional[int] = rag_tokenizer.generator if indexed_dataset is not None: lowercase__ : List[str] = 'custom' lowercase__ : Dict = CustomHFIndex(config.retrieval_vector_size , a ) else: lowercase__ : Dict = cls._build_index(a ) return cls( a , question_encoder_tokenizer=a , generator_tokenizer=a , retrieval_workers=a , index=a , )
77
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[List[PIL.Image.Image], np.ndarray] lowerCamelCase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : np.ndarray lowerCamelCase__ : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
77
1
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
77
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class UpperCAmelCase_ : def __init__( self , a , a , a , a , a , a=0.2 , a=0.2 ) -> Dict: lowercase__ : Any = bp_numa lowercase__ : Optional[int] = bp_numa lowercase__ : Tuple = bp_numa lowercase__ : Optional[Any] = conva_get[:2] lowercase__ : Optional[int] = conva_get[2] lowercase__ : Optional[Any] = size_pa lowercase__ : Union[str, Any] = rate_w lowercase__ : Union[str, Any] = rate_t lowercase__ : List[Any] = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ : Optional[Any] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : int = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : Any = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 def _UpperCAmelCase ( self , a ) -> Union[str, Any]: # save model dict with pickle lowercase__ : Optional[Any] = { 'num_bp1': self.num_bpa, 'num_bp2': self.num_bpa, 'num_bp3': self.num_bpa, 'conv1': self.conva, 'step_conv1': self.step_conva, 'size_pooling1': self.size_poolinga, 'rate_weight': self.rate_weight, 'rate_thre': self.rate_thre, 'w_conv1': self.w_conva, 'wkj': self.wkj, 'vji': self.vji, 'thre_conv1': self.thre_conva, 'thre_bp2': self.thre_bpa, 'thre_bp3': self.thre_bpa, } with open(a , 'wb' ) as f: pickle.dump(a , a ) print(f"""Model saved: {save_path}""" ) @classmethod def _UpperCAmelCase ( cls , a ) -> Any: # read saved model with open(a , 'rb' ) as f: lowercase__ : Optional[int] = pickle.load(a ) # noqa: S301 lowercase__ : Optional[int] = model_dic.get('conv1' ) conv_get.append(model_dic.get('step_conv1' ) ) lowercase__ : List[Any] = model_dic.get('size_pooling1' ) lowercase__ : Tuple = model_dic.get('num_bp1' ) lowercase__ : int = model_dic.get('num_bp2' ) lowercase__ : int = model_dic.get('num_bp3' ) lowercase__ : Union[str, Any] = model_dic.get('rate_weight' ) lowercase__ : Tuple = model_dic.get('rate_thre' ) # create model instance lowercase__ : Tuple = CNN(a , a , a , a , a , a , a ) # modify model parameter lowercase__ : str = model_dic.get('w_conv1' ) lowercase__ : Optional[int] = model_dic.get('wkj' ) lowercase__ : Tuple = model_dic.get('vji' ) lowercase__ : str = model_dic.get('thre_conv1' ) lowercase__ : Union[str, Any] = model_dic.get('thre_bp2' ) lowercase__ : List[str] = model_dic.get('thre_bp3' ) return conv_ins def _UpperCAmelCase ( self , a ) -> str: return 1 / (1 + np.exp(-1 * x )) def _UpperCAmelCase ( self , a ) -> Any: return round(a , 3 ) def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[str]: # convolution process lowercase__ : int = convs[0] lowercase__ : Optional[Any] = convs[1] lowercase__ : int = np.shape(a )[0] # get the data slice of original image data, data_focus lowercase__ : Optional[Any] = [] for i_focus in range(0 , size_data - size_conv + 1 , a ): for j_focus in range(0 , size_data - size_conv + 1 , a ): lowercase__ : Optional[int] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(a ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ : Union[str, Any] = [] lowercase__ : Dict = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(a ): lowercase__ : Any = [] for i_focus in range(len(a ) ): lowercase__ : Tuple = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(a ) ) lowercase__ : Optional[Any] = np.asmatrix(a ).reshape( a , a ) data_featuremap.append(a ) # expanding the data slice to One dimenssion lowercase__ : str = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(a ) ) lowercase__ : int = np.asarray(a ) return focus_list, data_featuremap def _UpperCAmelCase ( self , a , a , a="average_pool" ) -> str: # pooling process lowercase__ : List[str] = len(featuremaps[0] ) lowercase__ : List[str] = int(size_map / size_pooling ) lowercase__ : str = [] for i_map in range(len(a ) ): lowercase__ : List[str] = featuremaps[i_map] lowercase__ : Optional[int] = [] for i_focus in range(0 , a , a ): for j_focus in range(0 , a , a ): lowercase__ : List[Any] = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(a ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(a ) ) lowercase__ : List[Any] = np.asmatrix(a ).reshape(a , a ) featuremap_pooled.append(a ) return featuremap_pooled def _UpperCAmelCase ( self , a ) -> List[str]: # expanding three dimension data to one dimension list lowercase__ : Any = [] for i in range(len(a ) ): lowercase__ : Optional[int] = np.shape(data[i] ) lowercase__ : int = data[i].reshape(1 , shapes[0] * shapes[1] ) lowercase__ : str = data_listed.getA().tolist()[0] data_expanded.extend(a ) lowercase__ : int = np.asarray(a ) return data_expanded def _UpperCAmelCase ( self , a ) -> Dict: # expanding matrix to one dimension list lowercase__ : Dict = np.asarray(a ) lowercase__ : Union[str, Any] = np.shape(a ) lowercase__ : Optional[Any] = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[Any]: lowercase__ : Dict = [] lowercase__ : int = 0 for i_map in range(a ): lowercase__ : str = np.ones((size_map, size_map) ) for i in range(0 , a , a ): for j in range(0 , a , a ): lowercase__ : Optional[Any] = pd_pool[ i_pool ] lowercase__ : Union[str, Any] = i_pool + 1 lowercase__ : List[Any] = np.multiply( a , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(a ) return pd_all def _UpperCAmelCase ( self , a , a , a , a , a , a=bool ) -> str: # model traning print('----------------------Start Training-------------------------' ) print((' - - Shape: Train_Data ', np.shape(a )) ) print((' - - Shape: Teach_Data ', np.shape(a )) ) lowercase__ : int = 0 lowercase__ : List[Any] = [] lowercase__ : Union[str, Any] = 1_0_0_0_0 while rp < n_repeat and mse >= error_accuracy: lowercase__ : List[Any] = 0 print(f"""-------------Learning Time {rp}--------------""" ) for p in range(len(a ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ : Optional[int] = np.asmatrix(datas_train[p] ) lowercase__ : int = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ : Union[str, Any] = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Optional[Any] = self.pooling(a , self.size_poolinga ) lowercase__ : Tuple = np.shape(a ) lowercase__ : List[str] = self._expand(a ) lowercase__ : Optional[int] = data_bp_input lowercase__ : Optional[Any] = np.dot(a , self.vji.T ) - self.thre_bpa lowercase__ : str = self.sig(a ) lowercase__ : Tuple = np.dot(a , self.wkj.T ) - self.thre_bpa lowercase__ : Any = self.sig(a ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ : int = np.multiply( (data_teach - bp_outa) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Any = np.multiply( np.dot(a , self.wkj ) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Optional[int] = np.dot(a , self.vji ) lowercase__ : Union[str, Any] = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ : Any = pd_conva_pooled.T.getA().tolist() lowercase__ : List[str] = self._calculate_gradient_from_pool( a , a , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ : Optional[int] = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ : Tuple = self.rate_weight * np.dot(a , a ) lowercase__ : Union[str, Any] = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ : Any = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ : Tuple = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ : Optional[Any] = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ : Dict = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ : str = rp + 1 lowercase__ : List[str] = error_count / patterns all_mse.append(a ) def draw_error(): lowercase__ : Any = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(a , '+-' ) plt.plot(a , 'r--' ) plt.xlabel('Learning Times' ) plt.ylabel('All_mse' ) plt.grid(a , alpha=0.5 ) plt.show() print('------------------Training Complished---------------------' ) print((' - - Training epoch: ', rp, f""" - - Mse: {mse:.6f}""") ) if draw_e: draw_error() return mse def _UpperCAmelCase ( self , a ) -> List[Any]: # model predict lowercase__ : Optional[int] = [] print('-------------------Start Testing-------------------------' ) print((' - - Shape: Test_Data ', np.shape(a )) ) for p in range(len(a ) ): lowercase__ : List[str] = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ : Tuple = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Any = self.pooling(a , self.size_poolinga ) lowercase__ : Union[str, Any] = self._expand(a ) lowercase__ : Optional[Any] = data_bp_input lowercase__ : str = bp_outa * self.vji.T - self.thre_bpa lowercase__ : Optional[Any] = self.sig(a ) lowercase__ : Dict = bp_outa * self.wkj.T - self.thre_bpa lowercase__ : List[str] = self.sig(a ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ : Optional[int] = [list(map(self.do_round , a ) ) for each in produce_out] return np.asarray(a ) def _UpperCAmelCase ( self , a ) -> List[str]: # return the data of image after convoluting process so we can check it out lowercase__ : Any = np.asmatrix(a ) lowercase__ , lowercase__ : str = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Tuple = self.pooling(a , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
77
1
"""simple docstring""" import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Path , _lowerCAmelCase : str = None , _lowerCAmelCase : str = None , _lowerCAmelCase : str = None , ): '''simple docstring''' if config_name_or_path is None: lowercase__ : str = 'facebook/rag-token-base' if model_type == 'rag_token' else 'facebook/rag-sequence-base' if generator_tokenizer_name_or_path is None: lowercase__ : Optional[Any] = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: lowercase__ : Union[str, Any] = question_encoder_name_or_path lowercase__ : Union[str, Any] = RagTokenForGeneration if model_type == 'rag_token' else RagSequenceForGeneration # Save model. lowercase__ : Optional[Any] = RagConfig.from_pretrained(_lowerCAmelCase ) lowercase__ : List[str] = AutoConfig.from_pretrained(_lowerCAmelCase ) lowercase__ : str = AutoConfig.from_pretrained(_lowerCAmelCase ) lowercase__ : str = gen_config lowercase__ : List[Any] = question_encoder_config lowercase__ : List[str] = model_class.from_pretrained_question_encoder_generator( _lowerCAmelCase , _lowerCAmelCase , config=_lowerCAmelCase ) rag_model.save_pretrained(_lowerCAmelCase ) # Sanity check. model_class.from_pretrained(_lowerCAmelCase ) # Save tokenizers. lowercase__ : Optional[int] = AutoTokenizer.from_pretrained(_lowerCAmelCase ) gen_tokenizer.save_pretrained(dest_dir / 'generator_tokenizer/' ) lowercase__ : Dict = AutoTokenizer.from_pretrained(_lowerCAmelCase ) question_encoder_tokenizer.save_pretrained(dest_dir / 'question_encoder_tokenizer/' ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = argparse.ArgumentParser() parser.add_argument( "--model_type", choices=["rag_sequence", "rag_token"], required=True, type=str, help="RAG model type: rag_sequence, rag_token", ) parser.add_argument("--dest", type=str, required=True, help="Path to the output checkpoint directory.") parser.add_argument("--generator_name_or_path", type=str, required=True, help="Generator model identifier") parser.add_argument( "--question_encoder_name_or_path", type=str, required=True, help="Question encoder model identifier" ) parser.add_argument( "--generator_tokenizer_name_or_path", type=str, help="Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``", ) parser.add_argument( "--question_encoder_tokenizer_name_or_path", type=str, help="Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``", ) parser.add_argument( "--config_name_or_path", type=str, help=( "Identifier of the model config to use, if not provided, resolves to a base config for a given" " ``model_type``" ), ) _UpperCamelCase : Optional[Any] = parser.parse_args() _UpperCamelCase : Optional[Any] = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
77
"""simple docstring""" from collections.abc import Generator def a_ ( ): '''simple docstring''' lowercase__ , lowercase__ : List[str] = 0, 1 while True: lowercase__ , lowercase__ : Optional[int] = b, a + b yield b def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' lowercase__ : List[Any] = 1 lowercase__ : Any = fibonacci_generator() while len(str(next(_lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
77
1
"""simple docstring""" import argparse import copy def a_ ( _lowerCAmelCase : Optional[int] ): '''simple docstring''' lowercase__ : int = {} with open(_lowerCAmelCase ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: lowercase__ : str = [] _list.append([line.split()[1], line.split()[2]] ) lowercase__ : Dict = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: lowercase__ : Dict = [] _list.append([line.split()[0], line.split()[2]] ) lowercase__ : Optional[int] = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' with open(_lowerCAmelCase ) as f: lowercase__ : Union[str, Any] = f.read(1 ) lowercase__ : Union[str, Any] = start_node lowercase__ : int = [] lowercase__ : int = start_node lowercase__ : Optional[Any] = 0 while visiting not in first_solution: lowercase__ : str = 1_0000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(_lowerCAmelCase ) and k[0] not in first_solution: lowercase__ : Any = k[1] lowercase__ : Union[str, Any] = k[0] first_solution.append(_lowerCAmelCase ) lowercase__ : Dict = distance_of_first_solution + int(_lowerCAmelCase ) lowercase__ : Tuple = best_node first_solution.append(_lowerCAmelCase ) lowercase__ : int = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 lowercase__ : str = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0000 ) return first_solution, distance_of_first_solution def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Optional[int] = [] for n in solution[1:-1]: lowercase__ : int = solution.index(_lowerCAmelCase ) for kn in solution[1:-1]: lowercase__ : List[Any] = solution.index(_lowerCAmelCase ) if n == kn: continue lowercase__ : Optional[Any] = copy.deepcopy(_lowerCAmelCase ) lowercase__ : Tuple = kn lowercase__ : Dict = n lowercase__ : List[str] = 0 for k in _tmp[:-1]: lowercase__ : Optional[int] = _tmp[_tmp.index(_lowerCAmelCase ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: lowercase__ : Optional[Any] = distance + int(i[1] ) _tmp.append(_lowerCAmelCase ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) lowercase__ : Dict = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda _lowerCAmelCase : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def a_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple ): '''simple docstring''' lowercase__ : List[str] = 1 lowercase__ : Optional[int] = first_solution lowercase__ : List[Any] = [] lowercase__ : Optional[int] = distance_of_first_solution lowercase__ : List[str] = solution while count <= iters: lowercase__ : Any = find_neighborhood(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Any = 0 lowercase__ : int = neighborhood[index_of_best_solution] lowercase__ : List[Any] = len(_lowerCAmelCase ) - 1 lowercase__ : List[str] = False while not found: lowercase__ : Optional[Any] = 0 while i < len(_lowerCAmelCase ): if best_solution[i] != solution[i]: lowercase__ : Union[str, Any] = best_solution[i] lowercase__ : Tuple = solution[i] break lowercase__ : str = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) lowercase__ : List[str] = True lowercase__ : List[str] = best_solution[:-1] lowercase__ : List[str] = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: lowercase__ : List[Any] = cost lowercase__ : Tuple = solution else: lowercase__ : Tuple = index_of_best_solution + 1 lowercase__ : Any = neighborhood[index_of_best_solution] if len(_lowerCAmelCase ) >= size: tabu_list.pop(0 ) lowercase__ : List[Any] = count + 1 return best_solution_ever, best_cost def a_ ( _lowerCAmelCase : Tuple=None ): '''simple docstring''' lowercase__ : Tuple = generate_neighbours(args.File ) lowercase__ , lowercase__ : Optional[int] = generate_first_solution( args.File , _lowerCAmelCase ) lowercase__ , lowercase__ : Optional[int] = tabu_search( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , args.Iterations , args.Size , ) print(f"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = argparse.ArgumentParser(description="Tabu Search") parser.add_argument( "-f", "--File", type=str, help="Path to the file containing the data", required=True, ) parser.add_argument( "-i", "--Iterations", type=int, help="How many iterations the algorithm should perform", required=True, ) parser.add_argument( "-s", "--Size", type=int, help="Size of the tabu list", required=True ) # Pass the arguments to main method main(parser.parse_args())
77
"""simple docstring""" import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class UpperCAmelCase_ : def __init__( self , a ) -> List[str]: if isinstance(a , a ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden lowercase__ : Optional[Any] = deepcopy(a ) elif os.path.exists(a ): with io.open(a , 'r' , encoding='utf-8' ) as f: lowercase__ : List[Any] = json.load(a ) else: try: lowercase__ : Optional[int] = baseaa.urlsafe_baadecode(a ).decode('utf-8' ) lowercase__ : List[str] = json.loads(a ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""" ) lowercase__ : Any = config self.set_stage_and_offload() def _UpperCAmelCase ( self ) -> Dict: # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. lowercase__ : Tuple = self.get_value('zero_optimization.stage' , -1 ) # offload lowercase__ : int = False if self.is_zeroa() or self.is_zeroa(): lowercase__ : str = set(['cpu', 'nvme'] ) lowercase__ : Optional[Any] = set( [ self.get_value('zero_optimization.offload_optimizer.device' ), self.get_value('zero_optimization.offload_param.device' ), ] ) if len(offload_devices & offload_devices_valid ) > 0: lowercase__ : Optional[Any] = True def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Dict = self.config # find the config node of interest if it exists lowercase__ : int = ds_key_long.split('.' ) lowercase__ : Dict = nodes.pop() for node in nodes: lowercase__ : Optional[Any] = config.get(a ) if config is None: return None, ds_key return config, ds_key def _UpperCAmelCase ( self , a , a=None ) -> Union[str, Any]: lowercase__ , lowercase__ : Tuple = self.find_config_node(a ) if config is None: return default return config.get(a , a ) def _UpperCAmelCase ( self , a , a=False ) -> Any: lowercase__ : str = self.config # find the config node of interest if it exists lowercase__ : List[Any] = ds_key_long.split('.' ) for node in nodes: lowercase__ : str = config lowercase__ : str = config.get(a ) if config is None: if must_exist: raise ValueError(f"""Can't find {ds_key_long} entry in the config: {self.config}""" ) else: return # if found remove it if parent_config is not None: parent_config.pop(a ) def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : Union[str, Any] = self.get_value(a ) return False if value is None else bool(a ) def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Any = self.get_value(a ) return False if value is None else not bool(a ) def _UpperCAmelCase ( self ) -> Tuple: return self._stage == 2 def _UpperCAmelCase ( self ) -> List[Any]: return self._stage == 3 def _UpperCAmelCase ( self ) -> str: return self._offload class UpperCAmelCase_ : def __init__( self , a ) -> str: lowercase__ : Tuple = engine def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: # runs backpropagation and handles mixed precision self.engine.backward(a , **a ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class UpperCAmelCase_ ( _a): def __init__( self , a ) -> Dict: super().__init__(a , device_placement=a , scaler=a ) lowercase__ : Union[str, Any] = hasattr(self.optimizer , 'overflow' ) def _UpperCAmelCase ( self , a=None ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def _UpperCAmelCase ( self ) -> Optional[int]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def _UpperCAmelCase ( self ) -> Tuple: if self.__has_overflow__: return self.optimizer.overflow return False class UpperCAmelCase_ ( _a): def __init__( self , a , a ) -> Any: super().__init__(a , a ) def _UpperCAmelCase ( self ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class UpperCAmelCase_ : def __init__( self , a , a=0.001 , a=0 , **a ) -> Tuple: lowercase__ : List[Any] = params lowercase__ : int = lr lowercase__ : int = weight_decay lowercase__ : Union[str, Any] = kwargs class UpperCAmelCase_ : def __init__( self , a , a=None , a=0 , **a ) -> Tuple: lowercase__ : Dict = optimizer lowercase__ : List[str] = total_num_steps lowercase__ : Optional[int] = warmup_num_steps lowercase__ : List[Any] = kwargs
77
1
"""simple docstring""" import argparse import importlib from pathlib import Path # Test all the extensions added in the setup _UpperCamelCase : Any = [ "kernels/rwkv/wkv_cuda.cu", "kernels/rwkv/wkv_op.cpp", "kernels/deformable_detr/ms_deform_attn.h", "kernels/deformable_detr/cuda/ms_deform_im2col_cuda.cuh", "models/graphormer/algos_graphormer.pyx", ] def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' for file in FILES_TO_FIND: if not (transformers_path / file).exists(): return False return True if __name__ == "__main__": _UpperCamelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--check_lib", action="store_true", help="Whether to check the build or the actual package.") _UpperCamelCase : str = parser.parse_args() if args.check_lib: _UpperCamelCase : List[Any] = importlib.import_module("transformers") _UpperCamelCase : str = Path(transformers_module.__file__).parent else: _UpperCamelCase : List[str] = Path.cwd() / "build/lib/transformers" if not test_custom_files_are_present(transformers_path): raise ValueError("The built release does not contain the custom files. Fix this before going further!")
77
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _UpperCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , *a , **a ) -> Union[str, Any]: super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _UpperCAmelCase ( self , a=None ) -> Dict: lowercase__ : Any = {} if top_k is not None: lowercase__ : List[str] = top_k return {}, {}, postprocess_params def __call__( self , a , **a ) -> Tuple: return super().__call__(a , **a ) def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : List[Any] = load_image(a ) lowercase__ : Union[str, Any] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _UpperCAmelCase ( self , a ) -> List[str]: lowercase__ : Dict = self.model(**a ) return model_outputs def _UpperCAmelCase ( self , a , a=5 ) -> Dict: if top_k > self.model.config.num_labels: lowercase__ : List[Any] = self.model.config.num_labels if self.framework == "pt": lowercase__ : Tuple = model_outputs.logits.softmax(-1 )[0] lowercase__ , lowercase__ : Optional[Any] = probs.topk(a ) elif self.framework == "tf": lowercase__ : Union[str, Any] = stable_softmax(model_outputs.logits , axis=-1 )[0] lowercase__ : str = tf.math.top_k(a , k=a ) lowercase__ , lowercase__ : Dict = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) lowercase__ : Dict = scores.tolist() lowercase__ : Dict = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
77
1
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : List[Any] = 'ZinengTang/tvlt-base' lowercase__ : List[Any] = tempfile.mkdtemp() def _UpperCAmelCase ( self , **a ) -> Any: return TvltImageProcessor.from_pretrained(self.checkpoint , **a ) def _UpperCAmelCase ( self , **a ) -> Union[str, Any]: return TvltFeatureExtractor.from_pretrained(self.checkpoint , **a ) def _UpperCAmelCase ( self ) -> int: shutil.rmtree(self.tmpdirname ) def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Tuple = self.get_image_processor() lowercase__ : int = self.get_feature_extractor() lowercase__ : Optional[Any] = TvltProcessor(image_processor=a , feature_extractor=a ) processor.save_pretrained(self.tmpdirname ) lowercase__ : Optional[Any] = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , a ) self.assertIsInstance(processor.image_processor , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : Any = self.get_image_processor() lowercase__ : Dict = self.get_feature_extractor() lowercase__ : Optional[Any] = TvltProcessor(image_processor=a , feature_extractor=a ) lowercase__ : Optional[Any] = np.ones([1_2_0_0_0] ) lowercase__ : str = feature_extractor(a , return_tensors='np' ) lowercase__ : Union[str, Any] = processor(audio=a , return_tensors='np' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : Union[str, Any] = self.get_image_processor() lowercase__ : int = self.get_feature_extractor() lowercase__ : int = TvltProcessor(image_processor=a , feature_extractor=a ) lowercase__ : Tuple = np.ones([3, 2_2_4, 2_2_4] ) lowercase__ : Dict = image_processor(a , return_tensors='np' ) lowercase__ : List[str] = processor(images=a , return_tensors='np' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = self.get_image_processor() lowercase__ : Tuple = self.get_feature_extractor() lowercase__ : Union[str, Any] = TvltProcessor(image_processor=a , feature_extractor=a ) lowercase__ : Optional[Any] = np.ones([1_2_0_0_0] ) lowercase__ : Any = np.ones([3, 2_2_4, 2_2_4] ) lowercase__ : Union[str, Any] = processor(audio=a , images=a ) self.assertListEqual(list(inputs.keys() ) , ['audio_values', 'audio_mask', 'pixel_values', 'pixel_mask'] ) # test if it raises when no input is passed with pytest.raises(a ): processor() def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : List[str] = self.get_image_processor() lowercase__ : Optional[int] = self.get_feature_extractor() lowercase__ : Any = TvltProcessor(image_processor=a , feature_extractor=a ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='`processor` and `image_processor`+`feature_extractor` model input names do not match' , )
77
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : str = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all LED models at https://huggingface.co/models?filter=LED _UpperCamelCase : Optional[Any] = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } _UpperCamelCase : Optional[int] = { "allenai/led-base-16384": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def a_ ( ): '''simple docstring''' lowercase__ : int = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowercase__ : Union[str, Any] = bs[:] lowercase__ : str = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCAmelCase ) cs.append(2**8 + n ) n += 1 lowercase__ : str = [chr(_lowerCAmelCase ) for n in cs] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Dict = set() lowercase__ : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Optional[Any] = char return pairs class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = VOCAB_FILES_NAMES lowerCamelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , a , a , a="replace" , a="<s>" , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a=False , **a , ) -> Any: lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else bos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else eos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else sep_token lowercase__ : Dict = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else cls_token lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else unk_token lowercase__ : Tuple = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( errors=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , add_prefix_space=a , **a , ) with open(a , encoding='utf-8' ) as vocab_handle: lowercase__ : Tuple = json.load(a ) lowercase__ : Dict = {v: k for k, v in self.encoder.items()} lowercase__ : str = errors # how to handle errors in decoding lowercase__ : Optional[Any] = bytes_to_unicode() lowercase__ : Optional[Any] = {v: k for k, v in self.byte_encoder.items()} with open(a , encoding='utf-8' ) as merges_handle: lowercase__ : Optional[Any] = merges_handle.read().split('\n' )[1:-1] lowercase__ : Optional[int] = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ : Union[str, Any] = dict(zip(a , range(len(a ) ) ) ) lowercase__ : Tuple = {} lowercase__ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ : List[Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _UpperCAmelCase ( self ) -> List[Any]: return len(self.encoder ) def _UpperCAmelCase ( self ) -> str: return dict(self.encoder , **self.added_tokens_encoder ) def _UpperCAmelCase ( self , a ) -> List[str]: if token in self.cache: return self.cache[token] lowercase__ : Optional[Any] = tuple(a ) lowercase__ : int = get_pairs(a ) if not pairs: return token while True: lowercase__ : List[str] = min(a , key=lambda a : self.bpe_ranks.get(a , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ : List[str] = bigram lowercase__ : Union[str, Any] = [] lowercase__ : List[Any] = 0 while i < len(a ): try: lowercase__ : str = word.index(a , a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : Optional[int] = j if word[i] == first and i < len(a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : int = tuple(a ) lowercase__ : Dict = new_word if len(a ) == 1: break else: lowercase__ : Any = get_pairs(a ) lowercase__ : List[str] = ' '.join(a ) lowercase__ : Optional[Any] = word return word def _UpperCAmelCase ( self , a ) -> Union[str, Any]: lowercase__ : Tuple = [] for token in re.findall(self.pat , a ): lowercase__ : Union[str, Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(a ).split(' ' ) ) return bpe_tokens def _UpperCAmelCase ( self , a ) -> Optional[Any]: return self.encoder.get(a , self.encoder.get(self.unk_token ) ) def _UpperCAmelCase ( self , a ) -> Optional[int]: return self.decoder.get(a ) def _UpperCAmelCase ( self , a ) -> str: lowercase__ : Any = ''.join(a ) lowercase__ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Any = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : str = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(a , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a , ensure_ascii=a ) + '\n' ) lowercase__ : List[Any] = 0 with open(a , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) lowercase__ : Union[str, Any] = token_index writer.write(' '.join(a ) + '\n' ) index += 1 return vocab_file, merge_file def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ : Union[str, Any] = [self.cls_token_id] lowercase__ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) + [1] def _UpperCAmelCase ( self , a , a = None ) -> List[int]: lowercase__ : Dict = [self.sep_token_id] lowercase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCAmelCase ( self , a , a=False , **a ) -> Optional[int]: lowercase__ : Tuple = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(a ) > 0 and not text[0].isspace()): lowercase__ : List[str] = ' ' + text return (text, kwargs) def _UpperCAmelCase ( self , a , a = None , a = PaddingStrategy.DO_NOT_PAD , a = None , a = None , ) -> dict: lowercase__ : Dict = super()._pad( encoded_inputs=a , max_length=a , padding_strategy=a , pad_to_multiple_of=a , return_attention_mask=a , ) # Load from model defaults if return_attention_mask is None: lowercase__ : Union[str, Any] = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ : Any = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ : Tuple = len(encoded_inputs['global_attention_mask'] ) != len(a ) if needs_to_be_padded: lowercase__ : str = len(a ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ : Union[str, Any] = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowercase__ : List[str] = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
77
1
"""simple docstring""" from random import randint, random def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : bool = False , _lowerCAmelCase : bool = False , _lowerCAmelCase : int = 5 , ): '''simple docstring''' lowercase__ : str = [[-1] * number_of_cells] # Create a highway without any car lowercase__ : Union[str, Any] = 0 lowercase__ : List[Any] = max(_lowerCAmelCase , 0 ) while i < number_of_cells: lowercase__ : Tuple = ( randint(0 , _lowerCAmelCase ) if random_speed else initial_speed ) # Place the cars i += ( randint(1 , max_speed * 2 ) if random_frequency else frequency ) # Arbitrary number, may need tuning return highway def a_ ( _lowerCAmelCase : list , _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Tuple = 0 lowercase__ : List[Any] = highway_now[car_index + 1 :] for cell in range(len(_lowerCAmelCase ) ): # May need a better name for this if cells[cell] != -1: # If the cell is not empty then return distance # we have the distance we wanted distance += 1 # Here if the car is near the end of the highway return distance + get_distance(_lowerCAmelCase , -1 ) def a_ ( _lowerCAmelCase : list , _lowerCAmelCase : float , _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : List[Any] = len(_lowerCAmelCase ) # Beforce calculations, the highway is empty lowercase__ : Any = [-1] * number_of_cells for car_index in range(_lowerCAmelCase ): if highway_now[car_index] != -1: # Add 1 to the current speed of the car and cap the speed lowercase__ : Dict = min(highway_now[car_index] + 1 , _lowerCAmelCase ) # Number of empty cell before the next car lowercase__ : Any = get_distance(_lowerCAmelCase , _lowerCAmelCase ) - 1 # We can't have the car causing an accident lowercase__ : Any = min(next_highway[car_index] , _lowerCAmelCase ) if random() < probability: # Randomly, a driver will slow down lowercase__ : Tuple = max(next_highway[car_index] - 1 , 0 ) return next_highway def a_ ( _lowerCAmelCase : list , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Any = len(highway[0] ) for i in range(_lowerCAmelCase ): lowercase__ : List[Any] = update(highway[i] , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Union[str, Any] = [-1] * number_of_cells for car_index in range(_lowerCAmelCase ): lowercase__ : Any = next_speeds_calculated[car_index] if speed != -1: # Change the position based on the speed (with % to create the loop) lowercase__ : Optional[int] = (car_index + speed) % number_of_cells # Commit the change of position lowercase__ : Any = speed highway.append(_lowerCAmelCase ) return highway if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" _UpperCamelCase : Union[str, Any] = 8.3_1_4_4_5_9_8 def a_ ( _lowerCAmelCase : float , _lowerCAmelCase : float ): '''simple docstring''' if temperature < 0: raise Exception('Temperature cannot be less than 0 K' ) if molar_mass <= 0: raise Exception('Molar mass cannot be less than or equal to 0 kg/mol' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _UpperCamelCase : List[Any] = 3_00 _UpperCamelCase : Tuple = 28 _UpperCamelCase : Any = rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
77
1
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def a_ ( _lowerCAmelCase : Callable[[int | float], int | float] , _lowerCAmelCase : int | float , _lowerCAmelCase : int | float , _lowerCAmelCase : int = 100 , ): '''simple docstring''' lowercase__ : Dict = x_start lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) lowercase__ : Optional[Any] = 0.0 for _ in range(_lowerCAmelCase ): # Approximates curve as a sequence of linear lines and sums their length lowercase__ : Union[str, Any] = (x_end - x_start) / steps + xa lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step lowercase__ : Union[str, Any] = xa lowercase__ : int = fxa return length if __name__ == "__main__": def a_ ( _lowerCAmelCase : List[Any] ): '''simple docstring''' return math.sin(10 * x ) print("f(x) = sin(10 * x)") print("The length of the curve from x = -10 to x = 10 is:") _UpperCamelCase : str = 10 while i <= 10_00_00: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
77
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization lowerCamelCase__ : str = field(default="text-classification" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({"labels": ClassLabel}) lowerCamelCase__ : str = "text" lowerCamelCase__ : str = "labels" def _UpperCAmelCase ( self , a ) -> Tuple: if self.label_column not in features: raise ValueError(f"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , a ): raise ValueError(f"""Column {self.label_column} is not a ClassLabel.""" ) lowercase__ : Optional[Any] = copy.deepcopy(self ) lowercase__ : Optional[Any] = self.label_schema.copy() lowercase__ : Any = features[self.label_column] lowercase__ : Optional[Any] = label_schema return task_template @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return { self.text_column: "text", self.label_column: "labels", }
77
1
"""simple docstring""" import os import pytest from attr import dataclass _UpperCamelCase : Any = "us-east-1" # defaults region @dataclass class UpperCAmelCase_ : lowerCamelCase__ : str lowerCamelCase__ : List[Any] = "arn:aws:iam::558105141721:role/sagemaker_execution_role" lowerCamelCase__ : str = { "task_name": "mnli", "per_device_train_batch_size": 1_6, "per_device_eval_batch_size": 1_6, "do_train": True, "do_eval": True, "do_predict": True, "output_dir": "/opt/ml/model", "overwrite_output_dir": True, "max_steps": 5_0_0, "save_steps": 5_5_0_0, } lowerCamelCase__ : Optional[int] = {**hyperparameters, "max_steps": 1_0_0_0} @property def _UpperCAmelCase ( self ) -> str: if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def _UpperCAmelCase ( self ) -> str: return f"""{self.framework}-transfromers-test""" @property def _UpperCAmelCase ( self ) -> str: return f"""./tests/sagemaker/scripts/{self.framework}""" @property def _UpperCAmelCase ( self ) -> str: if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='class' ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Tuple = SageMakerTestEnvironment(framework=request.cls.framework )
77
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np _UpperCamelCase : Any = re.compile(r"\b(a|an|the)\b", re.UNICODE) _UpperCamelCase : Union[str, Any] = None def a_ ( ): '''simple docstring''' lowercase__ : Optional[int] = argparse.ArgumentParser('Official evaluation script for SQuAD version 2.0.' ) parser.add_argument('data_file' , metavar='data.json' , help='Input data JSON file.' ) parser.add_argument('pred_file' , metavar='pred.json' , help='Model predictions.' ) parser.add_argument( '--out-file' , '-o' , metavar='eval.json' , help='Write accuracy metrics to file (default is stdout).' ) parser.add_argument( '--na-prob-file' , '-n' , metavar='na_prob.json' , help='Model estimates of probability of no answer.' ) parser.add_argument( '--na-prob-thresh' , '-t' , type=_lowerCAmelCase , default=1.0 , help='Predict "" if no-answer probability exceeds this (default = 1.0).' , ) parser.add_argument( '--out-image-dir' , '-p' , metavar='out_images' , default=_lowerCAmelCase , help='Save precision-recall curves to directory.' ) parser.add_argument('--verbose' , '-v' , action='store_true' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Tuple = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Optional[int] = bool(qa['answers']['text'] ) return qid_to_has_ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' def remove_articles(_lowerCAmelCase : int ): return ARTICLES_REGEX.sub(' ' , _lowerCAmelCase ) def white_space_fix(_lowerCAmelCase : str ): return " ".join(text.split() ) def remove_punc(_lowerCAmelCase : List[Any] ): lowercase__ : int = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_lowerCAmelCase : List[str] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_lowerCAmelCase ) ) ) ) def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' if not s: return [] return normalize_answer(_lowerCAmelCase ).split() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ): '''simple docstring''' return int(normalize_answer(_lowerCAmelCase ) == normalize_answer(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ): '''simple docstring''' lowercase__ : Dict = get_tokens(_lowerCAmelCase ) lowercase__ : List[str] = get_tokens(_lowerCAmelCase ) lowercase__ : List[Any] = collections.Counter(_lowerCAmelCase ) & collections.Counter(_lowerCAmelCase ) lowercase__ : int = sum(common.values() ) if len(_lowerCAmelCase ) == 0 or len(_lowerCAmelCase ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 lowercase__ : Any = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Dict = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Any = (2 * precision * recall) / (precision + recall) return fa def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Optional[int] = {} lowercase__ : Union[str, Any] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Any = qa['id'] lowercase__ : Union[str, Any] = [t for t in qa['answers']['text'] if normalize_answer(_lowerCAmelCase )] if not gold_answers: # For unanswerable questions, only correct answer is empty string lowercase__ : Dict = [''] if qid not in preds: print(f"""Missing prediction for {qid}""" ) continue lowercase__ : Optional[int] = preds[qid] # Take max over all gold answers lowercase__ : int = max(compute_exact(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) lowercase__ : Optional[Any] = max(compute_fa(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) return exact_scores, fa_scores def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : str = {} for qid, s in scores.items(): lowercase__ : int = na_probs[qid] > na_prob_thresh if pred_na: lowercase__ : Optional[Any] = float(not qid_to_has_ans[qid] ) else: lowercase__ : Optional[Any] = s return new_scores def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str=None ): '''simple docstring''' if not qid_list: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores.values() ) / total), ('f1', 1_0_0.0 * sum(fa_scores.values() ) / total), ('total', total), ] ) else: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores[k] for k in qid_list ) / total), ('f1', 1_0_0.0 * sum(fa_scores[k] for k in qid_list ) / total), ('total', total), ] ) def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' for k in new_eval: lowercase__ : int = new_eval[k] def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple ): '''simple docstring''' plt.step(_lowerCAmelCase , _lowerCAmelCase , color='b' , alpha=0.2 , where='post' ) plt.fill_between(_lowerCAmelCase , _lowerCAmelCase , step='post' , alpha=0.2 , color='b' ) plt.xlabel('Recall' ) plt.ylabel('Precision' ) plt.xlim([0.0, 1.0_5] ) plt.ylim([0.0, 1.0_5] ) plt.title(_lowerCAmelCase ) plt.savefig(_lowerCAmelCase ) plt.clf() def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : str , _lowerCAmelCase : Any=None , _lowerCAmelCase : List[str]=None ): '''simple docstring''' lowercase__ : Optional[int] = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) lowercase__ : Tuple = 0.0 lowercase__ : List[str] = 1.0 lowercase__ : List[str] = 0.0 lowercase__ : Union[str, Any] = [1.0] lowercase__ : List[Any] = [0.0] lowercase__ : Optional[int] = 0.0 for i, qid in enumerate(_lowerCAmelCase ): if qid_to_has_ans[qid]: true_pos += scores[qid] lowercase__ : Tuple = true_pos / float(i + 1 ) lowercase__ : Union[str, Any] = true_pos / float(_lowerCAmelCase ) if i == len(_lowerCAmelCase ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_lowerCAmelCase ) recalls.append(_lowerCAmelCase ) if out_image: plot_pr_curve(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return {"ap": 1_0_0.0 * avg_prec} def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple ): '''simple docstring''' if out_image_dir and not os.path.exists(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) lowercase__ : List[str] = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return lowercase__ : Dict = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_exact.png' ) , title='Precision-Recall curve for Exact Match score' , ) lowercase__ : Tuple = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_f1.png' ) , title='Precision-Recall curve for F1 score' , ) lowercase__ : List[Any] = {k: float(_lowerCAmelCase ) for k, v in qid_to_has_ans.items()} lowercase__ : Any = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_oracle.png' ) , title='Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)' , ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_exact' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_f1' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_oracle' ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' if not qid_list: return lowercase__ : List[str] = [na_probs[k] for k in qid_list] lowercase__ : Tuple = np.ones_like(_lowerCAmelCase ) / float(len(_lowerCAmelCase ) ) plt.hist(_lowerCAmelCase , weights=_lowerCAmelCase , bins=20 , range=(0.0, 1.0) ) plt.xlabel('Model probability of no-answer' ) plt.ylabel('Proportion of dataset' ) plt.title(f"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(_lowerCAmelCase , f"""na_prob_hist_{name}.png""" ) ) plt.clf() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : Tuple = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) lowercase__ : int = num_no_ans lowercase__ : Optional[int] = cur_score lowercase__ : Tuple = 0.0 lowercase__ : Dict = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) for i, qid in enumerate(_lowerCAmelCase ): if qid not in scores: continue if qid_to_has_ans[qid]: lowercase__ : Optional[int] = scores[qid] else: if preds[qid]: lowercase__ : List[Any] = -1 else: lowercase__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: lowercase__ : Dict = cur_score lowercase__ : Optional[int] = na_probs[qid] return 1_0_0.0 * best_score / len(_lowerCAmelCase ), best_thresh def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str ): '''simple docstring''' lowercase__ , lowercase__ : List[Any] = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ , lowercase__ : Dict = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Any = best_exact lowercase__ : Tuple = exact_thresh lowercase__ : Optional[Any] = best_fa lowercase__ : Any = fa_thresh def a_ ( ): '''simple docstring''' with open(OPTS.data_file ) as f: lowercase__ : List[Any] = json.load(_lowerCAmelCase ) lowercase__ : Union[str, Any] = dataset_json['data'] with open(OPTS.pred_file ) as f: lowercase__ : str = json.load(_lowerCAmelCase ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: lowercase__ : Union[str, Any] = json.load(_lowerCAmelCase ) else: lowercase__ : str = {k: 0.0 for k in preds} lowercase__ : int = make_qid_to_has_ans(_lowerCAmelCase ) # maps qid to True/False lowercase__ : List[str] = [k for k, v in qid_to_has_ans.items() if v] lowercase__ : Any = [k for k, v in qid_to_has_ans.items() if not v] lowercase__ , lowercase__ : Any = get_raw_scores(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Optional[Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Union[str, Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Tuple = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase ) if has_ans_qids: lowercase__ : int = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'HasAns' ) if no_ans_qids: lowercase__ : Optional[Any] = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'NoAns' ) if OPTS.na_prob_file: find_all_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'hasAns' ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'noAns' ) if OPTS.out_file: with open(OPTS.out_file , 'w' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) else: print(json.dumps(_lowerCAmelCase , indent=2 ) ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
77
1
"""simple docstring""" # DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class UpperCAmelCase_ ( _a , _a): lowerCamelCase__ : Optional[int] = 1 @register_to_config def __init__( self , a=2_0_0_0 , a=0.1 , a=2_0 , a=1e-3 ) -> Dict: lowercase__ : Any = None lowercase__ : Optional[Any] = None lowercase__ : List[Any] = None def _UpperCAmelCase ( self , a , a = None ) -> List[Any]: lowercase__ : str = torch.linspace(1 , self.config.sampling_eps , a , device=a ) def _UpperCAmelCase ( self , a , a , a , a=None ) -> str: if self.timesteps is None: raise ValueError( '`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler' ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score lowercase__ : Optional[Any] = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) lowercase__ : List[Any] = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) lowercase__ : Union[str, Any] = std.flatten() while len(std.shape ) < len(score.shape ): lowercase__ : List[str] = std.unsqueeze(-1 ) lowercase__ : Optional[Any] = -score / std # compute lowercase__ : Optional[int] = -1.0 / len(self.timesteps ) lowercase__ : Any = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) lowercase__ : Union[str, Any] = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): lowercase__ : List[Any] = beta_t.unsqueeze(-1 ) lowercase__ : str = -0.5 * beta_t * x lowercase__ : Dict = torch.sqrt(a ) lowercase__ : Optional[Any] = drift - diffusion**2 * score lowercase__ : Optional[int] = x + drift * dt # add noise lowercase__ : int = randn_tensor(x.shape , layout=x.layout , generator=a , device=x.device , dtype=x.dtype ) lowercase__ : List[str] = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__( self ) -> Dict: return self.config.num_train_timesteps
77
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = 0 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Tuple = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Any: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : str = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : int = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = CLIPConfig() # Create a dummy config file with image_proceesor_type lowercase__ : Optional[int] = Path(a ) / 'preprocessor_config.json' lowercase__ : Optional[int] = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowercase__ : int = AutoImageProcessor.from_pretrained(a ).to_dict() config_dict.pop('image_processor_type' ) lowercase__ : Tuple = CLIPImageProcessor(**a ) # save in new folder model_config.save_pretrained(a ) config.save_pretrained(a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) # make sure private variable is not incorrectly saved lowercase__ : Optional[int] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = Path(a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'clip-base is not a local folder and is not a valid model identifier' ): lowercase__ : Any = AutoImageProcessor.from_pretrained('clip-base' ) def _UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex( a , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowercase__ : Dict = AutoImageProcessor.from_pretrained(a , revision='aaaaaa' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def _UpperCAmelCase ( self ) -> Optional[int]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(a ): lowercase__ : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : str = AutoImageProcessor.from_pretrained(a , trust_remote_code=a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def _UpperCAmelCase ( self ) -> int: try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoImageProcessor.register(a , a ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Optional[Any] = Path(a ) / 'preprocessor_config.json' lowercase__ : List[Any] = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = CustomImageProcessor.from_pretrained(a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCAmelCase ( self ) -> Dict: class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = True try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # If remote code is not set, the default is to use local lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowercase__ : int = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
77
1