code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from __future__ import annotations from collections import deque class __A : def __init__(self : List[Any] , __a : list[str] ): UpperCAmelCase_ = [] self.adlist.append( {"value": "", "next_states": [], "fail_state": 0, "output": []} ) for keyword in keywords: self.add_keyword(__a ) self.set_fail_transitions() def _lowercase (self : Union[str, Any] , __a : int , __a : str ): for state in self.adlist[current_state]["next_states"]: if char == self.adlist[state]["value"]: return state return None def _lowercase (self : Optional[Any] , __a : str ): UpperCAmelCase_ = 0 for character in keyword: UpperCAmelCase_ = self.find_next_state(__a , __a ) if next_state is None: self.adlist.append( { "value": character, "next_states": [], "fail_state": 0, "output": [], } ) self.adlist[current_state]["next_states"].append(len(self.adlist ) - 1 ) UpperCAmelCase_ = len(self.adlist ) - 1 else: UpperCAmelCase_ = next_state self.adlist[current_state]["output"].append(__a ) def _lowercase (self : int ): UpperCAmelCase_ = deque() for node in self.adlist[0]["next_states"]: q.append(__a ) UpperCAmelCase_ = 0 while q: UpperCAmelCase_ = q.popleft() for child in self.adlist[r]["next_states"]: q.append(__a ) UpperCAmelCase_ = self.adlist[r]["fail_state"] while ( self.find_next_state(__a , self.adlist[child]["value"] ) is None and state != 0 ): UpperCAmelCase_ = self.adlist[state]["fail_state"] UpperCAmelCase_ = self.find_next_state( __a , self.adlist[child]["value"] ) if self.adlist[child]["fail_state"] is None: UpperCAmelCase_ = 0 UpperCAmelCase_ = ( self.adlist[child]["output"] + self.adlist[self.adlist[child]["fail_state"]]["output"] ) def _lowercase (self : Dict , __a : str ): UpperCAmelCase_ = {} # returns a dict with keywords and list of its occurrences UpperCAmelCase_ = 0 for i in range(len(__a ) ): while ( self.find_next_state(__a , string[i] ) is None and current_state != 0 ): UpperCAmelCase_ = self.adlist[current_state]["fail_state"] UpperCAmelCase_ = self.find_next_state(__a , string[i] ) if next_state is None: UpperCAmelCase_ = 0 else: UpperCAmelCase_ = next_state for key in self.adlist[current_state]["output"]: if key not in result: UpperCAmelCase_ = [] result[key].append(i - len(__a ) + 1 ) return result if __name__ == "__main__": import doctest doctest.testmod()
1
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy __lowerCamelCase = logging.getLogger(__name__) __lowerCamelCase = """pytorch_model.bin""" @dataclasses.dataclass class UpperCAmelCase : A__ : str = dataclasses.field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models."} ) A__ : Optional[str] = dataclasses.field( default=A_ ,metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co."} ,) @dataclasses.dataclass class UpperCAmelCase : A__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the training data."} ) A__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the data to predict on."} ) A__ : Optional[str] = dataclasses.field( default=A_ ,metadata={"help": "A csv or a json file containing the validation data."} ) A__ : Optional[str] = dataclasses.field( default=A_ ,metadata={"help": "The name of the task to train on."} ,) A__ : Optional[List[str]] = dataclasses.field( default=A_ ,metadata={"help": "The list of labels for the task."} ) @dataclasses.dataclass class UpperCAmelCase : A__ : str = dataclasses.field( metadata={"help": "The output directory where the model predictions and checkpoints will be written."} ) A__ : Optional[str] = dataclasses.field( default="accuracy" ,metadata={"help": "The evaluation metric used for the task."} ) A__ : Optional[str] = dataclasses.field( default="no" ,metadata={ "help": "The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]" } ,) A__ : Optional[int] = dataclasses.field( default=10 ,metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} ,) A__ : Optional[float] = dataclasses.field( default=0.0 ,metadata={ "help": "How much the specified evaluation metric must improve to satisfy early stopping conditions." } ,) A__ : Optional[bool] = dataclasses.field( default=A_ ,metadata={"help": "Whether to filter the pseudo-labeled data based on the confidence score."} ,) A__ : Optional[bool] = dataclasses.field( default=A_ ,metadata={"help": "Whether to filter the pseudo-labeled data based on the validation performance."} ,) A__ : Optional[bool] = dataclasses.field( default=A_ ,metadata={"help": "Whether to fine-tune on labeled data after pseudo training."} ,) A__ : Optional[float] = dataclasses.field( default=0.0 ,metadata={"help": "Confidence threshold for pseudo-labeled data filtering."} ,) A__ : Optional[int] = dataclasses.field( default=1_00 ,metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} ,) A__ : Optional[int] = dataclasses.field( default=A_ ,metadata={"help": "Random seed for initialization."} ,) def UpperCamelCase ( __lowerCamelCase : int , __lowerCamelCase : Tuple , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : Optional[int] ): snake_case : Tuple = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: snake_case : Optional[int] = dataset.filter(lambda __lowerCamelCase : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 snake_case : int = int(eval_result * len(__lowerCamelCase ) ) print(__lowerCamelCase ) snake_case : List[str] = dataset.sort("probability" , reverse=__lowerCamelCase ) snake_case : Tuple = dataset.select(range(__lowerCamelCase ) ) snake_case : List[Any] = dataset.remove_columns(["label", "probability"] ) snake_case : Any = dataset.rename_column("prediction" , "label" ) snake_case : str = dataset.map(lambda __lowerCamelCase : {"label": idalabel[example["label"]]} ) snake_case : List[str] = dataset.shuffle(seed=args.seed ) snake_case : int = os.path.join(__lowerCamelCase , f"""train_pseudo.{args.data_file_extension}""" ) if args.data_file_extension == "csv": dataset.to_csv(__lowerCamelCase , index=__lowerCamelCase ) else: dataset.to_json(__lowerCamelCase ) def UpperCamelCase ( __lowerCamelCase : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple , **__lowerCamelCase : List[Any] ): snake_case : int = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() snake_case : Dict = STModelArguments(model_name_or_path=__lowerCamelCase ) snake_case : Tuple = STDataArguments(train_file=__lowerCamelCase , infer_file=__lowerCamelCase ) snake_case : str = STTrainingArguments(output_dir=__lowerCamelCase ) snake_case : int = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(__lowerCamelCase ).items(): setattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) for key, value in kwargs.items(): if hasattr(__lowerCamelCase , __lowerCamelCase ): setattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Sanity checks snake_case : List[str] = {} snake_case : Optional[int] = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None snake_case : str = args.train_file snake_case : Tuple = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None snake_case : Tuple = args.eval_file for key in data_files: snake_case : List[Any] = data_files[key].split("." )[-1] assert extension in ["csv", "json"], f"""`{key}_file` should be a csv or a json file.""" if args.data_file_extension is None: snake_case : Union[str, Any] = extension else: assert extension == args.data_file_extension, f"""`{key}_file` should be a {args.data_file_extension} file`.""" assert ( args.eval_metric in datasets.list_metrics() ), f"""{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.""" # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info("Creating the initial data directory for self-training..." ) snake_case : List[Any] = f"""{args.output_dir}/self-train_iter-{{}}""".format snake_case : Optional[int] = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) accelerator.wait_for_everyone() snake_case : Dict = None snake_case : Union[str, Any] = None snake_case : Tuple = 0 snake_case : List[Any] = False # Show the progress bar snake_case : List[Any] = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): snake_case : str = data_dir_format(__lowerCamelCase ) assert os.path.exists(__lowerCamelCase ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 snake_case : Dict = os.path.join(__lowerCamelCase , "stage-1" ) snake_case : Optional[Any] = { "accelerator": accelerator, "model_name_or_path": args.model_name_or_path, "cache_dir": args.cache_dir, "do_train": True, "train_file": data_files["train"] if iteration == 0 else data_files["train_pseudo"], "do_eval": True if args.eval_file is not None else False, "eval_file": data_files["eval"], "do_predict": True, "infer_file": data_files["infer"], "task_name": args.task_name, "label_list": args.label_list, "output_dir": current_output_dir, "eval_metric": args.eval_metric, "evaluation_strategy": args.evaluation_strategy, "early_stopping_patience": args.early_stopping_patience, "early_stopping_threshold": args.early_stopping_threshold, "seed": args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(__lowerCamelCase , __lowerCamelCase ): arguments_dict.update({key: value} ) snake_case : int = os.path.join(__lowerCamelCase , "best-checkpoint" , __lowerCamelCase ) if os.path.exists(__lowerCamelCase ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1." , __lowerCamelCase , __lowerCamelCase , ) else: logger.info("***** Running self-training: iteration: %d, stage: 1 *****" , __lowerCamelCase ) finetune(**__lowerCamelCase ) accelerator.wait_for_everyone() assert os.path.exists(__lowerCamelCase ) logger.info("Self-training job completed: iteration: %d, stage: 1." , __lowerCamelCase ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data snake_case : str = os.path.join(__lowerCamelCase , "best-checkpoint" ) snake_case : Dict = os.path.join(__lowerCamelCase , "stage-2" ) # Update arguments_dict snake_case : List[str] = model_path snake_case : Optional[Any] = data_files["train"] snake_case : Optional[Any] = current_output_dir snake_case : Union[str, Any] = os.path.join(__lowerCamelCase , "best-checkpoint" , __lowerCamelCase ) if os.path.exists(__lowerCamelCase ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2." , __lowerCamelCase , __lowerCamelCase , ) else: logger.info("***** Running self-training: iteration: %d, stage: 2 *****" , __lowerCamelCase ) finetune(**__lowerCamelCase ) accelerator.wait_for_everyone() assert os.path.exists(__lowerCamelCase ) logger.info("Self-training job completed: iteration: %d, stage: 2." , __lowerCamelCase ) snake_case : int = iteration snake_case : Tuple = data_dir_format(iteration + 1 ) snake_case : Tuple = AutoConfig.from_pretrained(os.path.join(__lowerCamelCase , "best-checkpoint" ) ) snake_case : Optional[int] = config.idalabel snake_case : List[Any] = os.path.join(__lowerCamelCase , "eval_results_best-checkpoint.json" ) snake_case : Union[str, Any] = os.path.join(__lowerCamelCase , "test_results_best-checkpoint.json" ) assert os.path.exists(__lowerCamelCase ) with open(__lowerCamelCase , "r" ) as f: snake_case : Dict = float(json.load(__lowerCamelCase )[args.eval_metric] ) snake_case : Optional[int] = os.path.join(__lowerCamelCase , "infer_output_best-checkpoint.csv" ) assert os.path.exists(__lowerCamelCase ) # Loading the dataset from local csv or json files. snake_case : Optional[Any] = load_dataset(args.data_file_extension , data_files={"data": data_files["infer"]} )["data"] snake_case : Dict = load_dataset("csv" , data_files={"data": infer_output_file} )["data"] if accelerator.is_main_process: os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) shutil.copy(__lowerCamelCase , os.path.join(__lowerCamelCase , f"""eval_results_iter-{iteration}.json""" ) ) if os.path.exists(__lowerCamelCase ): shutil.copy(__lowerCamelCase , os.path.join(__lowerCamelCase , f"""test_results_iter-{iteration}.json""" ) ) create_pseudo_labeled_data(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) accelerator.wait_for_everyone() snake_case : str = os.path.join(__lowerCamelCase , f"""train_pseudo.{args.data_file_extension}""" ) if args.evaluation_strategy != IntervalStrategy.NO.value: snake_case : List[Any] = eval_result if best_iteration is None: snake_case : List[Any] = new_iteration snake_case : int = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: snake_case : int = new_iteration snake_case : Union[str, Any] = new_eval_result snake_case : str = 0 else: if new_eval_result == best_eval_result: snake_case : Any = new_iteration snake_case : Union[str, Any] = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: snake_case : Tuple = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info("Best iteration: %d" , __lowerCamelCase ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , __lowerCamelCase ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(__lowerCamelCase , f"""eval_results_iter-{iteration}.json""" ) , os.path.join(__lowerCamelCase , "eval_results_best-iteration.json" ) , ) else: # Assume that the last iteration is the best logger.info("Best iteration: %d" , args.max_selftrain_iterations - 1 ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , __lowerCamelCase ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(__lowerCamelCase , f"""eval_results_iter-{args.max_selftrain_iterations - 1}.json""" ) , os.path.join(__lowerCamelCase , "eval_results_best-iteration.json" ) , )
59
0
'''simple docstring''' import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 _lowercase = data_utils.TransfoXLTokenizer _lowercase = data_utils.TransfoXLCorpus _lowercase = data_utils _lowercase = data_utils def A (__lowerCamelCase :Optional[Any] , __lowerCamelCase :List[str] , __lowerCamelCase :Dict , __lowerCamelCase :Tuple ): if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(lowerCAmelCase__ , """rb""" ) as fp: _lowerCAmelCase = pickle.load(lowerCAmelCase__ , encoding="""latin1""" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) _lowerCAmelCase = pytorch_dump_folder_path + '''/''' + VOCAB_FILES_NAMES['''pretrained_vocab_file'''] print(f'Save vocabulary to {pytorch_vocab_dump_path}' ) _lowerCAmelCase = corpus.vocab.__dict__ torch.save(lowerCAmelCase__ , lowerCAmelCase__ ) _lowerCAmelCase = corpus.__dict__ corpus_dict_no_vocab.pop("""vocab""" , lowerCAmelCase__ ) _lowerCAmelCase = pytorch_dump_folder_path + '''/''' + CORPUS_NAME print(f'Save dataset to {pytorch_dataset_dump_path}' ) torch.save(lowerCAmelCase__ , lowerCAmelCase__ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model _lowerCAmelCase = os.path.abspath(lowerCAmelCase__ ) _lowerCAmelCase = os.path.abspath(lowerCAmelCase__ ) print(f'Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.' ) # Initialise PyTorch model if transfo_xl_config_file == "": _lowerCAmelCase = TransfoXLConfig() else: _lowerCAmelCase = TransfoXLConfig.from_json_file(lowerCAmelCase__ ) print(f'Building PyTorch model from configuration: {config}' ) _lowerCAmelCase = TransfoXLLMHeadModel(lowerCAmelCase__ ) _lowerCAmelCase = load_tf_weights_in_transfo_xl(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Save pytorch-model _lowerCAmelCase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) _lowerCAmelCase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) print(f'Save PyTorch model to {os.path.abspath(lowerCAmelCase__ )}' ) torch.save(model.state_dict() , lowerCAmelCase__ ) print(f'Save configuration file to {os.path.abspath(lowerCAmelCase__ )}' ) with open(lowerCAmelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the folder to store the PyTorch model or dataset/vocab.""", ) parser.add_argument( """--tf_checkpoint_path""", default="""""", type=str, help="""An optional path to a TensorFlow checkpoint path to be converted.""", ) parser.add_argument( """--transfo_xl_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained BERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--transfo_xl_dataset_file""", default="""""", type=str, help="""An optional dataset file to be converted in a vocabulary.""", ) _lowercase = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
352
'''simple docstring''' import numpy as np import qiskit def A (__lowerCamelCase :int = 8 , __lowerCamelCase :int | None = None ): _lowerCAmelCase = np.random.default_rng(seed=__lowerCamelCase ) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. _lowerCAmelCase = 6 * key_len # Measurement basis for Alice's qubits. _lowerCAmelCase = rng.integers(2 , size=__lowerCamelCase ) # The set of states Alice will prepare. _lowerCAmelCase = rng.integers(2 , size=__lowerCamelCase ) # Measurement basis for Bob's qubits. _lowerCAmelCase = rng.integers(2 , size=__lowerCamelCase ) # Quantum Circuit to simulate BB84 _lowerCAmelCase = qiskit.QuantumCircuit(__lowerCamelCase , name="""BB84""" ) # Alice prepares her qubits according to rules above. for index, _ in enumerate(__lowerCamelCase ): if alice_state[index] == 1: bbaa_circ.x(__lowerCamelCase ) if alice_basis[index] == 1: bbaa_circ.h(__lowerCamelCase ) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(__lowerCamelCase ): if bob_basis[index] == 1: bbaa_circ.h(__lowerCamelCase ) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. _lowerCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. _lowerCAmelCase = qiskit.execute(__lowerCamelCase , __lowerCamelCase , shots=1 , seed_simulator=__lowerCamelCase ) # Returns the result of measurement. _lowerCAmelCase = job.result().get_counts(__lowerCamelCase ).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. _lowerCAmelCase = """""".join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if alice_basis_bit == bob_basis_bit ] ) # Get final key. Pad with 0 if too short, otherwise truncate. _lowerCAmelCase = gen_key[:key_len] if len(__lowerCamelCase ) >= key_len else gen_key.ljust(__lowerCamelCase , """0""" ) return key if __name__ == "__main__": print(F"""The generated key is : {bbaa(8, seed=0)}""") from doctest import testmod testmod()
229
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:List[str] = { """google/mobilenet_v2_1.4_224""": """https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json""", """google/mobilenet_v2_1.0_224""": """https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json""", """google/mobilenet_v2_0.75_160""": """https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json""", """google/mobilenet_v2_0.35_96""": """https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json""", # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class snake_case__ ( snake_case_ ): _snake_case : Tuple = """mobilenet_v2""" def __init__( self , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=1.0 , lowerCamelCase=8 , lowerCamelCase=8 , lowerCamelCase=6 , lowerCamelCase=32 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="relu6" , lowerCamelCase=True , lowerCamelCase=0.8 , lowerCamelCase=0.02 , lowerCamelCase=0.001 , lowerCamelCase=255 , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) if depth_multiplier <= 0: raise ValueError("depth_multiplier must be greater than zero." ) __a = num_channels __a = image_size __a = depth_multiplier __a = depth_divisible_by __a = min_depth __a = expand_ratio __a = output_stride __a = first_layer_is_expansion __a = finegrained_output __a = hidden_act __a = tf_padding __a = classifier_dropout_prob __a = initializer_range __a = layer_norm_eps __a = semantic_loss_ignore_index class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = version.parse("""1.11""" ) @property def a__ ( self ): return OrderedDict([("pixel_values", {0: "batch"})] ) @property def a__ ( self ): if self.task == "image-classification": return OrderedDict([("logits", {0: "batch"})] ) else: return OrderedDict([("last_hidden_state", {0: "batch"}), ("pooler_output", {0: "batch"})] ) @property def a__ ( self ): return 1E-4
261
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class snake_case__ ( snake_case_, snake_case_, snake_case_, unittest.TestCase ): _snake_case : str = StableUnCLIPImgaImgPipeline _snake_case : Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS _snake_case : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _snake_case : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _snake_case : List[Any] = frozenset([] ) def a__ ( self ): __a = 32 __a = embedder_hidden_size # image encoding components __a = CLIPImageProcessor(crop_size=32 , size=32 ) torch.manual_seed(0 ) __a = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=lowerCamelCase , projection_dim=lowerCamelCase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) ) # regular denoising components torch.manual_seed(0 ) __a = StableUnCLIPImageNormalizer(embedding_dim=lowerCamelCase ) __a = DDPMScheduler(beta_schedule="squaredcos_cap_v2" ) torch.manual_seed(0 ) __a = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) __a = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=lowerCamelCase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) __a = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="projection" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=lowerCamelCase , layers_per_block=1 , upcast_attention=lowerCamelCase , use_linear_projection=lowerCamelCase , ) torch.manual_seed(0 ) __a = DDIMScheduler( beta_schedule="scaled_linear" , beta_start=0.0_0085 , beta_end=0.012 , prediction_type="v_prediction" , set_alpha_to_one=lowerCamelCase , steps_offset=1 , ) torch.manual_seed(0 ) __a = AutoencoderKL() __a = { # image encoding components "feature_extractor": feature_extractor, "image_encoder": image_encoder.eval(), # image noising components "image_normalizer": image_normalizer.eval(), "image_noising_scheduler": image_noising_scheduler, # regular denoising components "tokenizer": tokenizer, "text_encoder": text_encoder.eval(), "unet": unet.eval(), "scheduler": scheduler, "vae": vae.eval(), } return components def a__ ( self , lowerCamelCase , lowerCamelCase=0 , lowerCamelCase=True ): if str(lowerCamelCase ).startswith("mps" ): __a = torch.manual_seed(lowerCamelCase ) else: __a = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) __a = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) if pil_image: __a = input_image * 0.5 + 0.5 __a = input_image.clamp(0 , 1 ) __a = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() __a = DiffusionPipeline.numpy_to_pil(lowerCamelCase )[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def a__ ( self ): __a = "cpu" # ensure determinism for the device-dependent torch.Generator __a = self.get_dummy_components() __a = StableUnCLIPImgaImgPipeline(**lowerCamelCase ) __a = sd_pipe.to(lowerCamelCase ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase ) __a = self.get_dummy_inputs(lowerCamelCase ) inputs.update({"image_embeds": None} ) __a = sd_pipe(**lowerCamelCase ).images __a = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __a = np.array([0.3872, 0.7224, 0.5601, 0.4741, 0.6872, 0.5814, 0.4636, 0.3867, 0.5078] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def a__ ( self ): __a = torch_device in ["cpu", "mps"] self._test_attention_slicing_forward_pass(test_max_difference=lowerCamelCase ) def a__ ( self ): __a = torch_device in ["cpu", "mps"] self._test_inference_batch_single_identical(test_max_difference=lowerCamelCase ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def a__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_max_difference=lowerCamelCase ) @slow @require_torch_gpu class snake_case__ ( unittest.TestCase ): def a__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def a__ ( self ): __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) __a = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy" ) __a = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-l-img2img" , torch_dtype=torch.floataa ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __a = torch.Generator(device="cpu" ).manual_seed(0 ) __a = pipe(lowerCamelCase , "anime turle" , generator=lowerCamelCase , output_type="np" ) __a = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowerCamelCase , lowerCamelCase ) def a__ ( self ): __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) __a = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy" ) __a = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-h-img2img" , torch_dtype=torch.floataa ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __a = torch.Generator(device="cpu" ).manual_seed(0 ) __a = pipe(lowerCamelCase , "anime turle" , generator=lowerCamelCase , output_type="np" ) __a = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowerCamelCase , lowerCamelCase ) def a__ ( self ): __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __a = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-h-img2img" , torch_dtype=torch.floataa ) __a = pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __a = pipe( lowerCamelCase , "anime turtle" , num_inference_steps=2 , output_type="np" , ) __a = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
261
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_deformable_detr import DeformableDetrImageProcessor SCREAMING_SNAKE_CASE_: List[Any] =logging.get_logger(__name__) class __A ( UpperCamelCase__ ): def __init__(self : str , *__a : List[Any] , **__a : Optional[Any] ): warnings.warn( "The class DeformableDetrFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use DeformableDetrImageProcessor instead." , __a , ) super().__init__(*__a , **__a )
106
'''simple docstring''' from random import randint from tempfile import TemporaryFile import numpy as np def lowerCAmelCase_ ( snake_case_ : int , snake_case_ : Any , snake_case_ : int ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = 0 if start < end: UpperCAmelCase_ = randint(snake_case_ , snake_case_ ) UpperCAmelCase_ = a[end] UpperCAmelCase_ = a[pivot] UpperCAmelCase_ = temp UpperCAmelCase_ , UpperCAmelCase_ = _in_place_partition(snake_case_ , snake_case_ , snake_case_ ) count += _in_place_quick_sort(snake_case_ , snake_case_ , p - 1 ) count += _in_place_quick_sort(snake_case_ , p + 1 , snake_case_ ) return count def lowerCAmelCase_ ( snake_case_ : List[Any] , snake_case_ : Dict , snake_case_ : List[str] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = 0 UpperCAmelCase_ = randint(snake_case_ , snake_case_ ) UpperCAmelCase_ = a[end] UpperCAmelCase_ = a[pivot] UpperCAmelCase_ = temp UpperCAmelCase_ = start - 1 for index in range(snake_case_ , snake_case_ ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value UpperCAmelCase_ = new_pivot_index + 1 UpperCAmelCase_ = a[new_pivot_index] UpperCAmelCase_ = a[index] UpperCAmelCase_ = temp UpperCAmelCase_ = a[new_pivot_index + 1] UpperCAmelCase_ = a[end] UpperCAmelCase_ = temp return new_pivot_index + 1, count SCREAMING_SNAKE_CASE_: List[str] =TemporaryFile() SCREAMING_SNAKE_CASE_: int =1_00 # 1000 elements are to be sorted SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_: str =0, 1 # mean and standard deviation SCREAMING_SNAKE_CASE_: List[str] =np.random.normal(mu, sigma, p) np.save(outfile, X) print('The array is') print(X) outfile.seek(0) # using the same array SCREAMING_SNAKE_CASE_: str =np.load(outfile) SCREAMING_SNAKE_CASE_: List[Any] =len(M) - 1 SCREAMING_SNAKE_CASE_: Dict =_in_place_quick_sort(M, 0, r) print( 'No of Comparisons for 100 elements selected from a standard normal distribution' 'is :' ) print(z)
106
1
"""simple docstring""" import json import os import shutil import tempfile import unittest from transformers import BatchEncoding, CanineTokenizer from transformers.testing_utils import require_tokenizers, require_torch from transformers.tokenization_utils import AddedToken from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin class __a (a_ , unittest.TestCase): '''simple docstring''' _SCREAMING_SNAKE_CASE :str = CanineTokenizer _SCREAMING_SNAKE_CASE :Dict = False def _a ( self ) -> Tuple: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ : List[str] = CanineTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _a ( self ) -> Optional[int]: """simple docstring""" return CanineTokenizer.from_pretrained("""google/canine-s""" ) def _a ( self , **_a ) -> CanineTokenizer: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = self.tokenizer_class.from_pretrained(self.tmpdirname , **_a ) SCREAMING_SNAKE_CASE__ : List[Any] = 1_024 return tokenizer @require_torch def _a ( self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = self.canine_tokenizer SCREAMING_SNAKE_CASE__ : List[Any] = ["""Life is like a box of chocolates.""", """You never know what you're gonna get."""] # fmt: off SCREAMING_SNAKE_CASE__ : Union[str, Any] = [57_344, 76, 105, 102, 101, 32, 105, 115, 32, 108, 105, 107, 101, 32, 97, 32, 98, 111, 120, 32, 111, 102, 32, 99, 104, 111, 99, 111, 108, 97, 116, 101, 115, 46, 57_345, 0, 0, 0, 0] # fmt: on SCREAMING_SNAKE_CASE__ : Dict = tokenizer(_a , padding=_a , return_tensors="""pt""" ) self.assertIsInstance(_a , _a ) SCREAMING_SNAKE_CASE__ : List[Any] = list(batch.input_ids.numpy()[0] ) self.assertListEqual(_a , _a ) self.assertEqual((2, 39) , batch.input_ids.shape ) self.assertEqual((2, 39) , batch.attention_mask.shape ) @require_torch def _a ( self ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = self.canine_tokenizer SCREAMING_SNAKE_CASE__ : List[Any] = ["""Once there was a man.""", """He wrote a test in HuggingFace Tranformers."""] SCREAMING_SNAKE_CASE__ : int = tokenizer(_a , padding=_a , return_tensors="""pt""" ) # check if input_ids, attention_mask and token_type_ids are returned self.assertIn("""input_ids""" , _a ) self.assertIn("""attention_mask""" , _a ) self.assertIn("""token_type_ids""" , _a ) @require_torch def _a ( self ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = self.canine_tokenizer SCREAMING_SNAKE_CASE__ : Optional[int] = [ """What's the weater?""", """It's about 25 degrees.""", ] SCREAMING_SNAKE_CASE__ : str = tokenizer( text_target=_a , max_length=32 , padding="""max_length""" , truncation=_a , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def _a ( self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE__ : Optional[Any] = tempfile.mkdtemp() SCREAMING_SNAKE_CASE__ : Optional[Any] = """ He is very happy, UNwant\u00E9d,running""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = tokenizer.encode(_a , add_special_tokens=_a ) tokenizer.save_pretrained(_a ) SCREAMING_SNAKE_CASE__ : Tuple = tokenizer.__class__.from_pretrained(_a ) SCREAMING_SNAKE_CASE__ : Optional[Any] = after_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) shutil.rmtree(_a ) SCREAMING_SNAKE_CASE__ : List[str] = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE__ : int = tempfile.mkdtemp() SCREAMING_SNAKE_CASE__ : Any = """ He is very happy, UNwant\u00E9d,running""" SCREAMING_SNAKE_CASE__ : Dict = tokenizer.additional_special_tokens # We can add a new special token for Canine as follows: SCREAMING_SNAKE_CASE__ : Tuple = chr(0Xe007 ) additional_special_tokens.append(_a ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) SCREAMING_SNAKE_CASE__ : Tuple = tokenizer.encode(_a , add_special_tokens=_a ) tokenizer.save_pretrained(_a ) SCREAMING_SNAKE_CASE__ : int = tokenizer.__class__.from_pretrained(_a ) SCREAMING_SNAKE_CASE__ : int = after_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) self.assertIn(_a , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) SCREAMING_SNAKE_CASE__ : List[Any] = tokenizer.__class__.from_pretrained(_a , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(_a ) def _a ( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = self.get_tokenizers(do_lower_case=_a ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.get_clean_sequence(_a ) # a special token for Canine can be defined as follows: SCREAMING_SNAKE_CASE__ : Optional[int] = 0Xe005 SCREAMING_SNAKE_CASE__ : Union[str, Any] = chr(_a ) tokenizer.add_special_tokens({"""cls_token""": special_token} ) SCREAMING_SNAKE_CASE__ : str = tokenizer.encode(_a , add_special_tokens=_a ) self.assertEqual(len(_a ) , 1 ) SCREAMING_SNAKE_CASE__ : List[str] = tokenizer.decode(ids + encoded_special_token , clean_up_tokenization_spaces=_a ) SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer.encode(_a , add_special_tokens=_a ) SCREAMING_SNAKE_CASE__ : Optional[Any] = tokenizer.encode(_a , add_special_tokens=_a ) SCREAMING_SNAKE_CASE__ : Optional[Any] = tokenizer.encode(_a , add_special_tokens=_a ) self.assertEqual(_a , input_encoded + special_token_id ) SCREAMING_SNAKE_CASE__ : List[Any] = tokenizer.decode(_a , skip_special_tokens=_a ) self.assertTrue(special_token not in decoded ) def _a ( self ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = self.get_tokenizers(do_lower_case=_a ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ : Optional[int] = chr(0Xe005 ) SCREAMING_SNAKE_CASE__ : Tuple = chr(0Xe006 ) # `add_tokens` method stores special tokens only in `tokenizer.unique_no_split_tokens`. (in tokenization_utils.py) tokenizer.add_tokens([SPECIAL_TOKEN_1] , special_tokens=_a ) # `add_special_tokens` method stores special tokens in `tokenizer.additional_special_tokens`, # which also occur in `tokenizer.all_special_tokens`. (in tokenization_utils_base.py) tokenizer.add_special_tokens({"""additional_special_tokens""": [SPECIAL_TOKEN_2]} ) SCREAMING_SNAKE_CASE__ : Any = tokenizer.tokenize(_a ) SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer.tokenize(_a ) self.assertEqual(len(_a ) , 1 ) self.assertEqual(len(_a ) , 1 ) self.assertEqual(token_a[0] , _a ) self.assertEqual(token_a[0] , _a ) @require_tokenizers def _a ( self ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = self.get_tokenizers(do_lower_case=_a ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # a special token for Canine can be defined as follows: SCREAMING_SNAKE_CASE__ : Any = 0Xe006 SCREAMING_SNAKE_CASE__ : List[str] = chr(_a ) SCREAMING_SNAKE_CASE__ : Optional[int] = AddedToken(_a , lstrip=_a ) tokenizer.add_special_tokens({"""additional_special_tokens""": [new_token]} ) with tempfile.TemporaryDirectory() as tmp_dir_name: tokenizer.save_pretrained(_a ) tokenizer.from_pretrained(_a ) def _a ( self ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_a ) with open(os.path.join(_a , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: SCREAMING_SNAKE_CASE__ : str = json.load(_a ) with open(os.path.join(_a , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: SCREAMING_SNAKE_CASE__ : Tuple = json.load(_a ) # a special token for Canine can be defined as follows: SCREAMING_SNAKE_CASE__ : int = 0Xe006 SCREAMING_SNAKE_CASE__ : int = chr(_a ) SCREAMING_SNAKE_CASE__ : List[str] = [new_token_a] SCREAMING_SNAKE_CASE__ : List[str] = [new_token_a] with open(os.path.join(_a , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(_a , _a ) with open(os.path.join(_a , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(_a , _a ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files SCREAMING_SNAKE_CASE__ : Union[str, Any] = tokenizer_class.from_pretrained(_a , extra_ids=0 ) self.assertIn(_a , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids([new_token_a] ) ) , ) SCREAMING_SNAKE_CASE__ : List[Any] = 0Xe007 SCREAMING_SNAKE_CASE__ : str = chr(_a ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained SCREAMING_SNAKE_CASE__ : str = [AddedToken(_a , lstrip=_a )] SCREAMING_SNAKE_CASE__ : Dict = tokenizer_class.from_pretrained( _a , additional_special_tokens=_a , extra_ids=0 ) self.assertIn(_a , tokenizer.additional_special_tokens ) # self.assertIn(new_token_2,tokenizer.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer.convert_ids_to_tokens(tokenizer.convert_tokens_to_ids([new_token_a] ) ) ) @require_tokenizers def _a ( self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = self.get_tokenizers(do_lower_case=_a ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ : Optional[int] = """hello world""" if self.space_between_special_tokens: SCREAMING_SNAKE_CASE__ : Optional[int] = """[CLS] hello world [SEP]""" else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = input SCREAMING_SNAKE_CASE__ : Union[str, Any] = tokenizer.encode(_a , add_special_tokens=_a ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = tokenizer.decode(_a , spaces_between_special_tokens=self.space_between_special_tokens ) self.assertIn(_a , [output, output.lower()] ) def _a ( self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ : Optional[int] = [ """bos_token""", """eos_token""", """unk_token""", """sep_token""", """pad_token""", """cls_token""", """mask_token""", ] SCREAMING_SNAKE_CASE__ : Optional[int] = """a""" SCREAMING_SNAKE_CASE__ : Optional[int] = ord(_a ) for attr in attributes_list: setattr(_a , attr + """_id""" , _a ) self.assertEqual(getattr(_a , _a ) , _a ) self.assertEqual(getattr(_a , attr + """_id""" ) , _a ) setattr(_a , attr + """_id""" , _a ) self.assertEqual(getattr(_a , _a ) , _a ) self.assertEqual(getattr(_a , attr + """_id""" ) , _a ) setattr(_a , """additional_special_tokens_ids""" , [] ) self.assertListEqual(getattr(_a , """additional_special_tokens""" ) , [] ) self.assertListEqual(getattr(_a , """additional_special_tokens_ids""" ) , [] ) SCREAMING_SNAKE_CASE__ : List[str] = 0Xe006 SCREAMING_SNAKE_CASE__ : Any = chr(_a ) setattr(_a , """additional_special_tokens_ids""" , [additional_special_token_id] ) self.assertListEqual(getattr(_a , """additional_special_tokens""" ) , [additional_special_token] ) self.assertListEqual(getattr(_a , """additional_special_tokens_ids""" ) , [additional_special_token_id] ) def _a ( self ) -> str: """simple docstring""" pass def _a ( self ) -> Optional[Any]: """simple docstring""" pass def _a ( self ) -> Any: """simple docstring""" pass def _a ( self ) -> str: """simple docstring""" pass def _a ( self ) -> int: """simple docstring""" pass def _a ( self ) -> List[Any]: """simple docstring""" pass def _a ( self ) -> Tuple: """simple docstring""" pass def _a ( self ) -> Tuple: """simple docstring""" pass
132
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase : Any = { "configuration_altclip": [ "ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "AltCLIPConfig", "AltCLIPTextConfig", "AltCLIPVisionConfig", ], "processing_altclip": ["AltCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Tuple = [ "ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "AltCLIPPreTrainedModel", "AltCLIPModel", "AltCLIPTextModel", "AltCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
252
0
'''simple docstring''' import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.esm.modeling_esmfold import EsmForProteinFolding class UpperCamelCase_ : """simple docstring""" def __init__( self : Optional[Any] , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Tuple=13 , _lowerCamelCase : Optional[Any]=7 , _lowerCamelCase : List[str]=False , _lowerCamelCase : Dict=True , _lowerCamelCase : List[str]=False , _lowerCamelCase : List[Any]=False , _lowerCamelCase : str=19 , _lowerCamelCase : Dict=32 , _lowerCamelCase : str=5 , _lowerCamelCase : Optional[int]=4 , _lowerCamelCase : List[Any]=37 , _lowerCamelCase : List[str]="gelu" , _lowerCamelCase : Optional[int]=0.1 , _lowerCamelCase : List[Any]=0.1 , _lowerCamelCase : List[Any]=512 , _lowerCamelCase : List[str]=16 , _lowerCamelCase : Union[str, Any]=2 , _lowerCamelCase : Dict=0.02 , _lowerCamelCase : str=3 , _lowerCamelCase : Optional[int]=4 , _lowerCamelCase : str=None , ): """simple docstring""" A_ : List[str] = parent A_ : Tuple = batch_size A_ : Dict = seq_length A_ : List[Any] = is_training A_ : Optional[int] = use_input_mask A_ : int = use_token_type_ids A_ : Dict = use_labels A_ : List[str] = vocab_size A_ : Optional[Any] = hidden_size A_ : Optional[Any] = num_hidden_layers A_ : Optional[Any] = num_attention_heads A_ : Optional[Any] = intermediate_size A_ : int = hidden_act A_ : int = hidden_dropout_prob A_ : List[Any] = attention_probs_dropout_prob A_ : Any = max_position_embeddings A_ : Optional[Any] = type_vocab_size A_ : Optional[Any] = type_sequence_label_size A_ : Tuple = initializer_range A_ : Any = num_labels A_ : Optional[Any] = num_choices A_ : int = scope def _a ( self : List[Any] ): """simple docstring""" A_ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : int = None if self.use_input_mask: A_ : Dict = random_attention_mask([self.batch_size, self.seq_length] ) A_ : Any = None A_ : Tuple = None A_ : List[str] = None if self.use_labels: A_ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : Tuple = ids_tensor([self.batch_size] , self.num_choices ) A_ : List[Any] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self : Tuple ): """simple docstring""" A_ : Optional[Any] = EsmConfig( vocab_size=33 , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , is_folding_model=_lowerCamelCase , esmfold_config={'''trunk''': {'''num_blocks''': 2}, '''fp16_esm''': False} , ) return config def _a ( self : Optional[Any] , _lowerCamelCase : List[Any] , _lowerCamelCase : Dict , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[int] , _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any] ): """simple docstring""" A_ : List[str] = EsmForProteinFolding(config=_lowerCamelCase ).float() model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase , attention_mask=_lowerCamelCase ) A_ : Dict = model(_lowerCamelCase ) A_ : str = model(_lowerCamelCase ) self.parent.assertEqual(result.positions.shape , (8, self.batch_size, self.seq_length, 14, 3) ) self.parent.assertEqual(result.angles.shape , (8, self.batch_size, self.seq_length, 7, 2) ) def _a ( self : Optional[Any] ): """simple docstring""" A_ : List[str] = self.prepare_config_and_inputs() ( A_ ) : str = config_and_inputs A_ : Any = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ (a__, a__, unittest.TestCase ): """simple docstring""" _lowerCAmelCase : Optional[int] = False _lowerCAmelCase : Optional[int] = (EsmForProteinFolding,) if is_torch_available() else () _lowerCAmelCase : List[str] = () _lowerCAmelCase : Tuple = {} if is_torch_available() else {} _lowerCAmelCase : str = False def _a ( self : Tuple ): """simple docstring""" A_ : List[str] = EsmFoldModelTester(self ) A_ : Union[str, Any] = ConfigTester(self , config_class=_lowerCamelCase , hidden_size=37 ) def _a ( self : Optional[int] ): """simple docstring""" self.config_tester.run_common_tests() def _a ( self : int ): """simple docstring""" A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) @unittest.skip('''Does not support attention outputs''' ) def _a ( self : Tuple ): """simple docstring""" pass @unittest.skip def _a ( self : Dict ): """simple docstring""" pass @unittest.skip('''Esm does not support embedding resizing''' ) def _a ( self : int ): """simple docstring""" pass @unittest.skip('''Esm does not support embedding resizing''' ) def _a ( self : Any ): """simple docstring""" pass @unittest.skip('''ESMFold does not support passing input embeds!''' ) def _a ( self : str ): """simple docstring""" pass @unittest.skip('''ESMFold does not support head pruning.''' ) def _a ( self : Tuple ): """simple docstring""" pass @unittest.skip('''ESMFold does not support head pruning.''' ) def _a ( self : Union[str, Any] ): """simple docstring""" pass @unittest.skip('''ESMFold does not support head pruning.''' ) def _a ( self : List[str] ): """simple docstring""" pass @unittest.skip('''ESMFold does not support head pruning.''' ) def _a ( self : Optional[Any] ): """simple docstring""" pass @unittest.skip('''ESMFold does not support head pruning.''' ) def _a ( self : int ): """simple docstring""" pass @unittest.skip('''ESMFold does not output hidden states in the normal way.''' ) def _a ( self : Tuple ): """simple docstring""" pass @unittest.skip('''ESMfold does not output hidden states in the normal way.''' ) def _a ( self : List[Any] ): """simple docstring""" pass @unittest.skip('''ESMFold only has one output format.''' ) def _a ( self : Dict ): """simple docstring""" pass @unittest.skip('''This test doesn\'t work for ESMFold and doesn\'t test core functionality''' ) def _a ( self : Any ): """simple docstring""" pass @unittest.skip('''ESMFold does not support input chunking.''' ) def _a ( self : Any ): """simple docstring""" pass @unittest.skip('''ESMFold doesn\'t respect you and it certainly doesn\'t respect your initialization arguments.''' ) def _a ( self : List[Any] ): """simple docstring""" pass @unittest.skip('''ESMFold doesn\'t support torchscript compilation.''' ) def _a ( self : Any ): """simple docstring""" pass @unittest.skip('''ESMFold doesn\'t support torchscript compilation.''' ) def _a ( self : Optional[Any] ): """simple docstring""" pass @unittest.skip('''ESMFold doesn\'t support torchscript compilation.''' ) def _a ( self : List[str] ): """simple docstring""" pass @unittest.skip('''ESMFold doesn\'t support data parallel.''' ) def _a ( self : Union[str, Any] ): """simple docstring""" pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def _a ( self : Optional[int] ): """simple docstring""" pass @require_torch class UpperCamelCase_ (a__ ): """simple docstring""" @slow def _a ( self : Union[str, Any] ): """simple docstring""" A_ : Tuple = EsmForProteinFolding.from_pretrained('''facebook/esmfold_v1''' ).float() model.eval() A_ : List[Any] = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) A_ : str = model(_lowerCamelCase )['''positions'''] A_ : Any = torch.tensor([2.58_28, 0.79_93, -10.93_34] , dtype=torch.floataa ) self.assertTrue(torch.allclose(position_outputs[0, 0, 0, 0] , _lowerCamelCase , atol=1E-4 ) )
354
'''simple docstring''' from __future__ import annotations def snake_case__ ( lowerCamelCase__ : list[int] , lowerCamelCase__ : int ) -> list[int]: A_ : int = 0 A_ : str = len(lowerCamelCase__ ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: A_ : Tuple = i + 1 else: A_ : List[str] = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(F'{two_pointer([2, 7, 11, 15], 9) = }')
4
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} # See all MVP models at https://huggingface.co/models?filter=mvp _UpperCAmelCase = { """vocab_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json""", }, """added_tokens.json""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json""", }, """merges_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt""", }, """tokenizer_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json""", }, } _UpperCAmelCase = { """RUCAIBox/mvp""": 1024, } class UpperCAmelCase ( __A ): '''simple docstring''' lowerCamelCase_ = VOCAB_FILES_NAMES lowerCamelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ = ['''input_ids''', '''attention_mask'''] lowerCamelCase_ = MvpTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=None , lowercase="replace" , lowercase="<s>" , lowercase="</s>" , lowercase="</s>" , lowercase="<s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase="<mask>" , lowercase=False , lowercase=True , **lowercase , ): """simple docstring""" super().__init__( lowercase , lowercase , tokenizer_file=lowercase , errors=lowercase , bos_token=lowercase , eos_token=lowercase , sep_token=lowercase , cls_token=lowercase , unk_token=lowercase , pad_token=lowercase , mask_token=lowercase , add_prefix_space=lowercase , trim_offsets=lowercase , **lowercase , ) A_ : str = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , lowercase ) != add_prefix_space: A_ : int = getattr(lowercase , pre_tok_state.pop('type' ) ) A_ : Union[str, Any] = add_prefix_space A_ : Dict = pre_tok_class(**lowercase ) A_ : Union[str, Any] = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` A_ : Any = 'post_processor' A_ : List[str] = getattr(self.backend_tokenizer , lowercase , lowercase ) if tokenizer_component_instance: A_ : List[str] = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: A_ : int = tuple(state['sep'] ) if "cls" in state: A_ : Optional[int] = tuple(state['cls'] ) A_ : Tuple = False if state.get('add_prefix_space' , lowercase ) != add_prefix_space: A_ : Union[str, Any] = add_prefix_space A_ : Tuple = True if state.get('trim_offsets' , lowercase ) != trim_offsets: A_ : str = trim_offsets A_ : str = True if changes_to_apply: A_ : List[str] = getattr(lowercase , state.pop('type' ) ) A_ : List[str] = component_class(**lowercase ) setattr(self.backend_tokenizer , lowercase , lowercase ) @property def lowerCAmelCase_ ( self ): """simple docstring""" if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def lowerCAmelCase_ ( self , lowercase ): """simple docstring""" A_ : Optional[Any] = AddedToken(lowercase , lstrip=lowercase , rstrip=lowercase ) if isinstance(lowercase , lowercase ) else value A_ : Dict = value def lowerCAmelCase_ ( self , *lowercase , **lowercase ): """simple docstring""" A_ : Any = kwargs.get('is_split_into_words' , lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' 'to use it with pretokenized inputs.' ) return super()._batch_encode_plus(*lowercase , **lowercase ) def lowerCAmelCase_ ( self , *lowercase , **lowercase ): """simple docstring""" A_ : Dict = kwargs.get('is_split_into_words' , lowercase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' 'to use it with pretokenized inputs.' ) return super()._encode_plus(*lowercase , **lowercase ) def lowerCAmelCase_ ( self , lowercase , lowercase = None ): """simple docstring""" A_ : Any = self._tokenizer.model.save(lowercase , name=lowercase ) return tuple(lowercase ) def lowerCAmelCase_ ( self , lowercase , lowercase=None ): """simple docstring""" A_ : Any = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def lowerCAmelCase_ ( self , lowercase , lowercase = None ): """simple docstring""" A_ : Union[str, Any] = [self.sep_token_id] A_ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
140
import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class UpperCAmelCase ( __A ): '''simple docstring''' lowerCamelCase_ = '''char''' lowerCamelCase_ = '''bpe''' lowerCamelCase_ = '''wp''' _UpperCAmelCase = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class UpperCAmelCase ( __A ): '''simple docstring''' lowerCamelCase_ = ['''image_processor''', '''char_tokenizer'''] lowerCamelCase_ = '''ViTImageProcessor''' lowerCamelCase_ = '''MgpstrTokenizer''' def __init__( self , lowercase=None , lowercase=None , **lowercase ): """simple docstring""" A_ : str = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , lowercase , ) A_ : Optional[int] = kwargs.pop('feature_extractor' ) A_ : List[str] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) A_ : Union[str, Any] = tokenizer A_ : List[Any] = AutoTokenizer.from_pretrained('gpt2' ) A_ : Any = AutoTokenizer.from_pretrained('bert-base-uncased' ) super().__init__(lowercase , lowercase ) def __call__( self , lowercase=None , lowercase=None , lowercase=None , **lowercase ): """simple docstring""" if images is None and text is None: raise ValueError('You need to specify either an `images` or `text` input to process.' ) if images is not None: A_ : List[Any] = self.image_processor(lowercase , return_tensors=lowercase , **lowercase ) if text is not None: A_ : Union[str, Any] = self.char_tokenizer(lowercase , return_tensors=lowercase , **lowercase ) if text is None: return inputs elif images is None: return encodings else: A_ : Optional[int] = encodings['input_ids'] return inputs def lowerCAmelCase_ ( self , lowercase ): """simple docstring""" A_ , A_ , A_ : int = sequences A_ : Union[str, Any] = char_preds.size(0 ) A_ , A_ : Union[str, Any] = self._decode_helper(lowercase , 'char' ) A_ , A_ : List[str] = self._decode_helper(lowercase , 'bpe' ) A_ , A_ : Optional[int] = self._decode_helper(lowercase , 'wp' ) A_ : Dict = [] A_ : Optional[int] = [] for i in range(lowercase ): A_ : List[str] = [char_scores[i], bpe_scores[i], wp_scores[i]] A_ : int = [char_strs[i], bpe_strs[i], wp_strs[i]] A_ : Union[str, Any] = scores.index(max(lowercase ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) A_ : Dict = {} A_ : str = final_strs A_ : Union[str, Any] = final_scores A_ : Optional[Any] = char_strs A_ : Dict = bpe_strs A_ : str = wp_strs return out def lowerCAmelCase_ ( self , lowercase , lowercase ): """simple docstring""" if format == DecodeType.CHARACTER: A_ : List[Any] = self.char_decode A_ : List[Any] = 1 A_ : str = '[s]' elif format == DecodeType.BPE: A_ : List[Any] = self.bpe_decode A_ : Optional[int] = 2 A_ : Tuple = '#' elif format == DecodeType.WORDPIECE: A_ : Optional[int] = self.wp_decode A_ : Optional[int] = 1_0_2 A_ : List[Any] = '[SEP]' else: raise ValueError(F'''Format {format} is not supported.''' ) A_ , A_ : Dict = [], [] A_ : Any = pred_logits.size(0 ) A_ : Optional[int] = pred_logits.size(1 ) A_ , A_ : int = pred_logits.topk(1 , dim=-1 , largest=lowercase , sorted=lowercase ) A_ : Dict = preds_index.view(-1 , lowercase )[:, 1:] A_ : Any = decoder(lowercase ) A_ , A_ : List[Any] = torch.nn.functional.softmax(lowercase , dim=2 ).max(dim=2 ) A_ : List[str] = preds_max_prob[:, 1:] for index in range(lowercase ): A_ : int = preds_str[index].find(lowercase ) A_ : Union[str, Any] = preds_str[index][:pred_eos] A_ : Dict = preds_index[index].cpu().tolist() A_ : List[str] = pred_index.index(lowercase ) if eos_token in pred_index else -1 A_ : List[str] = preds_max_prob[index][: pred_eos_index + 1] A_ : Union[str, Any] = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(lowercase ) conf_scores.append(lowercase ) return dec_strs, conf_scores def lowerCAmelCase_ ( self , lowercase ): """simple docstring""" A_ : int = [seq.replace(' ' , '' ) for seq in self.char_tokenizer.batch_decode(lowercase )] return decode_strs def lowerCAmelCase_ ( self , lowercase ): """simple docstring""" return self.bpe_tokenizer.batch_decode(lowercase ) def lowerCAmelCase_ ( self , lowercase ): """simple docstring""" A_ : Dict = [seq.replace(' ' , '' ) for seq in self.wp_tokenizer.batch_decode(lowercase )] return decode_strs
140
1
import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() __A =logging.get_logger(__name__) __A ={ '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } __A =[ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): for attribute in key.split("." ): lowerCamelCase_ = getattr(_lowerCamelCase , _lowerCamelCase ) if weight_type is not None: lowerCamelCase_ = getattr(_lowerCamelCase , _lowerCamelCase ).shape else: lowerCamelCase_ = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": lowerCamelCase_ = value elif weight_type == "weight_g": lowerCamelCase_ = value elif weight_type == "weight_v": lowerCamelCase_ = value elif weight_type == "bias": lowerCamelCase_ = value else: lowerCamelCase_ = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = [] lowerCamelCase_ = fairseq_model.state_dict() lowerCamelCase_ = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight lowerCamelCase_ = None for name, value in fairseq_dict.items(): lowerCamelCase_ = False if "conv_layers" in name: load_conv_layer( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , hf_model.config.feat_extract_norm == "group" , ) lowerCamelCase_ = True elif name.split("." )[0] == "proj": lowerCamelCase_ = fairseq_model.proj lowerCamelCase_ = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: lowerCamelCase_ = True if "*" in mapped_key: lowerCamelCase_ = name.split(_lowerCamelCase )[0].split("." )[-2] lowerCamelCase_ = mapped_key.replace("*" , _lowerCamelCase ) if "weight_g" in name: lowerCamelCase_ = "weight_g" elif "weight_v" in name: lowerCamelCase_ = "weight_v" elif "bias" in name: lowerCamelCase_ = "bias" elif "weight" in name: lowerCamelCase_ = "weight" else: lowerCamelCase_ = None set_recursively(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) continue if not is_used: unused_weights.append(_lowerCamelCase ) logger.warning(F'Unused weights: {unused_weights}' ) return proj_weight def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = full_name.split("conv_layers." )[-1] lowerCamelCase_ = name.split("." ) lowerCamelCase_ = int(items[0] ) lowerCamelCase_ = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) lowerCamelCase_ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) lowerCamelCase_ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) lowerCamelCase_ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) lowerCamelCase_ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(_lowerCamelCase ) def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = emb.weight.shape lowerCamelCase_ = nn.Linear(_lowerCamelCase , _lowerCamelCase , bias=_lowerCamelCase ) lowerCamelCase_ = emb.weight.data return lin_layer def lowerCamelCase_ ( lowerCamelCase__ ): with open(_lowerCamelCase , "r" , encoding="utf-8" ) as f: lowerCamelCase_ = f.readlines() lowerCamelCase_ = [line.split(" " )[0] for line in lines] lowerCamelCase_ = len(_lowerCamelCase ) lowerCamelCase_ = { "<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3, } vocab_dict.update(dict(zip(_lowerCamelCase , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): lowerCamelCase_ = WavaVecaConfig.from_pretrained(_lowerCamelCase ) lowerCamelCase_ = SpeechaTextaConfig.from_pretrained( _lowerCamelCase , vocab_size=_lowerCamelCase , decoder_layers=_lowerCamelCase , do_stable_layer_norm=_lowerCamelCase ) lowerCamelCase_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=_lowerCamelCase , return_attention_mask=_lowerCamelCase , ) lowerCamelCase_ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] )} ) lowerCamelCase_ = model[0].eval() # set weights for wav2vec2 encoder lowerCamelCase_ = WavaVecaModel(_lowerCamelCase ) lowerCamelCase_ = recursively_load_weights_wavaveca(model.encoder , _lowerCamelCase ) lowerCamelCase_ = SpeechaTextaForCausalLM(_lowerCamelCase ) lowerCamelCase_ = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=_lowerCamelCase ) # set output linear layer unexpected_keys.remove("embed_out" ) lowerCamelCase_ = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F'The following keys are missing when loading the decoder weights: {missing_keys}' ) logger.warning(F'The following keys are unexpected when loading the decoder weights: {unexpected_keys}' ) lowerCamelCase_ = SpeechEncoderDecoderModel(encoder=_lowerCamelCase , decoder=_lowerCamelCase ) lowerCamelCase_ = False # add projection layer lowerCamelCase_ = nn.Parameter(projection_layer.weight ) lowerCamelCase_ = nn.Parameter(projection_layer.bias ) lowerCamelCase_ = create_vocab_dict(_lowerCamelCase ) with open(os.path.join(_lowerCamelCase , "vocab.json" ) , "w" ) as fp: json.dump(_lowerCamelCase , _lowerCamelCase ) lowerCamelCase_ = SpeechaTextaTokenizer(os.path.join(_lowerCamelCase , "vocab.json" ) ) tokenizer.save_pretrained(_lowerCamelCase ) lowerCamelCase_ = hf_wavavec.config.to_dict() lowerCamelCase_ = tokenizer.pad_token_id lowerCamelCase_ = tokenizer.bos_token_id lowerCamelCase_ = tokenizer.eos_token_id lowerCamelCase_ = "speech_to_text_2" lowerCamelCase_ = "wav2vec2" lowerCamelCase_ = SpeechEncoderDecoderConfig.from_dict(_lowerCamelCase ) hf_wavavec.save_pretrained(_lowerCamelCase ) feature_extractor.save_pretrained(_lowerCamelCase ) if __name__ == "__main__": __A =argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-large-lv60''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/s2t-small-mustc-en-fr-st''', type=str, help='''Path to hf decoder s2t checkpoint config''', ) parser.add_argument('''--vocab_size''', default=1_0_2_2_4, type=int, help='''Vocab size of decoder''') parser.add_argument('''--num_decoder_layers''', default=7, type=int, help='''Number of decoder layers''') __A =parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
369
import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class _SCREAMING_SNAKE_CASE : def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=False , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=16 , lowercase=2 , lowercase=0.0_2 , lowercase=3 , lowercase=4 , lowercase=None , ) -> Optional[int]: lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = seq_length lowerCamelCase_ = is_training lowerCamelCase_ = use_input_mask lowerCamelCase_ = use_token_type_ids lowerCamelCase_ = use_labels lowerCamelCase_ = vocab_size lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = max_position_embeddings lowerCamelCase_ = type_vocab_size lowerCamelCase_ = type_sequence_label_size lowerCamelCase_ = initializer_range lowerCamelCase_ = num_labels lowerCamelCase_ = num_choices lowerCamelCase_ = scope def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase_ = None if self.use_input_mask: lowerCamelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase_ = None if self.use_token_type_ids: lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase_ = None lowerCamelCase_ = None lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase_ = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE_( self ) -> Tuple: return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> Union[str, Any]: lowerCamelCase_ = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase ) lowerCamelCase_ = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> Any: lowerCamelCase_ = BioGptForCausalLM(config=lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ) -> Tuple: lowerCamelCase_ = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() # create attention mask lowerCamelCase_ = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) lowerCamelCase_ = self.seq_length // 2 lowerCamelCase_ = 0 # first forward pass lowerCamelCase_ , lowerCamelCase_ = model(lowercase , attention_mask=lowercase ).to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase_ = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids lowerCamelCase_ = ids_tensor((1,) , lowercase ).item() + 1 lowerCamelCase_ = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) lowerCamelCase_ = random_other_next_tokens # append to next input_ids and attn_mask lowerCamelCase_ = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCamelCase_ = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=lowercase )] , dim=1 , ) # get two different outputs lowerCamelCase_ = model(lowercase , attention_mask=lowercase )["last_hidden_state"] lowerCamelCase_ = model(lowercase , past_key_values=lowercase , attention_mask=lowercase )["last_hidden_state"] # select random slice lowerCamelCase_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCamelCase_ = output_from_no_past[:, -1, random_slice_idx].detach() lowerCamelCase_ = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1e-3 ) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ) -> Dict: lowerCamelCase_ = BioGptModel(config=lowercase ).to(lowercase ).eval() lowerCamelCase_ = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) # first forward pass lowerCamelCase_ = model(lowercase , attention_mask=lowercase , use_cache=lowercase ) lowerCamelCase_ , lowerCamelCase_ = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids lowerCamelCase_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCamelCase_ = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and lowerCamelCase_ = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCamelCase_ = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) lowerCamelCase_ = model(lowercase , attention_mask=lowercase )["last_hidden_state"] lowerCamelCase_ = model(lowercase , attention_mask=lowercase , past_key_values=lowercase )[ "last_hidden_state" ] # select random slice lowerCamelCase_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCamelCase_ = output_from_no_past[:, -3:, random_slice_idx].detach() lowerCamelCase_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1e-3 ) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase , lowercase=False ) -> Optional[int]: lowerCamelCase_ = BioGptForCausalLM(lowercase ) model.to(lowercase ) if gradient_checkpointing: model.gradient_checkpointing_enable() lowerCamelCase_ = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def SCREAMING_SNAKE_CASE_( self , lowercase , *lowercase ) -> List[Any]: lowerCamelCase_ = BioGptModel(lowercase ) lowerCamelCase_ = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.0_0_1 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.0_1 ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ) -> Any: lowerCamelCase_ = self.num_labels lowerCamelCase_ = BioGptForTokenClassification(lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = self.prepare_config_and_inputs() ( ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ) = config_and_inputs lowerCamelCase_ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): lowerCAmelCase__ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) lowerCAmelCase__ = (BioGptForCausalLM,) if is_torch_available() else () lowerCAmelCase__ = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase__ = False def SCREAMING_SNAKE_CASE_( self ) -> Any: lowerCamelCase_ = BioGptModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=lowercase , hidden_size=37 ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCamelCase_ = type self.model_tester.create_and_check_model(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*lowercase , gradient_checkpointing=lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> int: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*lowercase ) @slow def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ = BioGptForCausalLM.from_pretrained("microsoft/biogpt" ) model.to(lowercase ) lowerCamelCase_ = BioGptTokenizer.from_pretrained("microsoft/biogpt" ) lowerCamelCase_ = "left" # Define PAD Token = EOS Token = 50256 lowerCamelCase_ = tokenizer.eos_token lowerCamelCase_ = model.config.eos_token_id # use different length sentences to test batching lowerCamelCase_ = [ "Hello, my dog is a little", "Today, I", ] lowerCamelCase_ = tokenizer(lowercase , return_tensors="pt" , padding=lowercase ) lowerCamelCase_ = inputs["input_ids"].to(lowercase ) lowerCamelCase_ = model.generate( input_ids=lowercase , attention_mask=inputs["attention_mask"].to(lowercase ) , ) lowerCamelCase_ = tokenizer(sentences[0] , return_tensors="pt" ).input_ids.to(lowercase ) lowerCamelCase_ = model.generate(input_ids=lowercase ) lowerCamelCase_ = inputs_non_padded.shape[-1] - inputs["attention_mask"][-1].long().sum().cpu().item() lowerCamelCase_ = tokenizer(sentences[1] , return_tensors="pt" ).input_ids.to(lowercase ) lowerCamelCase_ = model.generate(input_ids=lowercase , max_length=model.config.max_length - num_paddings ) lowerCamelCase_ = tokenizer.batch_decode(lowercase , skip_special_tokens=lowercase ) lowerCamelCase_ = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowercase ) lowerCamelCase_ = tokenizer.decode(output_padded[0] , skip_special_tokens=lowercase ) lowerCamelCase_ = [ "Hello, my dog is a little bit bigger than a little bit.", "Today, I have a good idea of how to use the information", ] self.assertListEqual(lowercase , lowercase ) self.assertListEqual(lowercase , [non_padded_sentence, padded_sentence] ) @slow def SCREAMING_SNAKE_CASE_( self ) -> int: for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ = BioGptModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Tuple: lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase_ = 3 lowerCamelCase_ = input_dict["input_ids"] lowerCamelCase_ = input_ids.ne(1 ).to(lowercase ) lowerCamelCase_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowerCamelCase_ = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase_ = 3 lowerCamelCase_ = "multi_label_classification" lowerCamelCase_ = input_dict["input_ids"] lowerCamelCase_ = input_ids.ne(1 ).to(lowercase ) lowerCamelCase_ = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) lowerCamelCase_ = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = BioGptForCausalLM.from_pretrained("microsoft/biogpt" ) lowerCamelCase_ = torch.tensor([[2, 4805, 9, 656, 21]] ) lowerCamelCase_ = model(lowercase )[0] lowerCamelCase_ = 42384 lowerCamelCase_ = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , lowercase ) lowerCamelCase_ = torch.tensor( [[[-9.5_2_3_6, -9.8_9_1_8, 1_0.4_5_5_7], [-1_1.0_4_6_9, -9.6_4_2_3, 8.1_0_2_2], [-8.8_6_6_4, -7.8_8_2_6, 5.5_3_2_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase , atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = BioGptTokenizer.from_pretrained("microsoft/biogpt" ) lowerCamelCase_ = BioGptForCausalLM.from_pretrained("microsoft/biogpt" ) model.to(lowercase ) torch.manual_seed(0 ) lowerCamelCase_ = tokenizer("COVID-19 is" , return_tensors="pt" ).to(lowercase ) lowerCamelCase_ = model.generate( **lowercase , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=lowercase , ) lowerCamelCase_ = tokenizer.decode(output_ids[0] , skip_special_tokens=lowercase ) lowerCamelCase_ = ( "COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the" " causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and" " territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK)," " and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and" " more than 800,000 deaths." ) self.assertEqual(lowercase , lowercase )
47
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase = {'''configuration_mbart''': ['''MBART_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MBartConfig''', '''MBartOnnxConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['''MBartTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['''MBartTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ '''MBART_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MBartForCausalLM''', '''MBartForConditionalGeneration''', '''MBartForQuestionAnswering''', '''MBartForSequenceClassification''', '''MBartModel''', '''MBartPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ '''TFMBartForConditionalGeneration''', '''TFMBartModel''', '''TFMBartPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ '''FlaxMBartForConditionalGeneration''', '''FlaxMBartForQuestionAnswering''', '''FlaxMBartForSequenceClassification''', '''FlaxMBartModel''', '''FlaxMBartPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mbart import MBART_PRETRAINED_CONFIG_ARCHIVE_MAP, MBartConfig, MBartOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart import MBartTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart_fast import MBartTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mbart import ( MBART_PRETRAINED_MODEL_ARCHIVE_LIST, MBartForCausalLM, MBartForConditionalGeneration, MBartForQuestionAnswering, MBartForSequenceClassification, MBartModel, MBartPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mbart import TFMBartForConditionalGeneration, TFMBartModel, TFMBartPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mbart import ( FlaxMBartForConditionalGeneration, FlaxMBartForQuestionAnswering, FlaxMBartForSequenceClassification, FlaxMBartModel, FlaxMBartPreTrainedModel, ) else: import sys lowerCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
188
from collections import UserDict from typing import Union import numpy as np import requests from ..utils import ( add_end_docstrings, logging, ) from .audio_classification import ffmpeg_read from .base import PIPELINE_INIT_ARGS, Pipeline lowerCamelCase = logging.get_logger(__name__) @add_end_docstrings(lowerCamelCase__ ) class __magic_name__ ( lowerCamelCase__ ): '''simple docstring''' def __init__( self, **lowercase_ ) -> Dict: """simple docstring""" super().__init__(**lowercase_ ) if self.framework != "pt": raise ValueError(F"""The {self.__class__} is only available in PyTorch.""" ) # No specific FOR_XXX available yet def __call__( self, lowercase_, **lowercase_ ) -> Tuple: """simple docstring""" return super().__call__(lowercase_, **lowercase_ ) def _UpperCAmelCase ( self, **lowercase_ ) -> int: """simple docstring""" a__ ={} if "candidate_labels" in kwargs: a__ =kwargs['''candidate_labels'''] if "hypothesis_template" in kwargs: a__ =kwargs['''hypothesis_template'''] return preprocess_params, {}, {} def _UpperCAmelCase ( self, lowercase_, lowercase_=None, lowercase_="This is a sound of {}." ) -> Union[str, Any]: """simple docstring""" if isinstance(lowercase_, lowercase_ ): if audio.startswith('''http://''' ) or audio.startswith('''https://''' ): # We need to actually check for a real protocol, otherwise it's impossible to use a local file # like http_huggingface_co.png a__ =requests.get(lowercase_ ).content else: with open(lowercase_, '''rb''' ) as f: a__ =f.read() if isinstance(lowercase_, lowercase_ ): a__ =ffmpeg_read(lowercase_, self.feature_extractor.sampling_rate ) if not isinstance(lowercase_, np.ndarray ): raise ValueError('''We expect a numpy ndarray as input''' ) if len(audio.shape ) != 1: raise ValueError('''We expect a single channel audio input for ZeroShotAudioClassificationPipeline''' ) a__ =self.feature_extractor( [audio], sampling_rate=self.feature_extractor.sampling_rate, return_tensors='''pt''' ) a__ =candidate_labels a__ =[hypothesis_template.format(lowercase_ ) for x in candidate_labels] a__ =self.tokenizer(lowercase_, return_tensors=self.framework, padding=lowercase_ ) a__ =[text_inputs] return inputs def _UpperCAmelCase ( self, lowercase_ ) -> str: """simple docstring""" a__ =model_inputs.pop('''candidate_labels''' ) a__ =model_inputs.pop('''text_inputs''' ) if isinstance(text_inputs[0], lowercase_ ): a__ =text_inputs[0] else: # Batching case. a__ =text_inputs[0][0] a__ =self.model(**lowercase_, **lowercase_ ) a__ ={ '''candidate_labels''': candidate_labels, '''logits''': outputs.logits_per_audio, } return model_outputs def _UpperCAmelCase ( self, lowercase_ ) -> Any: """simple docstring""" a__ =model_outputs.pop('''candidate_labels''' ) a__ =model_outputs['''logits'''][0] if self.framework == "pt": a__ =logits.softmax(dim=0 ) a__ =probs.tolist() else: raise ValueError('''`tf` framework not supported.''' ) a__ =[ {'''score''': score, '''label''': candidate_label} for score, candidate_label in sorted(zip(lowercase_, lowercase_ ), key=lambda lowercase_ : -x[0] ) ] return result
188
1
from ..utils import DummyObject, requires_backends class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : str , *__lowerCamelCase : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> Optional[Any]: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Optional[int] , *__lowerCamelCase : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> Dict: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Any , *__lowerCamelCase : int , **__lowerCamelCase : Optional[int] ) -> int: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : Any , *__lowerCamelCase : Optional[int] , **__lowerCamelCase : Dict ) -> Tuple: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Union[str, Any] , *__lowerCamelCase : Union[str, Any] , **__lowerCamelCase : Optional[Any] ) -> int: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Optional[int] , *__lowerCamelCase : Optional[Any] , **__lowerCamelCase : str ) -> List[str]: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : Optional[Any] , *__lowerCamelCase : Union[str, Any] , **__lowerCamelCase : Tuple ) -> Union[str, Any]: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Optional[int] , *__lowerCamelCase : int , **__lowerCamelCase : Union[str, Any] ) -> Any: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Any , *__lowerCamelCase : List[str] , **__lowerCamelCase : Optional[Any] ) -> Tuple: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : List[str] , *__lowerCamelCase : List[str] , **__lowerCamelCase : List[Any] ) -> Tuple: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : List[Any] , *__lowerCamelCase : List[Any] , **__lowerCamelCase : Dict ) -> Union[str, Any]: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Union[str, Any] , *__lowerCamelCase : str , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : Union[str, Any] , *__lowerCamelCase : str , **__lowerCamelCase : List[str] ) -> Tuple: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Tuple , *__lowerCamelCase : Any , **__lowerCamelCase : str ) -> Dict: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Optional[Any] , *__lowerCamelCase : List[str] , **__lowerCamelCase : Optional[Any] ) -> int: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : List[str] , *__lowerCamelCase : str , **__lowerCamelCase : Optional[Any] ) -> Any: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Union[str, Any] , *__lowerCamelCase : Any , **__lowerCamelCase : Union[str, Any] ) -> Union[str, Any]: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Dict , *__lowerCamelCase : Optional[Any] , **__lowerCamelCase : int ) -> Optional[Any]: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : List[Any] , *__lowerCamelCase : Dict , **__lowerCamelCase : Union[str, Any] ) -> int: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Optional[Any] , *__lowerCamelCase : int , **__lowerCamelCase : Dict ) -> Optional[int]: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Tuple , *__lowerCamelCase : Tuple , **__lowerCamelCase : Tuple ) -> Any: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : Optional[int] , *__lowerCamelCase : Any , **__lowerCamelCase : List[str] ) -> List[Any]: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : List[str] , *__lowerCamelCase : Tuple , **__lowerCamelCase : Optional[int] ) -> List[str]: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Tuple , *__lowerCamelCase : int , **__lowerCamelCase : int ) -> Dict: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : Dict , *__lowerCamelCase : List[Any] , **__lowerCamelCase : List[str] ) -> Any: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : int , *__lowerCamelCase : Dict , **__lowerCamelCase : Any ) -> Union[str, Any]: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : List[Any] , *__lowerCamelCase : Any , **__lowerCamelCase : str ) -> Union[str, Any]: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : List[str] , *__lowerCamelCase : Tuple , **__lowerCamelCase : Optional[int] ) -> Optional[Any]: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : int , *__lowerCamelCase : List[str] , **__lowerCamelCase : str ) -> Dict: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : List[str] , *__lowerCamelCase : str , **__lowerCamelCase : Tuple ) -> Optional[Any]: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : str , *__lowerCamelCase : Optional[int] , **__lowerCamelCase : Tuple ) -> Optional[Any]: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Dict , *__lowerCamelCase : Tuple , **__lowerCamelCase : Union[str, Any] ) -> Tuple: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Union[str, Any] , *__lowerCamelCase : List[str] , **__lowerCamelCase : Optional[int] ) -> Optional[int]: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : Union[str, Any] , *__lowerCamelCase : str , **__lowerCamelCase : str ) -> Union[str, Any]: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Any , *__lowerCamelCase : List[str] , **__lowerCamelCase : str ) -> Union[str, Any]: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : List[str] , *__lowerCamelCase : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> Optional[Any]: requires_backends(cls , ["flax"] ) class lowerCamelCase_ ( metaclass=_A ): '''simple docstring''' a__ = ["flax"] def __init__( self : Union[str, Any] , *__lowerCamelCase : Any , **__lowerCamelCase : int ) -> str: requires_backends(self , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : int , *__lowerCamelCase : Union[str, Any] , **__lowerCamelCase : Dict ) -> List[str]: requires_backends(cls , ["flax"] ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Any , *__lowerCamelCase : Dict , **__lowerCamelCase : Any ) -> Dict: requires_backends(cls , ["flax"] )
256
from collections.abc import Generator from math import sin def UpperCAmelCase ( _lowerCamelCase ): if len(_lowerCamelCase ) != 32: raise ValueError("Input must be of length 32" ) A : Any = B"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCAmelCase ( _lowerCamelCase ): if i < 0: raise ValueError("Input must be non-negative" ) A : List[Any] = format(_lowerCamelCase , "08x" )[-8:] A : List[str] = B"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8" ) return little_endian_hex def UpperCAmelCase ( _lowerCamelCase ): A : Optional[Any] = B"" for char in message: bit_string += format(_lowerCamelCase , "08b" ).encode("utf-8" ) A : int = format(len(_lowerCamelCase ) , "064b" ).encode("utf-8" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(_lowerCamelCase ) % 512 != 448: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def UpperCAmelCase ( _lowerCamelCase ): if len(_lowerCamelCase ) % 512 != 0: raise ValueError("Input must have length that's a multiple of 512" ) for pos in range(0 , len(_lowerCamelCase ) , 512 ): A : Optional[int] = bit_string[pos : pos + 512] A : List[str] = [] for i in range(0 , 512 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def UpperCAmelCase ( _lowerCamelCase ): if i < 0: raise ValueError("Input must be non-negative" ) A : Union[str, Any] = format(_lowerCamelCase , "032b" ) A : List[str] = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(_lowerCamelCase , 2 ) def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): return (a + b) % 2**32 def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): if i < 0: raise ValueError("Input must be non-negative" ) if shift < 0: raise ValueError("Shift must be non-negative" ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCAmelCase ( _lowerCamelCase ): A : Union[str, Any] = preprocess(_lowerCamelCase ) A : Any = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states A : Optional[int] = 0X67452301 A : Any = 0Xefcdab89 A : Tuple = 0X98badcfe A : Union[str, Any] = 0X10325476 A : Optional[Any] = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(_lowerCamelCase ): A : Optional[Any] = aa A : Optional[Any] = ba A : List[Any] = ca A : Optional[int] = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f A : Dict = d ^ (b & (c ^ d)) A : Optional[Any] = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f A : Optional[int] = c ^ (d & (b ^ c)) A : List[Any] = (5 * i + 1) % 16 elif i <= 47: A : Tuple = b ^ c ^ d A : str = (3 * i + 5) % 16 else: A : Union[str, Any] = c ^ (b | not_aa(_lowerCamelCase )) A : Any = (7 * i) % 16 A : Union[str, Any] = (f + a + added_consts[i] + block_words[g]) % 2**32 A : Dict = d A : Optional[int] = c A : Optional[int] = b A : Any = sum_aa(_lowerCamelCase , left_rotate_aa(_lowerCamelCase , shift_amounts[i] ) ) # Add hashed chunk to running total A : Dict = sum_aa(_lowerCamelCase , _lowerCamelCase ) A : Any = sum_aa(_lowerCamelCase , _lowerCamelCase ) A : Dict = sum_aa(_lowerCamelCase , _lowerCamelCase ) A : Union[str, Any] = sum_aa(_lowerCamelCase , _lowerCamelCase ) A : Optional[Any] = reformat_hex(_lowerCamelCase ) + reformat_hex(_lowerCamelCase ) + reformat_hex(_lowerCamelCase ) + reformat_hex(_lowerCamelCase ) return digest if __name__ == "__main__": import doctest doctest.testmod()
256
1
'''simple docstring''' import sys a__ : Optional[int] =( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def lowercase__ ( __lowercase : str = N ) -> Any: """simple docstring""" __UpperCamelCase = -sys.maxsize - 1 for i in range(len(SCREAMING_SNAKE_CASE__ ) - 12 ): __UpperCamelCase = 1 for j in range(13 ): product *= int(n[i + j] ) if product > largest_product: __UpperCamelCase = product return largest_product if __name__ == "__main__": print(f'{solution() = }')
53
from numpy import exp, pi, sqrt def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : float = 0.0 , SCREAMING_SNAKE_CASE__ : float = 1.0 ): return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
62
0
import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) __SCREAMING_SNAKE_CASE : Optional[int] = logging.getLogger(__name__) __SCREAMING_SNAKE_CASE : Any = tf.data.AUTOTUNE def snake_case () -> int: '''simple docstring''' _snake_case : str = argparse.ArgumentParser(description="Train a masked language model on TPU." ) parser.add_argument( "--pretrained_model_config" , type=__lowercase , default="roberta-base" , help="The model config to use. Note that we don't copy the model's weights, only the config!" , ) parser.add_argument( "--tokenizer" , type=__lowercase , default="unigram-tokenizer-wikitext" , help="The name of the tokenizer to load. We use the pretrained tokenizer to initialize the model's vocab size." , ) parser.add_argument( "--per_replica_batch_size" , type=__lowercase , default=8 , help="Batch size per TPU core." , ) parser.add_argument( "--no_tpu" , action="store_true" , help="If set, run on CPU and don't try to initialize a TPU. Useful for debugging on non-TPU instances." , ) parser.add_argument( "--tpu_name" , type=__lowercase , help="Name of TPU resource to initialize. Should be blank on Colab, and 'local' on TPU VMs." , default="local" , ) parser.add_argument( "--tpu_zone" , type=__lowercase , help="Google cloud zone that TPU resource is located in. Only used for non-Colab TPU nodes." , ) parser.add_argument( "--gcp_project" , type=__lowercase , help="Google cloud project name. Only used for non-Colab TPU nodes." ) parser.add_argument( "--bfloat16" , action="store_true" , help="Use mixed-precision bfloat16 for training. This is the recommended lower-precision format for TPU." , ) parser.add_argument( "--train_dataset" , type=__lowercase , help="Path to training dataset to load. If the path begins with `gs://`" " then the dataset will be loaded from a Google Cloud Storage bucket." , ) parser.add_argument( "--shuffle_buffer_size" , type=__lowercase , default=2**18 , help="Size of the shuffle buffer (in samples)" , ) parser.add_argument( "--eval_dataset" , type=__lowercase , help="Path to evaluation dataset to load. If the path begins with `gs://`" " then the dataset will be loaded from a Google Cloud Storage bucket." , ) parser.add_argument( "--num_epochs" , type=__lowercase , default=1 , help="Number of epochs to train for." , ) parser.add_argument( "--learning_rate" , type=__lowercase , default=1e-4 , help="Learning rate to use for training." , ) parser.add_argument( "--weight_decay_rate" , type=__lowercase , default=1e-3 , help="Weight decay rate to use for training." , ) parser.add_argument( "--max_length" , type=__lowercase , default=512 , help="Maximum length of tokenized sequences. Should match the setting used in prepare_tfrecord_shards.py" , ) parser.add_argument( "--mlm_probability" , type=__lowercase , default=0.15 , help="Fraction of tokens to mask during training." , ) parser.add_argument("--output_dir" , type=__lowercase , required=__lowercase , help="Path to save model checkpoints to." ) parser.add_argument("--hub_model_id" , type=__lowercase , help="Model ID to upload to on the Hugging Face Hub." ) _snake_case : List[Any] = parser.parse_args() return args def snake_case (__lowercase ) -> Any: '''simple docstring''' try: if args.tpu_name: _snake_case : Optional[Any] = tf.distribute.cluster_resolver.TPUClusterResolver( args.tpu_name , zone=args.tpu_zone , project=args.gcp_project ) else: _snake_case : List[str] = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: raise RuntimeError( "Couldn't connect to TPU! Most likely you need to specify --tpu_name, --tpu_zone, or " "--gcp_project. When running on a TPU VM, use --tpu_name local." ) tf.config.experimental_connect_to_cluster(__lowercase ) tf.tpu.experimental.initialize_tpu_system(__lowercase ) return tpu def snake_case (__lowercase ) -> Tuple: '''simple docstring''' _snake_case : List[str] = 0 for file in file_list: _snake_case : Tuple = file.split("/" )[-1] _snake_case : List[str] = re.search(r"-\d+-(\d+)\.tfrecord" , __lowercase ).group(1 ) _snake_case : str = int(__lowercase ) num_samples += sample_count return num_samples def snake_case (__lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase=None ) -> int: '''simple docstring''' _snake_case : Union[str, Any] = count_samples(__lowercase ) _snake_case : List[Any] = tf.data.Dataset.from_tensor_slices(__lowercase ) if shuffle: _snake_case : List[str] = dataset.shuffle(len(__lowercase ) ) _snake_case : str = tf.data.TFRecordDataset(__lowercase , num_parallel_reads=__lowercase ) # TF can't infer the total sample count because it doesn't read all the records yet, so we assert it here _snake_case : str = dataset.apply(tf.data.experimental.assert_cardinality(__lowercase ) ) _snake_case : List[Any] = dataset.map(__lowercase , num_parallel_calls=__lowercase ) if shuffle: assert shuffle_buffer_size is not None _snake_case : Any = dataset.shuffle(args.shuffle_buffer_size ) _snake_case : Dict = dataset.batch(__lowercase , drop_remainder=__lowercase ) _snake_case : Union[str, Any] = dataset.map(__lowercase , num_parallel_calls=__lowercase ) _snake_case : int = dataset.prefetch(__lowercase ) return dataset def snake_case (__lowercase ) -> Optional[Any]: '''simple docstring''' if not args.no_tpu: _snake_case : List[Any] = initialize_tpu(__lowercase ) _snake_case : Optional[int] = tf.distribute.TPUStrategy(__lowercase ) else: _snake_case : Optional[Any] = tf.distribute.OneDeviceStrategy(device="/gpu:0" ) if args.bfloataa: tf.keras.mixed_precision.set_global_policy("mixed_bfloat16" ) _snake_case : List[Any] = AutoTokenizer.from_pretrained(args.tokenizer ) _snake_case : Optional[Any] = AutoConfig.from_pretrained(args.pretrained_model_config ) _snake_case : Union[str, Any] = tokenizer.vocab_size _snake_case : List[Any] = tf.io.gfile.glob(os.path.join(args.train_dataset , "*.tfrecord" ) ) if not training_records: raise ValueError(F"""No .tfrecord files found in {args.train_dataset}.""" ) _snake_case : Union[str, Any] = tf.io.gfile.glob(os.path.join(args.eval_dataset , "*.tfrecord" ) ) if not eval_records: raise ValueError(F"""No .tfrecord files found in {args.eval_dataset}.""" ) _snake_case : Optional[int] = count_samples(__lowercase ) _snake_case : Dict = num_train_samples // (args.per_replica_batch_size * strategy.num_replicas_in_sync) _snake_case : Optional[Any] = steps_per_epoch * args.num_epochs with strategy.scope(): _snake_case : Optional[Any] = TFAutoModelForMaskedLM.from_config(__lowercase ) model(model.dummy_inputs ) # Pass some dummy inputs through the model to ensure all the weights are built _snake_case ,_snake_case : Dict = create_optimizer( num_train_steps=__lowercase , num_warmup_steps=total_train_steps // 20 , init_lr=args.learning_rate , weight_decay_rate=args.weight_decay_rate , ) # Transformers models compute the right loss for their task by default when labels are passed, and will # use this for training unless you specify your own loss function in compile(). model.compile(optimizer=__lowercase , metrics=["accuracy"] ) def decode_fn(__lowercase ): _snake_case : Optional[Any] = { "input_ids": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), "attention_mask": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), } return tf.io.parse_single_example(__lowercase , __lowercase ) # Many of the data collators in Transformers are TF-compilable when return_tensors == "tf", so we can # use their methods in our data pipeline. _snake_case : Optional[int] = DataCollatorForLanguageModeling( tokenizer=__lowercase , mlm_probability=args.mlm_probability , mlm=__lowercase , return_tensors="tf" ) def mask_with_collator(__lowercase ): # TF really needs an isin() function _snake_case : str = ( ~tf.cast(batch["attention_mask"] , tf.bool ) | (batch["input_ids"] == tokenizer.cls_token_id) | (batch["input_ids"] == tokenizer.sep_token_id) ) _snake_case ,_snake_case : Union[str, Any] = data_collator.tf_mask_tokens( batch["input_ids"] , vocab_size=len(__lowercase ) , mask_token_id=tokenizer.mask_token_id , special_tokens_mask=__lowercase , ) return batch _snake_case : Union[str, Any] = args.per_replica_batch_size * strategy.num_replicas_in_sync _snake_case : Optional[Any] = prepare_dataset( __lowercase , decode_fn=__lowercase , mask_fn=__lowercase , batch_size=__lowercase , shuffle=__lowercase , shuffle_buffer_size=args.shuffle_buffer_size , ) _snake_case : Any = prepare_dataset( __lowercase , decode_fn=__lowercase , mask_fn=__lowercase , batch_size=__lowercase , shuffle=__lowercase , ) _snake_case : List[str] = [] if args.hub_model_id: callbacks.append( PushToHubCallback(output_dir=args.output_dir , hub_model_id=args.hub_model_id , tokenizer=__lowercase ) ) model.fit( __lowercase , validation_data=__lowercase , epochs=args.num_epochs , callbacks=__lowercase , ) model.save_pretrained(args.output_dir ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : int = parse_args() main(args)
284
from __future__ import annotations import string from itertools import cycle, product from pathlib import Path __SCREAMING_SNAKE_CASE : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) __SCREAMING_SNAKE_CASE : list[int] = [ord(letter) for letter in string.ascii_lowercase] __SCREAMING_SNAKE_CASE : set[int] = {ord(char) for char in VALID_CHARS} __SCREAMING_SNAKE_CASE : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def snake_case (__lowercase , __lowercase ) -> str | None: '''simple docstring''' _snake_case : str = "" _snake_case : int _snake_case : int _snake_case : int for keychar, cipherchar in zip(cycle(__lowercase ) , __lowercase ): _snake_case : str = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(__lowercase ) return decoded def snake_case (__lowercase ) -> list[str]: '''simple docstring''' _snake_case : list[str] = [] for key in product(__lowercase , repeat=3 ): _snake_case : Union[str, Any] = try_key(__lowercase , __lowercase ) if encoded is not None: possibles.append(__lowercase ) return possibles def snake_case (__lowercase , __lowercase ) -> list[str]: '''simple docstring''' return [possible for possible in possibles if common_word in possible.lower()] def snake_case (__lowercase = "p059_cipher.txt" ) -> int: '''simple docstring''' _snake_case : list[int] _snake_case : list[str] _snake_case : str _snake_case : str _snake_case : str = Path(__lowercase ).parent.joinpath(__lowercase ).read_text(encoding="utf-8" ) _snake_case : Dict = [int(__lowercase ) for number in data.strip().split("," )] _snake_case : Tuple = filter_valid_chars(__lowercase ) for common_word in COMMON_WORDS: _snake_case : Optional[int] = filter_common_word(__lowercase , __lowercase ) if len(__lowercase ) == 1: break _snake_case : int = possibles[0] return sum(ord(__lowercase ) for char in decoded_text ) if __name__ == "__main__": print(F'''{solution() = }''')
284
1
'''simple docstring''' from collections import Counter import numpy as np from sklearn import datasets from sklearn.model_selection import train_test_split a : int = datasets.load_iris() a : Union[str, Any] = np.array(data["data"]) a : Optional[Any] = np.array(data["target"]) a : List[Any] = data["target_names"] a , a , a , a : Dict = train_test_split(X, y) def lowercase ( __magic_name__ , __magic_name__ ): '''simple docstring''' return np.linalg.norm(np.array(__magic_name__ ) - np.array(__magic_name__ ) ) def lowercase ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__=5 ): '''simple docstring''' UpperCAmelCase : int = zip(__magic_name__ , __magic_name__ ) # List of distances of all points from the point to be classified UpperCAmelCase : List[Any] = [] for data_point in data: UpperCAmelCase : List[str] = euclidean_distance(data_point[0] , __magic_name__ ) distances.append((distance, data_point[1]) ) # Choosing 'k' points with the least distances. UpperCAmelCase : Union[str, Any] = [i[1] for i in sorted(__magic_name__ )[:k]] # Most commonly occurring class among them # is the class into which the point is classified UpperCAmelCase : List[str] = Counter(__magic_name__ ).most_common(1 )[0][0] return classes[result] if __name__ == "__main__": print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
311
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() a : List[str] = logging.get_logger(__name__) a : Optional[Any] = ["model.decoder.embed_positions.weights"] def lowercase ( __magic_name__ ): '''simple docstring''' if "emb" in name: UpperCAmelCase : str = name.replace("emb" , "model.decoder.embed_tokens" ) if "transformer" in name: UpperCAmelCase : List[str] = name.replace("transformer" , "model.decoder" ) if "cross_attention" in name: UpperCAmelCase : int = name.replace("cross_attention" , "encoder_attn" ) if "linear1" in name: UpperCAmelCase : List[Any] = name.replace("linear1" , "fc1" ) if "linear2" in name: UpperCAmelCase : int = name.replace("linear2" , "fc2" ) if "norm1" in name: UpperCAmelCase : Dict = name.replace("norm1" , "self_attn_layer_norm" ) if "norm_cross" in name: UpperCAmelCase : Any = name.replace("norm_cross" , "encoder_attn_layer_norm" ) if "norm2" in name: UpperCAmelCase : Union[str, Any] = name.replace("norm2" , "final_layer_norm" ) if "out_norm" in name: UpperCAmelCase : Dict = name.replace("out_norm" , "model.decoder.layer_norm" ) if "linears" in name: UpperCAmelCase : List[Any] = name.replace("linears" , "lm_heads" ) if "condition_provider.conditioners.description.output_proj" in name: UpperCAmelCase : Any = name.replace("condition_provider.conditioners.description.output_proj" , "enc_to_dec_proj" ) return name def lowercase ( __magic_name__ , __magic_name__ ): '''simple docstring''' UpperCAmelCase : Any = list(state_dict.keys() ) UpperCAmelCase : List[Any] = {} for key in keys: UpperCAmelCase : Any = state_dict.pop(__magic_name__ ) UpperCAmelCase : str = rename_keys(__magic_name__ ) if "in_proj_weight" in key: # split fused qkv proj UpperCAmelCase : Optional[int] = val[:hidden_size, :] UpperCAmelCase : Optional[Any] = val[hidden_size : 2 * hidden_size, :] UpperCAmelCase : Optional[Any] = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: UpperCAmelCase : str = val else: UpperCAmelCase : int = val return state_dict, enc_dec_proj_state_dict def lowercase ( __magic_name__ ): '''simple docstring''' if checkpoint == "small": # default config values UpperCAmelCase : List[Any] = 1024 UpperCAmelCase : Tuple = 24 UpperCAmelCase : Union[str, Any] = 16 elif checkpoint == "medium": UpperCAmelCase : List[Any] = 1536 UpperCAmelCase : Optional[Any] = 48 UpperCAmelCase : List[str] = 24 elif checkpoint == "large": UpperCAmelCase : List[Any] = 2048 UpperCAmelCase : str = 48 UpperCAmelCase : Optional[Any] = 32 else: raise ValueError(F"Checkpoint should be one of `['small', 'medium', 'large']`, got {checkpoint}." ) UpperCAmelCase : Tuple = MusicgenDecoderConfig( hidden_size=__magic_name__ , ffn_dim=hidden_size * 4 , num_hidden_layers=__magic_name__ , num_attention_heads=__magic_name__ , ) return config @torch.no_grad() def lowercase ( __magic_name__ , __magic_name__=None , __magic_name__=None , __magic_name__="cpu" ): '''simple docstring''' UpperCAmelCase : Union[str, Any] = MusicGen.get_pretrained(__magic_name__ , device=__magic_name__ ) UpperCAmelCase : List[str] = decoder_config_from_checkpoint(__magic_name__ ) UpperCAmelCase : Dict = fairseq_model.lm.state_dict() UpperCAmelCase , UpperCAmelCase : List[str] = rename_state_dict( __magic_name__ , hidden_size=decoder_config.hidden_size ) UpperCAmelCase : Any = TaEncoderModel.from_pretrained("t5-base" ) UpperCAmelCase : Any = EncodecModel.from_pretrained("facebook/encodec_32khz" ) UpperCAmelCase : int = MusicgenForCausalLM(__magic_name__ ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection UpperCAmelCase , UpperCAmelCase : Optional[int] = decoder.load_state_dict(__magic_name__ , strict=__magic_name__ ) for key in missing_keys.copy(): if key.startswith(("text_encoder", "audio_encoder") ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(__magic_name__ ) if len(__magic_name__ ) > 0: raise ValueError(F"Missing key(s) in state_dict: {missing_keys}" ) if len(__magic_name__ ) > 0: raise ValueError(F"Unexpected key(s) in state_dict: {unexpected_keys}" ) # init the composite model UpperCAmelCase : List[Any] = MusicgenForConditionalGeneration(text_encoder=__magic_name__ , audio_encoder=__magic_name__ , decoder=__magic_name__ ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(__magic_name__ ) # check we can do a forward pass UpperCAmelCase : Union[str, Any] = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) UpperCAmelCase : Optional[Any] = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): UpperCAmelCase : str = model(input_ids=__magic_name__ , decoder_input_ids=__magic_name__ ).logits if logits.shape != (8, 1, 2048): raise ValueError("Incorrect shape for logits" ) # now construct the processor UpperCAmelCase : Dict = AutoTokenizer.from_pretrained("t5-base" ) UpperCAmelCase : List[str] = AutoFeatureExtractor.from_pretrained("facebook/encodec_32khz" , padding_side="left" ) UpperCAmelCase : Dict = MusicgenProcessor(feature_extractor=__magic_name__ , tokenizer=__magic_name__ ) # set the appropriate bos/pad token ids UpperCAmelCase : List[Any] = 2048 UpperCAmelCase : Tuple = 2048 # set other default generation config params UpperCAmelCase : Tuple = int(30 * audio_encoder.config.frame_rate ) UpperCAmelCase : str = True UpperCAmelCase : Tuple = 3.0 if pytorch_dump_folder is not None: Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) logger.info(F"Saving model {checkpoint} to {pytorch_dump_folder}" ) model.save_pretrained(__magic_name__ ) processor.save_pretrained(__magic_name__ ) if repo_id: logger.info(F"Pushing model {checkpoint} to {repo_id}" ) model.push_to_hub(__magic_name__ ) processor.push_to_hub(__magic_name__ ) if __name__ == "__main__": a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint", default="small", type=str, help="Checkpoint size of the MusicGen model you'd like to convert. Can be one of: `['small', 'medium', 'large']`.", ) parser.add_argument( "--pytorch_dump_folder", required=True, default=None, type=str, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) parser.add_argument( "--device", default="cpu", type=str, help="Torch device to run the conversion, either cpu or cuda." ) a : int = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
311
1
from __future__ import annotations import string from itertools import cycle, product from pathlib import Path __lowerCamelCase : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) __lowerCamelCase : list[int] = [ord(letter) for letter in string.ascii_lowercase] __lowerCamelCase : set[int] = {ord(char) for char in VALID_CHARS} __lowerCamelCase : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> str | None: UpperCamelCase : List[Any] = "" UpperCamelCase : Tuple = 42 UpperCamelCase : List[Any] = 42 UpperCamelCase : Optional[Any] = 42 for keychar, cipherchar in zip(cycle(_UpperCamelCase ) , _UpperCamelCase ): UpperCamelCase : int = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(_UpperCamelCase ) return decoded def A_ ( _lowerCAmelCase ) -> list[str]: UpperCamelCase : Tuple = [] for key in product(_UpperCamelCase , repeat=3 ): UpperCamelCase : List[str] = try_key(_UpperCamelCase , _UpperCamelCase ) if encoded is not None: possibles.append(_UpperCamelCase ) return possibles def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def A_ ( _lowerCAmelCase = "p059_cipher.txt" ) -> int: UpperCamelCase : Tuple = 42 UpperCamelCase : int = 42 UpperCamelCase : Dict = 42 UpperCamelCase : Dict = 42 UpperCamelCase : Tuple = Path(_UpperCamelCase ).parent.joinpath(_UpperCamelCase ).read_text(encoding="utf-8" ) UpperCamelCase : Optional[Any] = [int(_UpperCamelCase ) for number in data.strip().split("," )] UpperCamelCase : Optional[int] = filter_valid_chars(_UpperCamelCase ) for common_word in COMMON_WORDS: UpperCamelCase : str = filter_common_word(_UpperCamelCase , _UpperCamelCase ) if len(_UpperCamelCase ) == 1: break UpperCamelCase : List[str] = possibles[0] return sum(ord(_UpperCamelCase ) for char in decoded_text ) if __name__ == "__main__": print(f"""{solution() = }""")
359
import sys from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers __lowerCamelCase : Optional[int] = """python tqdm regex requests packaging filelock numpy tokenizers""".split() if sys.version_info < (3, 7): pkgs_to_check_at_runtime.append("""dataclasses""") if sys.version_info < (3, 8): pkgs_to_check_at_runtime.append("""importlib_metadata""") for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f"""can't find {pkg} in {deps.keys()}, check dependency_versions_table.py""") def A_ ( _lowerCAmelCase , _lowerCAmelCase=None ) -> Optional[Any]: require_version(deps[pkg] , _lowerCAmelCase )
140
0
"""simple docstring""" import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration a_ = [ # tf -> hf ('/', '.'), ('layer_', 'layers.'), ('kernel', 'weight'), ('beta', 'bias'), ('gamma', 'weight'), ('pegasus', 'model'), ] a_ = [ ('.output.dense', '.fc2'), ('intermediate.LayerNorm', 'final_layer_norm'), ('intermediate.dense', 'fc1'), ] a_ = ( INIT_COMMON + [ ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.out_proj'), ('attention.self', 'self_attn'), ('attention.encdec.LayerNorm', 'encoder_attn_layer_norm'), ('attention.encdec_output.dense', 'encoder_attn.out_proj'), ('attention.encdec', 'encoder_attn'), ('key', 'k_proj'), ('value', 'v_proj'), ('query', 'q_proj'), ('decoder.LayerNorm', 'decoder.layernorm_embedding'), ] + END_COMMON ) a_ = ( INIT_COMMON + [ ('embeddings.word_embeddings', 'shared.weight'), ('embeddings.position_embeddings', 'embed_positions.weight'), ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.output'), ('attention.self', 'self_attn.self'), ('encoder.LayerNorm', 'encoder.layernorm_embedding'), ] + END_COMMON ) a_ = [ 'encdec/key/bias', 'encdec/query/bias', 'encdec/value/bias', 'self/key/bias', 'self/query/bias', 'self/value/bias', 'encdec_output/dense/bias', 'attention/output/dense/bias', ] def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase ): for tf_name, hf_name in patterns: __lowercase : Optional[Any] = k.replace(__UpperCamelCase , __UpperCamelCase ) return k def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase ): __lowercase : Optional[int] = BigBirdPegasusConfig(**__UpperCamelCase ) __lowercase : Tuple = BigBirdPegasusForConditionalGeneration(__UpperCamelCase ) __lowercase : int = torch_model.state_dict() __lowercase : Optional[int] = {} # separating decoder weights __lowercase : Tuple = {k: tf_weights[k] for k in tf_weights if k.startswith('''pegasus/decoder''' )} __lowercase : int = {k: tf_weights[k] for k in tf_weights if not k.startswith('''pegasus/decoder''' )} for k, v in tqdm(decoder_weights.items() , '''tf -> hf conversion''' ): __lowercase : Any = [k.endswith(__UpperCamelCase ) for ending in KEYS_TO_IGNORE] if any(__UpperCamelCase ): continue __lowercase : Optional[int] = DECODER_PATTERNS __lowercase : int = rename_state_dict_key(__UpperCamelCase , __UpperCamelCase ) if new_k not in state_dict: raise ValueError(f"""could not find new key {new_k} in state dict. (converted from {k})""" ) if any(True if i in k else False for i in ['''dense''', '''query''', '''key''', '''value'''] ): __lowercase : Tuple = v.T __lowercase : int = torch.from_numpy(__UpperCamelCase ) assert v.shape == state_dict[new_k].shape, f"""{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}""" for k, v in tqdm(remaining_weights.items() , '''tf -> hf conversion''' ): __lowercase : List[Any] = [k.endswith(__UpperCamelCase ) for ending in KEYS_TO_IGNORE] if any(__UpperCamelCase ): continue __lowercase : Union[str, Any] = REMAINING_PATTERNS __lowercase : int = rename_state_dict_key(__UpperCamelCase , __UpperCamelCase ) if new_k not in state_dict and k != "pegasus/embeddings/position_embeddings": raise ValueError(f"""could not find new key {new_k} in state dict. (converted from {k})""" ) if any(True if i in k else False for i in ['''dense''', '''query''', '''key''', '''value'''] ): __lowercase : Dict = v.T __lowercase : Tuple = torch.from_numpy(__UpperCamelCase ) if k != "pegasus/embeddings/position_embeddings": assert v.shape == state_dict[new_k].shape, f"""{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}""" __lowercase : List[Any] = mapping['''model.embed_positions.weight'''] __lowercase : int = mapping.pop('''model.embed_positions.weight''' ) __lowercase ,__lowercase : Optional[Any] = torch_model.load_state_dict(__UpperCamelCase , strict=__UpperCamelCase ) __lowercase : Optional[int] = [ k for k in missing if k not in [ '''final_logits_bias''', '''model.encoder.embed_tokens.weight''', '''model.decoder.embed_tokens.weight''', '''lm_head.weight''', ] ] assert unexpected_missing == [], f"""no matches found for the following torch keys {unexpected_missing}""" assert extra == [], f"""no matches found for the following tf keys {extra}""" return torch_model def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : List[str] = tf.train.list_variables(__UpperCamelCase ) __lowercase : Tuple = {} __lowercase : List[str] = ['''global_step'''] for name, shape in tqdm(__UpperCamelCase , desc='''converting tf checkpoint to dict''' ): __lowercase : str = any(pat in name for pat in ignore_name ) if skip_key: continue __lowercase : Optional[Any] = tf.train.load_variable(__UpperCamelCase , __UpperCamelCase ) __lowercase : Optional[Any] = array return tf_weights def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): __lowercase : Union[str, Any] = get_tf_weights_as_numpy(__UpperCamelCase ) __lowercase : Dict = convert_bigbird_pegasus(__UpperCamelCase , __UpperCamelCase ) torch_model.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument('--tf_ckpt_path', type=str, help='passed to tf.train.list_variables') parser.add_argument('--save_dir', default=None, type=str, help='Path to the output PyTorch model.') a_ = parser.parse_args() a_ = {} convert_bigbird_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir, config_update=config_update)
249
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a_ = { 'configuration_xlm_roberta': [ 'XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaConfig', 'XLMRobertaOnnxConfig', ], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['XLMRobertaTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['XLMRobertaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ 'XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMRobertaForCausalLM', 'XLMRobertaForMaskedLM', 'XLMRobertaForMultipleChoice', 'XLMRobertaForQuestionAnswering', 'XLMRobertaForSequenceClassification', 'XLMRobertaForTokenClassification', 'XLMRobertaModel', 'XLMRobertaPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ 'TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMRobertaForCausalLM', 'TFXLMRobertaForMaskedLM', 'TFXLMRobertaForMultipleChoice', 'TFXLMRobertaForQuestionAnswering', 'TFXLMRobertaForSequenceClassification', 'TFXLMRobertaForTokenClassification', 'TFXLMRobertaModel', 'TFXLMRobertaPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ 'FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'FlaxXLMRobertaForMaskedLM', 'FlaxXLMRobertaForCausalLM', 'FlaxXLMRobertaForMultipleChoice', 'FlaxXLMRobertaForQuestionAnswering', 'FlaxXLMRobertaForSequenceClassification', 'FlaxXLMRobertaForTokenClassification', 'FlaxXLMRobertaModel', 'FlaxXLMRobertaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaConfig, XLMRobertaOnnxConfig, ) try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta import XLMRobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta_fast import XLMRobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaForCausalLM, XLMRobertaForMaskedLM, XLMRobertaForMultipleChoice, XLMRobertaForQuestionAnswering, XLMRobertaForSequenceClassification, XLMRobertaForTokenClassification, XLMRobertaModel, XLMRobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm_roberta import ( TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMRobertaForCausalLM, TFXLMRobertaForMaskedLM, TFXLMRobertaForMultipleChoice, TFXLMRobertaForQuestionAnswering, TFXLMRobertaForSequenceClassification, TFXLMRobertaForTokenClassification, TFXLMRobertaModel, TFXLMRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xlm_roberta import ( FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxXLMRobertaForCausalLM, FlaxXLMRobertaForMaskedLM, FlaxXLMRobertaForMultipleChoice, FlaxXLMRobertaForQuestionAnswering, FlaxXLMRobertaForSequenceClassification, FlaxXLMRobertaForTokenClassification, FlaxXLMRobertaModel, FlaxXLMRobertaPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
249
1
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' UpperCamelCase_ : int = ['''image_processor''', '''tokenizer'''] UpperCamelCase_ : str = '''ViltImageProcessor''' UpperCamelCase_ : List[str] = ('''BertTokenizer''', '''BertTokenizerFast''') def __init__( self : int , UpperCAmelCase_ : Dict=None , UpperCAmelCase_ : Optional[int]=None , **UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , UpperCAmelCase_ , ) SCREAMING_SNAKE_CASE : Tuple = kwargs.pop("feature_extractor" ) SCREAMING_SNAKE_CASE : Tuple = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : str = self.image_processor def __call__( self : List[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , UpperCAmelCase_ : bool = True , UpperCAmelCase_ : Union[bool, str, PaddingStrategy] = False , UpperCAmelCase_ : Union[bool, str, TruncationStrategy] = None , UpperCAmelCase_ : Optional[int] = None , UpperCAmelCase_ : int = 0 , UpperCAmelCase_ : Optional[int] = None , UpperCAmelCase_ : Optional[bool] = None , UpperCAmelCase_ : Optional[bool] = None , UpperCAmelCase_ : bool = False , UpperCAmelCase_ : bool = False , UpperCAmelCase_ : bool = False , UpperCAmelCase_ : bool = False , UpperCAmelCase_ : bool = True , UpperCAmelCase_ : Optional[Union[str, TensorType]] = None , **UpperCAmelCase_ : str , ): SCREAMING_SNAKE_CASE : Dict = self.tokenizer( text=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=UpperCAmelCase_ , stride=UpperCAmelCase_ , pad_to_multiple_of=UpperCAmelCase_ , return_token_type_ids=UpperCAmelCase_ , return_attention_mask=UpperCAmelCase_ , return_overflowing_tokens=UpperCAmelCase_ , return_special_tokens_mask=UpperCAmelCase_ , return_offsets_mapping=UpperCAmelCase_ , return_length=UpperCAmelCase_ , verbose=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ , ) # add pixel_values + pixel_mask SCREAMING_SNAKE_CASE : List[str] = self.image_processor(UpperCAmelCase_ , return_tensors=UpperCAmelCase_ ) encoding.update(UpperCAmelCase_ ) return encoding def _A ( self : Optional[int] , *UpperCAmelCase_ : str , **UpperCAmelCase_ : Any ): return self.tokenizer.batch_decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) def _A ( self : Any , *UpperCAmelCase_ : int , **UpperCAmelCase_ : str ): return self.tokenizer.decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) @property def _A ( self : Any ): SCREAMING_SNAKE_CASE : Tuple = self.tokenizer.model_input_names SCREAMING_SNAKE_CASE : str = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _A ( self : int ): warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , UpperCAmelCase_ , ) return self.image_processor_class @property def _A ( self : List[str] ): warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , UpperCAmelCase_ , ) return self.image_processor
319
import argparse import os import torch from transformers.utils import WEIGHTS_NAME snake_case = ["""small""", """medium""", """large"""] snake_case = """lm_head.decoder.weight""" snake_case = """lm_head.weight""" def lowerCamelCase__ ( lowercase , lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = torch.load(lowercase ) SCREAMING_SNAKE_CASE : Any = d.pop(lowercase ) os.makedirs(lowercase , exist_ok=lowercase ) torch.save(lowercase , os.path.join(lowercase , lowercase ) ) if __name__ == "__main__": snake_case = argparse.ArgumentParser() parser.add_argument("""--dialogpt_path""", default=""".""", type=str) snake_case = parser.parse_args() for MODEL in DIALOGPT_MODELS: snake_case = os.path.join(args.dialogpt_path, F"""{MODEL}_ft.pkl""") snake_case = F"""./DialoGPT-{MODEL}""" convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
319
1
'''simple docstring''' import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType lowerCAmelCase : Optional[List[str]] =None lowerCAmelCase : Optional[Any] ='''<''' if sys.byteorder == '''little''' else '''>''' # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image lowerCAmelCase : Union[str, Any] =[ np.dtype('''|b1'''), np.dtype('''|u1'''), np.dtype('''<u2'''), np.dtype('''>u2'''), np.dtype('''<i2'''), np.dtype('''>i2'''), np.dtype('''<u4'''), np.dtype('''>u4'''), np.dtype('''<i4'''), np.dtype('''>i4'''), np.dtype('''<f4'''), np.dtype('''>f4'''), np.dtype('''<f8'''), np.dtype('''>f8'''), ] @dataclass class a_ : __A = True __A = None # Automatically constructed __A = "PIL.Image.Image" __A = pa.struct({"bytes": pa.binary(), "path": pa.string()} ) __A = field(default="Image" , init=_lowerCAmelCase , repr=_lowerCAmelCase ) def __call__( self : Optional[int] ): """simple docstring""" return self.pa_type def lowercase__ ( self : Optional[int] , lowercase : Union[str, bytes, dict, np.ndarray, "PIL.Image.Image"] ): """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) if isinstance(lowercase , lowercase ): lowercase_ :Union[str, Any] = np.array(lowercase ) if isinstance(lowercase , lowercase ): return {"path": value, "bytes": None} elif isinstance(lowercase , lowercase ): return {"path": None, "bytes": value} elif isinstance(lowercase , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(lowercase ) elif isinstance(lowercase , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(lowercase ) elif value.get("path" ) is not None and os.path.isfile(value["path"] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get("path" )} elif value.get("bytes" ) is not None or value.get("path" ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get("bytes" ), "path": value.get("path" )} else: raise ValueError( F'An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.' ) def lowercase__ ( self : int , lowercase : dict , lowercase : List[str]=None ): """simple docstring""" if not self.decode: raise RuntimeError("Decoding is disabled for this feature. Please use Image(decode=True) instead." ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support decoding images, please install 'Pillow'." ) if token_per_repo_id is None: lowercase_ :str = {} lowercase_ , lowercase_ :List[str] = value["path"], value["bytes"] if bytes_ is None: if path is None: raise ValueError(F'An image should have one of \'path\' or \'bytes\' but both are None in {value}.' ) else: if is_local_path(lowercase ): lowercase_ :Dict = PIL.Image.open(lowercase ) else: lowercase_ :Optional[int] = path.split("::" )[-1] try: lowercase_ :List[Any] = string_to_dict(lowercase , config.HUB_DATASETS_URL )["repo_id"] lowercase_ :Dict = token_per_repo_id.get(lowercase ) except ValueError: lowercase_ :Union[str, Any] = None with xopen(lowercase , "rb" , use_auth_token=lowercase ) as f: lowercase_ :List[str] = BytesIO(f.read() ) lowercase_ :str = PIL.Image.open(bytes_ ) else: lowercase_ :List[str] = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def lowercase__ ( self : Optional[Any] ): """simple docstring""" from .features import Value return ( self if self.decode else { "bytes": Value("binary" ), "path": Value("string" ), } ) def lowercase__ ( self : Union[str, Any] , lowercase : Union[pa.StringArray, pa.StructArray, pa.ListArray] ): """simple docstring""" if pa.types.is_string(storage.type ): lowercase_ :Any = pa.array([None] * len(lowercase ) , type=pa.binary() ) lowercase_ :Dict = pa.StructArray.from_arrays([bytes_array, storage] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): lowercase_ :List[Any] = pa.array([None] * len(lowercase ) , type=pa.string() ) lowercase_ :Optional[int] = pa.StructArray.from_arrays([storage, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("bytes" ) >= 0: lowercase_ :Union[str, Any] = storage.field("bytes" ) else: lowercase_ :int = pa.array([None] * len(lowercase ) , type=pa.binary() ) if storage.type.get_field_index("path" ) >= 0: lowercase_ :List[str] = storage.field("path" ) else: lowercase_ :Any = pa.array([None] * len(lowercase ) , type=pa.string() ) lowercase_ :Any = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): lowercase_ :List[str] = pa.array( [encode_np_array(np.array(lowercase ) )["bytes"] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) lowercase_ :List[Any] = pa.array([None] * len(lowercase ) , type=pa.string() ) lowercase_ :Optional[int] = pa.StructArray.from_arrays( [bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(lowercase , self.pa_type ) def lowercase__ ( self : Optional[Any] , lowercase : pa.StructArray ): """simple docstring""" @no_op_if_value_is_null def path_to_bytes(lowercase : List[str] ): with xopen(lowercase , "rb" ) as f: lowercase_ :Optional[int] = f.read() return bytes_ lowercase_ :Tuple = pa.array( [ (path_to_bytes(x["path"] ) if x["bytes"] is None else x["bytes"]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) lowercase_ :int = pa.array( [os.path.basename(lowercase ) if path is not None else None for path in storage.field("path" ).to_pylist()] , type=pa.string() , ) lowercase_ :Tuple = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(lowercase , self.pa_type ) def UpperCAmelCase_ ( ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() lowercase_ :Union[str, Any] = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def UpperCAmelCase_ ( __lowerCamelCase : "PIL.Image.Image" ): lowercase_ :List[Any] = BytesIO() if image.format in list_image_compression_formats(): lowercase_ :Optional[int] = image.format else: lowercase_ :List[str] = "PNG" if image.mode in ["1", "L", "LA", "RGB", "RGBA"] else "TIFF" image.save(__lowerCamelCase ,format=__lowerCamelCase ) return buffer.getvalue() def UpperCAmelCase_ ( __lowerCamelCase : "PIL.Image.Image" ): if hasattr(__lowerCamelCase ,"filename" ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(__lowerCamelCase )} def UpperCAmelCase_ ( __lowerCamelCase : np.ndarray ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) lowercase_ :Tuple = array.dtype lowercase_ :List[Any] = dtype.byteorder if dtype.byteorder != "=" else _NATIVE_BYTEORDER lowercase_ :Dict = dtype.kind lowercase_ :Optional[Any] = dtype.itemsize lowercase_ :int = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: lowercase_ :Union[str, Any] = np.dtype("|u1" ) if dtype_kind not in ["u", "i"]: raise TypeError( F'Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.' ) if dtype is not dest_dtype: warnings.warn(F'Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'' ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: lowercase_ :List[str] = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: lowercase_ :Any = dtype_byteorder + dtype_kind + str(__lowerCamelCase ) lowercase_ :List[Any] = np.dtype(__lowerCamelCase ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(F'Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'' ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( F'Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}' ) lowercase_ :List[Any] = PIL.Image.fromarray(array.astype(__lowerCamelCase ) ) return {"path": None, "bytes": image_to_bytes(__lowerCamelCase )} def UpperCAmelCase_ ( __lowerCamelCase : Union[List[str], List[dict], List[np.ndarray], List["PIL.Image.Image"]] ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) if objs: lowercase_ , lowercase_ :int = first_non_null_value(__lowerCamelCase ) if isinstance(__lowerCamelCase ,__lowerCamelCase ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(__lowerCamelCase ,np.ndarray ): lowercase_ :Union[str, Any] = no_op_if_value_is_null(__lowerCamelCase ) return [obj_to_image_dict_func(__lowerCamelCase ) for obj in objs] elif isinstance(__lowerCamelCase ,PIL.Image.Image ): lowercase_ :int = no_op_if_value_is_null(__lowerCamelCase ) return [obj_to_image_dict_func(__lowerCamelCase ) for obj in objs] else: return objs else: return objs
223
'''simple docstring''' import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print('''Googling.....''') lowerCAmelCase : str ='''https://www.google.com/search?q=''' + ''' '''.join(sys.argv[1:]) lowerCAmelCase : List[str] =requests.get(url, headers={'''UserAgent''': UserAgent().random}) # res.raise_for_status() with open('''project1a.html''', '''wb''') as out_file: # only for knowing the class for data in res.iter_content(10_000): out_file.write(data) lowerCAmelCase : List[Any] =BeautifulSoup(res.text, '''html.parser''') lowerCAmelCase : List[Any] =list(soup.select('''.eZt8xd'''))[:5] print(len(links)) for link in links: if link.text == "Maps": webbrowser.open(link.get('''href''')) else: webbrowser.open(F'''https://google.com{link.get('href')}''')
223
1
import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __magic_name__ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): @register_to_config def __init__( self , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case = False , ) -> Dict: '''simple docstring''' super().__init__() __a =nn.Embedding(__snake_case , __snake_case ) __a =nn.Embedding(__snake_case , __snake_case ) __a =False __a =nn.Dropout(p=__snake_case ) __a =TaConfig( vocab_size=__snake_case , d_model=__snake_case , num_heads=__snake_case , d_kv=__snake_case , d_ff=__snake_case , dropout_rate=__snake_case , feed_forward_proj=__snake_case , is_decoder=__snake_case , is_encoder_decoder=__snake_case , ) __a =nn.ModuleList() for lyr_num in range(__snake_case ): __a =TaBlock(__snake_case ) self.encoders.append(__snake_case ) __a =TaLayerNorm(__snake_case ) __a =nn.Dropout(p=__snake_case ) def __magic_name__ ( self , __snake_case , __snake_case ) -> Optional[int]: '''simple docstring''' __a =self.token_embedder(__snake_case ) __a =encoder_input_tokens.shape[1] __a =torch.arange(__snake_case , device=encoder_input_tokens.device ) x += self.position_encoding(__snake_case ) __a =self.dropout_pre(__snake_case ) # inverted the attention mask __a =encoder_input_tokens.size() __a =self.get_extended_attention_mask(__snake_case , __snake_case ) for lyr in self.encoders: __a =lyr(__snake_case , __snake_case )[0] __a =self.layer_norm(__snake_case ) return self.dropout_post(__snake_case ), encoder_inputs_mask
308
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCAmelCase : Tuple = { "configuration_megatron_bert": ["MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MegatronBertConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : List[str] = [ "MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "MegatronBertForCausalLM", "MegatronBertForMaskedLM", "MegatronBertForMultipleChoice", "MegatronBertForNextSentencePrediction", "MegatronBertForPreTraining", "MegatronBertForQuestionAnswering", "MegatronBertForSequenceClassification", "MegatronBertForTokenClassification", "MegatronBertModel", "MegatronBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_megatron_bert import MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MegatronBertConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_megatron_bert import ( MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, MegatronBertPreTrainedModel, ) else: import sys _lowerCAmelCase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
308
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ : Tuple = logging.get_logger(__name__) lowerCamelCase_ : str = { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/config.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/config.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json""" ), """distilbert-base-uncased-finetuned-sst-2-english""": ( """https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json""" ), } class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" __lowerCAmelCase = "distilbert" __lowerCAmelCase = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self , __A=3_0522 , __A=512 , __A=False , __A=6 , __A=12 , __A=768 , __A=4 * 768 , __A=0.1 , __A=0.1 , __A="gelu" , __A=0.02 , __A=0.1 , __A=0.2 , __A=0 , **__A , ) -> List[Any]: a =vocab_size a =max_position_embeddings a =sinusoidal_pos_embds a =n_layers a =n_heads a =dim a =hidden_dim a =dropout a =attention_dropout a =activation a =initializer_range a =qa_dropout a =seq_classif_dropout super().__init__(**__A , pad_token_id=__A ) class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" @property def SCREAMING_SNAKE_CASE ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": a ={0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: a ={0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
81
"""simple docstring""" def _A ( lowercase = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" a =set() # Replace all the whitespace in our sentence a =input_str.replace(''' ''' , '''''' ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(lowercase ) == 26 def _A ( lowercase = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" a =[False] * 26 for char in input_str: if char.islower(): a =True elif char.isupper(): a =True return all(lowercase ) def _A ( lowercase = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def _A ( ): """simple docstring""" from timeit import timeit a ='''from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest''' print(timeit('''is_pangram()''' , setup=lowercase ) ) print(timeit('''is_pangram_faster()''' , setup=lowercase ) ) print(timeit('''is_pangram_fastest()''' , setup=lowercase ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
81
1
"""simple docstring""" import copy from typing import Dict, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING from ..detr import DetrConfig from ..swin import SwinConfig UpperCAmelCase : List[Any] = { 'facebook/maskformer-swin-base-ade': ( 'https://huggingface.co/facebook/maskformer-swin-base-ade/blob/main/config.json' ) # See all MaskFormer models at https://huggingface.co/models?filter=maskformer } UpperCAmelCase : Any = logging.get_logger(__name__) class lowerCamelCase__ ( A ): """simple docstring""" __a = """maskformer""" __a = {"""hidden_size""": """mask_feature_size"""} __a = ["""resnet""", """swin"""] __a = ["""detr"""] def __init__( self : Tuple , UpperCamelCase : int = 256 , UpperCamelCase : int = 256 , UpperCamelCase : float = 0.1 , UpperCamelCase : bool = False , UpperCamelCase : Optional[Dict] = None , UpperCamelCase : Optional[Dict] = None , UpperCamelCase : float = 0.02 , UpperCamelCase : float = 1.0 , UpperCamelCase : float = 1.0 , UpperCamelCase : float = 1.0 , UpperCamelCase : float = 20.0 , UpperCamelCase : Optional[bool] = None , **UpperCamelCase : List[str] , ): '''simple docstring''' if backbone_config is None: # fall back to https://huggingface.co/microsoft/swin-base-patch4-window12-384-in22k __UpperCAmelCase : List[Any] = SwinConfig( image_size=384 , in_channels=3 , patch_size=4 , embed_dim=128 , depths=[2, 2, 18, 2] , num_heads=[4, 8, 16, 32] , window_size=12 , drop_path_rate=0.3 , out_features=["""stage1""", """stage2""", """stage3""", """stage4"""] , ) if isinstance(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : Any = backbone_config.pop("""model_type""" ) __UpperCAmelCase : int = CONFIG_MAPPING[backbone_model_type] __UpperCAmelCase : Any = config_class.from_dict(UpperCamelCase ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( f'''Backbone {backbone_config.model_type} is not a supported model and may not be compatible with MaskFormer. ''' f'''Supported model types: {",".join(self.backbones_supported )}''' ) if decoder_config is None: # fall back to https://huggingface.co/facebook/detr-resnet-50 __UpperCAmelCase : Any = DetrConfig() else: # verify that the decoder is supported __UpperCAmelCase : Optional[int] = ( decoder_config.pop("""model_type""" ) if isinstance(UpperCamelCase , UpperCamelCase ) else decoder_config.model_type ) if decoder_type not in self.decoders_supported: raise ValueError( f'''Transformer Decoder {decoder_type} not supported, please use one of''' f''' {",".join(self.decoders_supported )}''' ) if isinstance(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : List[Any] = CONFIG_MAPPING[decoder_type] __UpperCAmelCase : List[Any] = config_class.from_dict(UpperCamelCase ) __UpperCAmelCase : Any = backbone_config __UpperCAmelCase : Dict = decoder_config # main feature dimension for the model __UpperCAmelCase : List[str] = fpn_feature_size __UpperCAmelCase : List[str] = mask_feature_size # initializer __UpperCAmelCase : List[str] = init_std __UpperCAmelCase : str = init_xavier_std # Hungarian matcher && loss __UpperCAmelCase : Optional[int] = cross_entropy_weight __UpperCAmelCase : str = dice_weight __UpperCAmelCase : List[str] = mask_weight __UpperCAmelCase : Tuple = use_auxiliary_loss __UpperCAmelCase : Union[str, Any] = no_object_weight __UpperCAmelCase : Optional[int] = output_auxiliary_logits __UpperCAmelCase : Optional[int] = self.decoder_config.encoder_attention_heads __UpperCAmelCase : List[Any] = self.decoder_config.num_hidden_layers super().__init__(**UpperCamelCase ) @classmethod def lowerCamelCase__ ( cls : Optional[Any] , UpperCamelCase : PretrainedConfig , UpperCamelCase : PretrainedConfig , **UpperCamelCase : Any ): '''simple docstring''' return cls( backbone_config=UpperCamelCase , decoder_config=UpperCamelCase , **UpperCamelCase , ) def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' __UpperCAmelCase : List[Any] = copy.deepcopy(self.__dict__ ) __UpperCAmelCase : List[str] = self.backbone_config.to_dict() __UpperCAmelCase : Dict = self.decoder_config.to_dict() __UpperCAmelCase : List[Any] = self.__class__.model_type return output
320
"""simple docstring""" from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
320
1
"""simple docstring""" class lowerCAmelCase_ : """simple docstring""" def __init__(self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = 0 SCREAMING_SNAKE_CASE__ : str = 0 SCREAMING_SNAKE_CASE__ : int = {} def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> Dict: """simple docstring""" if vertex not in self.adjacency: SCREAMING_SNAKE_CASE__ : Any = {} self.num_vertices += 1 def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Dict: """simple docstring""" self.add_vertex(SCREAMING_SNAKE_CASE__ ) self.add_vertex(SCREAMING_SNAKE_CASE__ ) if head == tail: return SCREAMING_SNAKE_CASE__ : int = weight SCREAMING_SNAKE_CASE__ : Optional[Any] = weight def __magic_name__ (self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = self.get_edges() for edge in edges: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = edge edges.remove((tail, head, weight) ) for i in range(len(SCREAMING_SNAKE_CASE__ ) ): SCREAMING_SNAKE_CASE__ : List[str] = list(edges[i] ) edges.sort(key=lambda SCREAMING_SNAKE_CASE__ : e[2] ) for i in range(len(SCREAMING_SNAKE_CASE__ ) - 1 ): if edges[i][2] >= edges[i + 1][2]: SCREAMING_SNAKE_CASE__ : Any = edges[i][2] + 1 for edge in edges: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Tuple = edge SCREAMING_SNAKE_CASE__ : Any = weight SCREAMING_SNAKE_CASE__ : List[str] = weight def __str__(self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = """""" for tail in self.adjacency: for head in self.adjacency[tail]: SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.adjacency[head][tail] string += F'''{head} -> {tail} == {weight}\n''' return string.rstrip("""\n""" ) def __magic_name__ (self ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = [] for tail in self.adjacency: for head in self.adjacency[tail]: output.append((tail, head, self.adjacency[head][tail]) ) return output def __magic_name__ (self ) -> List[str]: """simple docstring""" return self.adjacency.keys() @staticmethod def __magic_name__ (SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = Graph() if vertices is None: SCREAMING_SNAKE_CASE__ : Union[str, Any] = [] if edges is None: SCREAMING_SNAKE_CASE__ : List[str] = [] for vertex in vertices: g.add_vertex(SCREAMING_SNAKE_CASE__ ) for edge in edges: g.add_edge(*SCREAMING_SNAKE_CASE__ ) return g class lowerCAmelCase_ : """simple docstring""" def __init__(self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = {} SCREAMING_SNAKE_CASE__ : List[Any] = {} def __len__(self ) -> Any: """simple docstring""" return len(self.parent ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: """simple docstring""" if item in self.parent: return self.find(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : int = item SCREAMING_SNAKE_CASE__ : Dict = 0 return item def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> str: """simple docstring""" if item not in self.parent: return self.make_set(SCREAMING_SNAKE_CASE__ ) if item != self.parent[item]: SCREAMING_SNAKE_CASE__ : List[Any] = self.find(self.parent[item] ) return self.parent[item] def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = self.find(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : List[Any] = self.find(SCREAMING_SNAKE_CASE__ ) if roota == roota: return roota if self.rank[roota] > self.rank[roota]: SCREAMING_SNAKE_CASE__ : str = roota return roota if self.rank[roota] < self.rank[roota]: SCREAMING_SNAKE_CASE__ : Union[str, Any] = roota return roota if self.rank[roota] == self.rank[roota]: self.rank[roota] += 1 SCREAMING_SNAKE_CASE__ : Optional[Any] = roota return roota return None @staticmethod def __magic_name__ (SCREAMING_SNAKE_CASE__ ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = graph.num_vertices SCREAMING_SNAKE_CASE__ : List[str] = Graph.UnionFind() SCREAMING_SNAKE_CASE__ : Tuple = [] while num_components > 1: SCREAMING_SNAKE_CASE__ : List[str] = {} for vertex in graph.get_vertices(): SCREAMING_SNAKE_CASE__ : List[str] = -1 SCREAMING_SNAKE_CASE__ : str = graph.get_edges() for edge in edges: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Dict = edge edges.remove((tail, head, weight) ) for edge in edges: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = edge SCREAMING_SNAKE_CASE__ : int = union_find.find(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[int] = union_find.find(SCREAMING_SNAKE_CASE__ ) if seta != seta: if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: SCREAMING_SNAKE_CASE__ : Dict = [head, tail, weight] if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: SCREAMING_SNAKE_CASE__ : List[Any] = [head, tail, weight] for vertex in cheap_edge: if cheap_edge[vertex] != -1: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Tuple = cheap_edge[vertex] if union_find.find(SCREAMING_SNAKE_CASE__ ) != union_find.find(SCREAMING_SNAKE_CASE__ ): union_find.union(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) mst_edges.append(cheap_edge[vertex] ) SCREAMING_SNAKE_CASE__ : Tuple = num_components - 1 SCREAMING_SNAKE_CASE__ : int = Graph.build(edges=SCREAMING_SNAKE_CASE__ ) return mst
25
'''simple docstring''' import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class lowerCamelCase ( lowercase_ ): '''simple docstring''' def __init__( self : List[Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : Optional[int]=13 , lowerCAmelCase_ : Optional[int]=7 , lowerCAmelCase_ : List[str]=True , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : str=False , lowerCAmelCase_ : Optional[Any]=True , lowerCAmelCase_ : int=99 , lowerCAmelCase_ : List[Any]=32 , lowerCAmelCase_ : int=5 , lowerCAmelCase_ : Any=4 , lowerCAmelCase_ : Dict=64 , lowerCAmelCase_ : Any="gelu" , lowerCAmelCase_ : Dict=0.1 , lowerCAmelCase_ : Optional[int]=0.1 , lowerCAmelCase_ : str=5_12 , lowerCAmelCase_ : Optional[Any]=16 , lowerCAmelCase_ : Optional[Any]=2 , lowerCAmelCase_ : List[str]=0.02 , lowerCAmelCase_ : Any=3 , lowerCAmelCase_ : Union[str, Any]=4 , lowerCAmelCase_ : Optional[int]=None , lowerCAmelCase_ : List[Any]=2 , lowerCAmelCase_ : str=2 , lowerCAmelCase_ : Dict=2 , lowerCAmelCase_ : List[str]=2 , lowerCAmelCase_ : Union[str, Any]=4 , lowerCAmelCase_ : Union[str, Any]=1 , ) -> List[Any]: '''simple docstring''' A__ : Dict =parent A__ : Optional[int] =batch_size A__ : List[Any] =seq_length A__ : Any =is_training A__ : List[str] =use_input_mask A__ : str =use_token_type_ids A__ : Tuple =use_labels A__ : Tuple =vocab_size A__ : Optional[Any] =hidden_size A__ : Dict =num_hidden_layers A__ : str =num_attention_heads A__ : int =intermediate_size A__ : Union[str, Any] =hidden_act A__ : List[Any] =hidden_dropout_prob A__ : Union[str, Any] =attention_probs_dropout_prob A__ : Dict =max_position_embeddings A__ : Any =type_vocab_size A__ : Any =type_sequence_label_size A__ : int =initializer_range A__ : str =num_labels A__ : Optional[int] =num_choices A__ : Optional[int] =scope A__ : List[str] =q_groups A__ : Dict =k_groups A__ : Any =v_groups A__ : Optional[Any] =post_attention_groups A__ : Optional[int] =intermediate_groups A__ : Optional[int] =output_groups def lowercase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' A__ : str =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Optional[int] =None if self.use_input_mask: A__ : str =random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] =None A__ : Tuple =None A__ : Dict =None if self.use_labels: A__ : List[Any] =ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : List[str] =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : int =ids_tensor([self.batch_size] , self.num_choices ) A__ : str =self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase__ ( self : Dict ) -> int: '''simple docstring''' return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def lowercase__ ( self : Optional[Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : int ) -> List[str]: '''simple docstring''' A__ : Optional[Any] =SqueezeBertModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() A__ : Dict =model(lowerCAmelCase_ , lowerCAmelCase_ ) A__ : Dict =model(lowerCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self : Any , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Any , lowerCAmelCase_ : Any ) -> str: '''simple docstring''' A__ : Union[str, Any] =SqueezeBertForMaskedLM(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() A__ : Tuple =model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase__ ( self : Optional[int] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Optional[Any] ) -> Any: '''simple docstring''' A__ : str =SqueezeBertForQuestionAnswering(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() A__ : Union[str, Any] =model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , start_positions=lowerCAmelCase_ , end_positions=lowerCAmelCase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase__ ( self : Optional[int] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int] ) -> List[Any]: '''simple docstring''' A__ : Dict =self.num_labels A__ : int =SqueezeBertForSequenceClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() A__ : Dict =model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase__ ( self : Dict , lowerCAmelCase_ : Dict , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Dict ) -> Optional[int]: '''simple docstring''' A__ : str =self.num_labels A__ : int =SqueezeBertForTokenClassification(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() A__ : Dict =model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase__ ( self : Optional[Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[Any] ) -> Optional[int]: '''simple docstring''' A__ : Union[str, Any] =self.num_choices A__ : Dict =SqueezeBertForMultipleChoice(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() A__ : Optional[Any] =input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Any =input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Optional[Any] =model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , labels=lowerCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase__ ( self : Any ) -> int: '''simple docstring''' A__ : Any =self.prepare_config_and_inputs() ((A__) , (A__) , (A__) , (A__) , (A__) , (A__)) : Any =config_and_inputs A__ : str ={"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowerCamelCase ( lowercase_ , lowercase_ , unittest.TestCase ): '''simple docstring''' __snake_case = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) __snake_case = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) __snake_case = False __snake_case = True __snake_case = False def lowercase__ ( self : Optional[int] ) -> List[Any]: '''simple docstring''' A__ : Optional[Any] =SqueezeBertModelTester(self ) A__ : int =ConfigTester(self , config_class=lowerCAmelCase_ , dim=37 ) def lowercase__ ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def lowercase__ ( self : List[str] ) -> str: '''simple docstring''' A__ : Tuple =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*lowerCAmelCase_ ) def lowercase__ ( self : Optional[Any] ) -> str: '''simple docstring''' A__ : Tuple =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*lowerCAmelCase_ ) def lowercase__ ( self : Dict ) -> Any: '''simple docstring''' A__ : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*lowerCAmelCase_ ) def lowercase__ ( self : Union[str, Any] ) -> str: '''simple docstring''' A__ : Any =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*lowerCAmelCase_ ) def lowercase__ ( self : List[str] ) -> Dict: '''simple docstring''' A__ : List[str] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*lowerCAmelCase_ ) def lowercase__ ( self : str ) -> Optional[int]: '''simple docstring''' A__ : List[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*lowerCAmelCase_ ) @slow def lowercase__ ( self : Tuple ) -> Tuple: '''simple docstring''' for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : int =SqueezeBertModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) @require_sentencepiece @require_tokenizers @require_torch class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def lowercase__ ( self : Optional[Any] ) -> List[Any]: '''simple docstring''' A__ : List[str] =SqueezeBertForSequenceClassification.from_pretrained("""squeezebert/squeezebert-mnli""" ) A__ : List[str] =torch.tensor([[1, 2_94_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 13, 15_88, 2]] ) A__ : Tuple =model(lowerCAmelCase_ )[0] A__ : Union[str, Any] =torch.Size((1, 3) ) self.assertEqual(output.shape , lowerCAmelCase_ ) A__ : Tuple =torch.tensor([[0.6401, -0.0349, -0.6041]] ) self.assertTrue(torch.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) )
134
0
import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel lowerCAmelCase__ = logging.getLogger(__name__) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" if os.path.exists(snake_case_ ): if os.path.exists(os.path.join(snake_case_ , "config.json" ) ) and os.path.isfile( os.path.join(snake_case_ , "config.json" ) ): os.remove(os.path.join(snake_case_ , "config.json" ) ) if os.path.exists(os.path.join(snake_case_ , "pytorch_model.bin" ) ) and os.path.isfile( os.path.join(snake_case_ , "pytorch_model.bin" ) ): os.remove(os.path.join(snake_case_ , "pytorch_model.bin" ) ) else: os.makedirs(snake_case_ ) model.save_pretrained(snake_case_ ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ): """simple docstring""" UpperCamelCase = 2 if unlogit: UpperCamelCase = torch.pow(snake_case_ , snake_case_ ) UpperCamelCase = p * torch.log(snake_case_ ) UpperCamelCase = 0 return -plogp.sum(dim=-1 ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" logger.info("lv, h >\t" + "\t".join(F"{x + 1}" for x in range(len(snake_case_ ) ) ) ) for row in range(len(snake_case_ ) ): if tensor.dtype != torch.long: logger.info(F"layer {row + 1}:\t" + "\t".join(F"{x:.5f}" for x in tensor[row].cpu().data ) ) else: logger.info(F"layer {row + 1}:\t" + "\t".join(F"{x:d}" for x in tensor[row].cpu().data ) ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=False ): """simple docstring""" UpperCamelCase , UpperCamelCase = model.config.num_hidden_layers, model.config.num_attention_heads UpperCamelCase = torch.zeros(snake_case_ , snake_case_ ).to(args.device ) UpperCamelCase = torch.zeros(snake_case_ , snake_case_ ).to(args.device ) if head_mask is None: UpperCamelCase = torch.ones(snake_case_ , snake_case_ ).to(args.device ) head_mask.requires_grad_(requires_grad=snake_case_ ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: UpperCamelCase = None UpperCamelCase = 0.0 UpperCamelCase = 0.0 for step, inputs in enumerate(tqdm(snake_case_ , desc="Iteration" , disable=args.local_rank not in [-1, 0] ) ): UpperCamelCase = tuple(t.to(args.device ) for t in inputs ) ((UpperCamelCase ) , ) = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) UpperCamelCase = model(snake_case_ , labels=snake_case_ , head_mask=snake_case_ ) # (loss), lm_logits, presents, (all hidden_states), (attentions) UpperCamelCase , UpperCamelCase , UpperCamelCase = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(snake_case_ ): UpperCamelCase = entropy(attn.detach() , snake_case_ ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(snake_case_ ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: UpperCamelCase = 2 UpperCamelCase = torch.pow(torch.pow(snake_case_ , snake_case_ ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1e-20 if not args.dont_normalize_global_importance: UpperCamelCase = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info("Attention entropies" ) print_ad_tensor(snake_case_ ) if compute_importance: logger.info("Head importance scores" ) print_ad_tensor(snake_case_ ) logger.info("Head ranked by importance scores" ) UpperCamelCase = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) UpperCamelCase = torch.arange( head_importance.numel() , device=args.device ) UpperCamelCase = head_ranks.view_as(snake_case_ ) print_ad_tensor(snake_case_ ) return attn_entropy, head_importance, total_loss def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase = compute_heads_importance(snake_case_ , snake_case_ , snake_case_ , compute_entropy=snake_case_ ) UpperCamelCase = 1 / loss # instead of downsteam score use the LM loss logger.info("Pruning: original score: %f, threshold: %f" , snake_case_ , original_score * args.masking_threshold ) UpperCamelCase = torch.ones_like(snake_case_ ) UpperCamelCase = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) UpperCamelCase = original_score while current_score >= original_score * args.masking_threshold: UpperCamelCase = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads UpperCamelCase = float("Inf" ) UpperCamelCase = head_importance.view(-1 ).sort()[1] if len(snake_case_ ) <= num_to_mask: print("BREAK BY num_to_mask" ) break # mask heads UpperCamelCase = current_heads_to_mask[:num_to_mask] logger.info("Heads to mask: %s" , str(current_heads_to_mask.tolist() ) ) UpperCamelCase = new_head_mask.view(-1 ) UpperCamelCase = 0.0 UpperCamelCase = new_head_mask.view_as(snake_case_ ) UpperCamelCase = new_head_mask.clone().detach() print_ad_tensor(snake_case_ ) # Compute metric and head importance again UpperCamelCase , UpperCamelCase , UpperCamelCase = compute_heads_importance( snake_case_ , snake_case_ , snake_case_ , compute_entropy=snake_case_ , head_mask=snake_case_ ) UpperCamelCase = 1 / loss logger.info( "Masking: current score: %f, remaining heads %d (%.1f percents)" , snake_case_ , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 100 , ) logger.info("Final head mask" ) print_ad_tensor(snake_case_ ) np.save(os.path.join(args.output_dir , "head_mask.npy" ) , head_mask.detach().cpu().numpy() ) return head_mask def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = datetime.now() UpperCamelCase , UpperCamelCase , UpperCamelCase = compute_heads_importance( snake_case_ , snake_case_ , snake_case_ , compute_entropy=snake_case_ , compute_importance=snake_case_ , head_mask=snake_case_ ) UpperCamelCase = 1 / loss UpperCamelCase = datetime.now() - before_time UpperCamelCase = sum(p.numel() for p in model.parameters() ) UpperCamelCase = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(snake_case_ ) ) } for k, v in heads_to_prune.items(): if isinstance(snake_case_ , snake_case_ ): UpperCamelCase = [ v, ] assert sum(len(snake_case_ ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(snake_case_ ) UpperCamelCase = sum(p.numel() for p in model.parameters() ) UpperCamelCase = datetime.now() UpperCamelCase , UpperCamelCase , UpperCamelCase = compute_heads_importance( snake_case_ , snake_case_ , snake_case_ , compute_entropy=snake_case_ , compute_importance=snake_case_ , head_mask=snake_case_ , actually_pruned=snake_case_ , ) UpperCamelCase = 1 / loss UpperCamelCase = datetime.now() - before_time logger.info( "Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)" , snake_case_ , snake_case_ , pruned_num_params / original_num_params * 100 , ) logger.info("Pruning: score with masking: %f score with pruning: %f" , snake_case_ , snake_case_ ) logger.info("Pruning: speed ratio (original timing / new timing): %f percents" , original_time / new_time * 100 ) save_model(snake_case_ , args.output_dir ) def a__ ( ): """simple docstring""" UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--data_dir" , default=snake_case_ , type=snake_case_ , required=snake_case_ , help="The input data dir. Should contain the .tsv files (or other data files) for the task." , ) parser.add_argument( "--model_name_or_path" , default=snake_case_ , type=snake_case_ , required=snake_case_ , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--output_dir" , default=snake_case_ , type=snake_case_ , required=snake_case_ , help="The output directory where the model predictions and checkpoints will be written." , ) # Other parameters parser.add_argument( "--config_name" , default="" , type=snake_case_ , help="Pretrained config name or path if not the same as model_name_or_path" , ) parser.add_argument( "--tokenizer_name" , default="" , type=snake_case_ , help="Pretrained tokenizer name or path if not the same as model_name_or_path" , ) parser.add_argument( "--cache_dir" , default=snake_case_ , type=snake_case_ , help="Where do you want to store the pre-trained models downloaded from s3" , ) parser.add_argument( "--data_subset" , type=snake_case_ , default=-1 , help="If > 0: limit the data to a subset of data_subset instances." ) parser.add_argument( "--overwrite_output_dir" , action="store_true" , help="Whether to overwrite data in output directory" ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) parser.add_argument( "--dont_normalize_importance_by_layer" , action="store_true" , help="Don't normalize importance score by layers" ) parser.add_argument( "--dont_normalize_global_importance" , action="store_true" , help="Don't normalize all importance scores between 0 and 1" , ) parser.add_argument( "--try_masking" , action="store_true" , help="Whether to try to mask head until a threshold of accuracy." ) parser.add_argument( "--masking_threshold" , default=0.9 , type=snake_case_ , help="masking threshold in term of metrics (stop masking when metric < threshold * original metric value)." , ) parser.add_argument( "--masking_amount" , default=0.1 , type=snake_case_ , help="Amount to heads to masking at each masking step." ) parser.add_argument("--metric_name" , default="acc" , type=snake_case_ , help="Metric to use for head masking." ) parser.add_argument( "--max_seq_length" , default=128 , type=snake_case_ , help=( "The maximum total input sequence length after WordPiece tokenization. \n" "Sequences longer than this will be truncated, sequences shorter padded." ) , ) parser.add_argument("--batch_size" , default=1 , type=snake_case_ , help="Batch size." ) parser.add_argument("--seed" , type=snake_case_ , default=42 ) parser.add_argument("--local_rank" , type=snake_case_ , default=-1 , help="local_rank for distributed training on gpus" ) parser.add_argument("--no_cuda" , action="store_true" , help="Whether not to use CUDA when available" ) parser.add_argument("--server_ip" , type=snake_case_ , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=snake_case_ , default="" , help="Can be used for distant debugging." ) UpperCamelCase = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=snake_case_ ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: UpperCamelCase = torch.device("cuda" if torch.cuda.is_available() and not args.no_cuda else "cpu" ) UpperCamelCase = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) UpperCamelCase = torch.device("cuda" , args.local_rank ) UpperCamelCase = 1 torch.distributed.init_process_group(backend="nccl" ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info("device: {} n_gpu: {}, distributed: {}".format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) UpperCamelCase = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: UpperCamelCase = nn.parallel.DistributedDataParallel( snake_case_ , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=snake_case_ ) elif args.n_gpu > 1: UpperCamelCase = nn.DataParallel(snake_case_ ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=snake_case_ ) torch.save(snake_case_ , os.path.join(args.output_dir , "run_args.bin" ) ) logger.info("Training/evaluation parameters %s" , snake_case_ ) # Prepare dataset UpperCamelCase = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) UpperCamelCase = (torch.from_numpy(snake_case_ ),) UpperCamelCase = TensorDataset(*snake_case_ ) UpperCamelCase = RandomSampler(snake_case_ ) UpperCamelCase = DataLoader(snake_case_ , sampler=snake_case_ , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(snake_case_ , snake_case_ , snake_case_ ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: UpperCamelCase = mask_heads(snake_case_ , snake_case_ , snake_case_ ) prune_heads(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if __name__ == "__main__": main()
368
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml lowerCAmelCase__ = NewType('''DataClass''', Any) lowerCAmelCase__ = NewType('''DataClassType''', Any) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F"Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive)." ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = {str(_SCREAMING_SNAKE_CASE ): choice for choice in choices} return lambda _SCREAMING_SNAKE_CASE : str_to_choice.get(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def a__ ( *, _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = dataclasses.MISSING , _SCREAMING_SNAKE_CASE = dataclasses.MISSING , _SCREAMING_SNAKE_CASE = None , **_SCREAMING_SNAKE_CASE , ): """simple docstring""" if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls UpperCamelCase = {} if aliases is not None: UpperCamelCase = aliases if help is not None: UpperCamelCase = help return dataclasses.field(metadata=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , default_factory=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) class _lowerCamelCase ( _lowercase ): UpperCAmelCase_ = 42 def __init__(self , __a , **__a ) -> Any: # To make the default appear when using --help if "formatter_class" not in kwargs: UpperCamelCase = ArgumentDefaultsHelpFormatter super().__init__(**__a ) if dataclasses.is_dataclass(__a ): UpperCamelCase = [dataclass_types] UpperCamelCase = list(__a ) for dtype in self.dataclass_types: self._add_dataclass_arguments(__a ) @staticmethod def snake_case_ (__a , __a ) -> Optional[Any]: UpperCamelCase = F"--{field.name}" UpperCamelCase = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , __a ): raise RuntimeError( "Unresolved type detected, which should have been done with the help of " "`typing.get_type_hints` method by default" ) UpperCamelCase = kwargs.pop("aliases" , [] ) if isinstance(__a , __a ): UpperCamelCase = [aliases] UpperCamelCase = getattr(field.type , "__origin__" , field.type ) if origin_type is Union or (hasattr(__a , "UnionType" ) and isinstance(__a , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(__a ) not in field.type.__args__ ): raise ValueError( "Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because" " the argument parser only supports one type per argument." F" Problem encountered in field '{field.name}'." ) if type(__a ) not in field.type.__args__: # filter `str` in Union UpperCamelCase = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] UpperCamelCase = getattr(field.type , "__origin__" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) UpperCamelCase = ( field.type.__args__[0] if isinstance(__a , field.type.__args__[1] ) else field.type.__args__[1] ) UpperCamelCase = getattr(field.type , "__origin__" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) UpperCamelCase = {} if origin_type is Literal or (isinstance(field.type , __a ) and issubclass(field.type , __a )): if origin_type is Literal: UpperCamelCase = field.type.__args__ else: UpperCamelCase = [x.value for x in field.type] UpperCamelCase = make_choice_type_function(kwargs["choices"] ) if field.default is not dataclasses.MISSING: UpperCamelCase = field.default else: UpperCamelCase = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument UpperCamelCase = copy(__a ) # Hack because type=bool in argparse does not behave as we want. UpperCamelCase = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. UpperCamelCase = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way UpperCamelCase = default # This tells argparse we accept 0 or 1 value after --field_name UpperCamelCase = "?" # This is the value that will get picked if we do --field_name (without value) UpperCamelCase = True elif isclass(__a ) and issubclass(__a , __a ): UpperCamelCase = field.type.__args__[0] UpperCamelCase = "+" if field.default_factory is not dataclasses.MISSING: UpperCamelCase = field.default_factory() elif field.default is dataclasses.MISSING: UpperCamelCase = True else: UpperCamelCase = field.type if field.default is not dataclasses.MISSING: UpperCamelCase = field.default elif field.default_factory is not dataclasses.MISSING: UpperCamelCase = field.default_factory() else: UpperCamelCase = True parser.add_argument(__a , *__a , **__a ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): UpperCamelCase = False parser.add_argument(F"--no_{field.name}" , action="store_false" , dest=field.name , **__a ) def snake_case_ (self , __a ) -> List[Any]: if hasattr(__a , "_argument_group_name" ): UpperCamelCase = self.add_argument_group(dtype._argument_group_name ) else: UpperCamelCase = self try: UpperCamelCase = get_type_hints(__a ) except NameError: raise RuntimeError( F"Type resolution failed for {dtype}. Try declaring the class in global scope or " "removing line of `from __future__ import annotations` which opts in Postponed " "Evaluation of Annotations (PEP 563)" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(__a ): UpperCamelCase = ".".join(map(__a , sys.version_info[:3] ) ) raise RuntimeError( F"Type resolution failed for {dtype} on Python {python_version}. Try removing " "line of `from __future__ import annotations` which opts in union types as " "`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To " "support Python versions that lower than 3.10, you need to use " "`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of " "`X | None`." ) from ex raise for field in dataclasses.fields(__a ): if not field.init: continue UpperCamelCase = type_hints[field.name] self._parse_dataclass_field(__a , __a ) def snake_case_ (self , __a=None , __a=False , __a=True , __a=None , __a=None , ) -> Tuple[DataClass, ...]: if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): UpperCamelCase = [] if args_filename: args_files.append(Path(__a ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(".args" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values UpperCamelCase = ArgumentParser() args_file_parser.add_argument(__a , type=__a , action="append" ) # Use only remaining args for further parsing (remove the args_file_flag) UpperCamelCase , UpperCamelCase = args_file_parser.parse_known_args(args=__a ) UpperCamelCase = vars(__a ).get(args_file_flag.lstrip("-" ) , __a ) if cmd_args_file_paths: args_files.extend([Path(__a ) for p in cmd_args_file_paths] ) UpperCamelCase = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last UpperCamelCase = file_args + args if args is not None else file_args + sys.argv[1:] UpperCamelCase , UpperCamelCase = self.parse_known_args(args=__a ) UpperCamelCase = [] for dtype in self.dataclass_types: UpperCamelCase = {f.name for f in dataclasses.fields(__a ) if f.init} UpperCamelCase = {k: v for k, v in vars(__a ).items() if k in keys} for k in keys: delattr(__a , __a ) UpperCamelCase = dtype(**__a ) outputs.append(__a ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(__a ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F"Some specified arguments are not used by the HfArgumentParser: {remaining_args}" ) return (*outputs,) def snake_case_ (self , __a , __a = False ) -> Tuple[DataClass, ...]: UpperCamelCase = set(args.keys() ) UpperCamelCase = [] for dtype in self.dataclass_types: UpperCamelCase = {f.name for f in dataclasses.fields(__a ) if f.init} UpperCamelCase = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) UpperCamelCase = dtype(**__a ) outputs.append(__a ) if not allow_extra_keys and unused_keys: raise ValueError(F"Some keys are not used by the HfArgumentParser: {sorted(__a )}" ) return tuple(__a ) def snake_case_ (self , __a , __a = False ) -> Tuple[DataClass, ...]: with open(Path(__a ) , encoding="utf-8" ) as open_json_file: UpperCamelCase = json.loads(open_json_file.read() ) UpperCamelCase = self.parse_dict(__a , allow_extra_keys=__a ) return tuple(__a ) def snake_case_ (self , __a , __a = False ) -> Tuple[DataClass, ...]: UpperCamelCase = self.parse_dict(yaml.safe_load(Path(__a ).read_text() ) , allow_extra_keys=__a ) return tuple(__a )
244
0
"""simple docstring""" def _snake_case ( _snake_case : str , _snake_case : bool = False ): if not isinstance(_snake_case , _snake_case ): lowerCAmelCase : Optional[Any] = f'''Expected string as input, found {type(_snake_case )}''' raise ValueError(_snake_case ) if not isinstance(_snake_case , _snake_case ): lowerCAmelCase : Tuple = f'''Expected boolean as use_pascal parameter, found {type(_snake_case )}''' raise ValueError(_snake_case ) lowerCAmelCase : Any = input_str.split('''_''' ) lowerCAmelCase : str = 0 if use_pascal else 1 lowerCAmelCase : Tuple = words[start_index:] lowerCAmelCase : Dict = [word[0].upper() + word[1:] for word in words_to_capitalize] lowerCAmelCase : Dict = '''''' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
60
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Union[str, Any] =logging.get_logger(__name__) UpperCAmelCase : Optional[Any] ={ """sayakpaul/vit-msn-base""": """https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json""", # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class _lowercase (a_ ): '''simple docstring''' lowercase__ = """vit_msn""" def __init__( self , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3072 , snake_case__="gelu" , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.02 , snake_case__=1e-06 , snake_case__=224 , snake_case__=16 , snake_case__=3 , snake_case__=True , **snake_case__ , ): '''simple docstring''' super().__init__(**snake_case__ ) UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_attention_heads UpperCamelCase_ = intermediate_size UpperCamelCase_ = hidden_act UpperCamelCase_ = hidden_dropout_prob UpperCamelCase_ = attention_probs_dropout_prob UpperCamelCase_ = initializer_range UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = image_size UpperCamelCase_ = patch_size UpperCamelCase_ = num_channels UpperCamelCase_ = qkv_bias
128
0
'''simple docstring''' from __future__ import annotations import time a : Any = list[tuple[int, int]] a : Any = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] a : List[str] = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class a : def __init__( self : Optional[Any] , lowercase_ : int , lowercase_ : int , lowercase_ : int , lowercase_ : int , lowercase_ : Node | None ): snake_case_ = pos_x snake_case_ = pos_y snake_case_ = (pos_y, pos_x) snake_case_ = goal_x snake_case_ = goal_y snake_case_ = parent class a : def __init__( self : List[str] , lowercase_ : tuple[int, int] , lowercase_ : tuple[int, int] ): snake_case_ = Node(start[1] , start[0] , goal[1] , goal[0] , lowercase_ ) snake_case_ = Node(goal[1] , goal[0] , goal[1] , goal[0] , lowercase_ ) snake_case_ = [self.start] snake_case_ = False def A_ ( self : int ): while self.node_queue: snake_case_ = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: snake_case_ = True return self.retrace_path(lowercase_ ) snake_case_ = self.get_successors(lowercase_ ) for node in successors: self.node_queue.append(lowercase_ ) if not self.reached: return [self.start.pos] return None def A_ ( self : Union[str, Any] , lowercase_ : Node ): snake_case_ = [] for action in delta: snake_case_ = parent.pos_x + action[1] snake_case_ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowercase_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(lowercase_ , lowercase_ , self.target.pos_y , self.target.pos_x , lowercase_ ) ) return successors def A_ ( self : List[Any] , lowercase_ : Node | None ): snake_case_ = node snake_case_ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) snake_case_ = current_node.parent path.reverse() return path class a : def __init__( self : Optional[int] , lowercase_ : List[Any] , lowercase_ : List[Any] ): snake_case_ = BreadthFirstSearch(lowercase_ , lowercase_ ) snake_case_ = BreadthFirstSearch(lowercase_ , lowercase_ ) snake_case_ = False def A_ ( self : List[str] ): while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: snake_case_ = self.fwd_bfs.node_queue.pop(0 ) snake_case_ = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: snake_case_ = True return self.retrace_bidirectional_path( lowercase_ , lowercase_ ) snake_case_ = current_bwd_node snake_case_ = current_fwd_node snake_case_ = { self.fwd_bfs: self.fwd_bfs.get_successors(lowercase_ ), self.bwd_bfs: self.bwd_bfs.get_successors(lowercase_ ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(lowercase_ ) if not self.reached: return [self.fwd_bfs.start.pos] return None def A_ ( self : Optional[Any] , lowercase_ : Node , lowercase_ : Node ): snake_case_ = self.fwd_bfs.retrace_path(lowercase_ ) snake_case_ = self.bwd_bfs.retrace_path(lowercase_ ) bwd_path.pop() bwd_path.reverse() snake_case_ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() a : List[Any] = (0, 0) a : Optional[int] = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) a : str = time.time() a : int = BreadthFirstSearch(init, goal) a : str = bfs.search() a : Union[str, Any] = time.time() - start_bfs_time print('Unidirectional BFS computation time : ', bfs_time) a : List[Any] = time.time() a : Dict = BidirectionalBreadthFirstSearch(init, goal) a : Optional[int] = bd_bfs.search() a : List[Any] = time.time() - start_bd_bfs_time print('Bidirectional BFS computation time : ', bd_bfs_time)
72
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a : int = { 'configuration_xlm': ['XLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMConfig', 'XLMOnnxConfig'], 'tokenization_xlm': ['XLMTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ 'XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMForMultipleChoice', 'XLMForQuestionAnswering', 'XLMForQuestionAnsweringSimple', 'XLMForSequenceClassification', 'XLMForTokenClassification', 'XLMModel', 'XLMPreTrainedModel', 'XLMWithLMHeadModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[int] = [ 'TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMForMultipleChoice', 'TFXLMForQuestionAnsweringSimple', 'TFXLMForSequenceClassification', 'TFXLMForTokenClassification', 'TFXLMMainLayer', 'TFXLMModel', 'TFXLMPreTrainedModel', 'TFXLMWithLMHeadModel', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys a : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
72
1
'''simple docstring''' import pickle import numpy as np from matplotlib import pyplot as plt class a_ : def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=0.2 , _SCREAMING_SNAKE_CASE=0.2 ) -> Tuple: """simple docstring""" UpperCamelCase = bp_numa UpperCamelCase = bp_numa UpperCamelCase = bp_numa UpperCamelCase = conva_get[:2] UpperCamelCase = conva_get[2] UpperCamelCase = size_pa UpperCamelCase = rate_w UpperCamelCase = rate_t UpperCamelCase = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] UpperCamelCase = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCamelCase = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCamelCase = -2 * np.random.rand(self.conva[1] ) + 1 UpperCamelCase = -2 * np.random.rand(self.num_bpa ) + 1 UpperCamelCase = -2 * np.random.rand(self.num_bpa ) + 1 def A__ ( self , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" UpperCamelCase = { """num_bp1""": self.num_bpa, """num_bp2""": self.num_bpa, """num_bp3""": self.num_bpa, """conv1""": self.conva, """step_conv1""": self.step_conva, """size_pooling1""": self.size_poolinga, """rate_weight""": self.rate_weight, """rate_thre""": self.rate_thre, """w_conv1""": self.w_conva, """wkj""": self.wkj, """vji""": self.vji, """thre_conv1""": self.thre_conva, """thre_bp2""": self.thre_bpa, """thre_bp3""": self.thre_bpa, } with open(_SCREAMING_SNAKE_CASE , """wb""" ) as f: pickle.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print(F"Model saved: {save_path}" ) @classmethod def A__ ( cls , _SCREAMING_SNAKE_CASE ) -> int: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , """rb""" ) as f: UpperCamelCase = pickle.load(_SCREAMING_SNAKE_CASE ) # noqa: S301 UpperCamelCase = model_dic.get("""conv1""" ) conv_get.append(model_dic.get("""step_conv1""" ) ) UpperCamelCase = model_dic.get("""size_pooling1""" ) UpperCamelCase = model_dic.get("""num_bp1""" ) UpperCamelCase = model_dic.get("""num_bp2""" ) UpperCamelCase = model_dic.get("""num_bp3""" ) UpperCamelCase = model_dic.get("""rate_weight""" ) UpperCamelCase = model_dic.get("""rate_thre""" ) # create model instance UpperCamelCase = CNN(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # modify model parameter UpperCamelCase = model_dic.get("""w_conv1""" ) UpperCamelCase = model_dic.get("""wkj""" ) UpperCamelCase = model_dic.get("""vji""" ) UpperCamelCase = model_dic.get("""thre_conv1""" ) UpperCamelCase = model_dic.get("""thre_bp2""" ) UpperCamelCase = model_dic.get("""thre_bp3""" ) return conv_ins def A__ ( self , _SCREAMING_SNAKE_CASE ) -> List[str]: """simple docstring""" return 1 / (1 + np.exp(-1 * x )) def A__ ( self , _SCREAMING_SNAKE_CASE ) -> List[str]: """simple docstring""" return round(_SCREAMING_SNAKE_CASE , 3 ) def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" UpperCamelCase = convs[0] UpperCamelCase = convs[1] UpperCamelCase = np.shape(_SCREAMING_SNAKE_CASE )[0] # get the data slice of original image data, data_focus UpperCamelCase = [] for i_focus in range(0 , size_data - size_conv + 1 , _SCREAMING_SNAKE_CASE ): for j_focus in range(0 , size_data - size_conv + 1 , _SCREAMING_SNAKE_CASE ): UpperCamelCase = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(_SCREAMING_SNAKE_CASE ) # calculate the feature map of every single kernel, and saved as list of matrix UpperCamelCase = [] UpperCamelCase = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(_SCREAMING_SNAKE_CASE ): UpperCamelCase = [] for i_focus in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCamelCase = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(_SCREAMING_SNAKE_CASE ) ) UpperCamelCase = np.asmatrix(_SCREAMING_SNAKE_CASE ).reshape( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) data_featuremap.append(_SCREAMING_SNAKE_CASE ) # expanding the data slice to One dimenssion UpperCamelCase = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(_SCREAMING_SNAKE_CASE ) ) UpperCamelCase = np.asarray(_SCREAMING_SNAKE_CASE ) return focus_list, data_featuremap def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="average_pool" ) -> Optional[int]: """simple docstring""" UpperCamelCase = len(featuremaps[0] ) UpperCamelCase = int(size_map / size_pooling ) UpperCamelCase = [] for i_map in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCamelCase = featuremaps[i_map] UpperCamelCase = [] for i_focus in range(0 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): for j_focus in range(0 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): UpperCamelCase = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(_SCREAMING_SNAKE_CASE ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(_SCREAMING_SNAKE_CASE ) ) UpperCamelCase = np.asmatrix(_SCREAMING_SNAKE_CASE ).reshape(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) featuremap_pooled.append(_SCREAMING_SNAKE_CASE ) return featuremap_pooled def A__ ( self , _SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" UpperCamelCase = [] for i in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCamelCase = np.shape(data[i] ) UpperCamelCase = data[i].reshape(1 , shapes[0] * shapes[1] ) UpperCamelCase = data_listed.getA().tolist()[0] data_expanded.extend(_SCREAMING_SNAKE_CASE ) UpperCamelCase = np.asarray(_SCREAMING_SNAKE_CASE ) return data_expanded def A__ ( self , _SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" UpperCamelCase = np.asarray(_SCREAMING_SNAKE_CASE ) UpperCamelCase = np.shape(_SCREAMING_SNAKE_CASE ) UpperCamelCase = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" UpperCamelCase = [] UpperCamelCase = 0 for i_map in range(_SCREAMING_SNAKE_CASE ): UpperCamelCase = np.ones((size_map, size_map) ) for i in range(0 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): for j in range(0 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): UpperCamelCase = pd_pool[ i_pool ] UpperCamelCase = i_pool + 1 UpperCamelCase = np.multiply( _SCREAMING_SNAKE_CASE , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(_SCREAMING_SNAKE_CASE ) return pd_all def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=bool ) -> List[str]: """simple docstring""" print("""----------------------Start Training-------------------------""" ) print((""" - - Shape: Train_Data """, np.shape(_SCREAMING_SNAKE_CASE )) ) print((""" - - Shape: Teach_Data """, np.shape(_SCREAMING_SNAKE_CASE )) ) UpperCamelCase = 0 UpperCamelCase = [] UpperCamelCase = 10000 while rp < n_repeat and mse >= error_accuracy: UpperCamelCase = 0 print(F"-------------Learning Time {rp}--------------" ) for p in range(len(_SCREAMING_SNAKE_CASE ) ): # print('------------Learning Image: %d--------------'%p) UpperCamelCase = np.asmatrix(datas_train[p] ) UpperCamelCase = np.asarray(datas_teach[p] ) UpperCamelCase ,UpperCamelCase = self.convolute( _SCREAMING_SNAKE_CASE , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCamelCase = self.pooling(_SCREAMING_SNAKE_CASE , self.size_poolinga ) UpperCamelCase = np.shape(_SCREAMING_SNAKE_CASE ) UpperCamelCase = self._expand(_SCREAMING_SNAKE_CASE ) UpperCamelCase = data_bp_input UpperCamelCase = np.dot(_SCREAMING_SNAKE_CASE , self.vji.T ) - self.thre_bpa UpperCamelCase = self.sig(_SCREAMING_SNAKE_CASE ) UpperCamelCase = np.dot(_SCREAMING_SNAKE_CASE , self.wkj.T ) - self.thre_bpa UpperCamelCase = self.sig(_SCREAMING_SNAKE_CASE ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- UpperCamelCase = np.multiply( (data_teach - bp_outa) , np.multiply(_SCREAMING_SNAKE_CASE , (1 - bp_outa) ) ) UpperCamelCase = np.multiply( np.dot(_SCREAMING_SNAKE_CASE , self.wkj ) , np.multiply(_SCREAMING_SNAKE_CASE , (1 - bp_outa) ) ) UpperCamelCase = np.dot(_SCREAMING_SNAKE_CASE , self.vji ) UpperCamelCase = pd_i_all / (self.size_poolinga * self.size_poolinga) UpperCamelCase = pd_conva_pooled.T.getA().tolist() UpperCamelCase = self._calculate_gradient_from_pool( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): UpperCamelCase = self._expand_mat(pd_conva_all[k_conv] ) UpperCamelCase = self.rate_weight * np.dot(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) UpperCamelCase = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) UpperCamelCase = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer UpperCamelCase = self.wkj + pd_k_all.T * bp_outa * self.rate_weight UpperCamelCase = self.vji + pd_j_all.T * bp_outa * self.rate_weight UpperCamelCase = self.thre_bpa - pd_k_all * self.rate_thre UpperCamelCase = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image UpperCamelCase = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) UpperCamelCase = rp + 1 UpperCamelCase = error_count / patterns all_mse.append(_SCREAMING_SNAKE_CASE ) def draw_error(): UpperCamelCase = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(_SCREAMING_SNAKE_CASE , """+-""" ) plt.plot(_SCREAMING_SNAKE_CASE , """r--""" ) plt.xlabel("""Learning Times""" ) plt.ylabel("""All_mse""" ) plt.grid(_SCREAMING_SNAKE_CASE , alpha=0.5 ) plt.show() print("""------------------Training Complished---------------------""" ) print((""" - - Training epoch: """, rp, F" - - Mse: {mse:.6f}") ) if draw_e: draw_error() return mse def A__ ( self , _SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" UpperCamelCase = [] print("""-------------------Start Testing-------------------------""" ) print((""" - - Shape: Test_Data """, np.shape(_SCREAMING_SNAKE_CASE )) ) for p in range(len(_SCREAMING_SNAKE_CASE ) ): UpperCamelCase = np.asmatrix(datas_test[p] ) UpperCamelCase ,UpperCamelCase = self.convolute( _SCREAMING_SNAKE_CASE , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCamelCase = self.pooling(_SCREAMING_SNAKE_CASE , self.size_poolinga ) UpperCamelCase = self._expand(_SCREAMING_SNAKE_CASE ) UpperCamelCase = data_bp_input UpperCamelCase = bp_outa * self.vji.T - self.thre_bpa UpperCamelCase = self.sig(_SCREAMING_SNAKE_CASE ) UpperCamelCase = bp_outa * self.wkj.T - self.thre_bpa UpperCamelCase = self.sig(_SCREAMING_SNAKE_CASE ) produce_out.extend(bp_outa.getA().tolist() ) UpperCamelCase = [list(map(self.do_round , _SCREAMING_SNAKE_CASE ) ) for each in produce_out] return np.asarray(_SCREAMING_SNAKE_CASE ) def A__ ( self , _SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" UpperCamelCase = np.asmatrix(_SCREAMING_SNAKE_CASE ) UpperCamelCase ,UpperCamelCase = self.convolute( _SCREAMING_SNAKE_CASE , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCamelCase = self.pooling(_SCREAMING_SNAKE_CASE , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
321
'''simple docstring''' import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowercase__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase )-> List[str]: UpperCamelCase = 1.5 UpperCamelCase = int(factor * num_class_images ) UpperCamelCase = ClipClient( url="""https://knn.laion.ai/knn-service""" , indice_name="""laion_400m""" , num_images=__UpperCamelCase , aesthetic_weight=0.1 ) os.makedirs(F"{class_data_dir}/images" , exist_ok=__UpperCamelCase ) if len(list(Path(F"{class_data_dir}/images" ).iterdir() ) ) >= num_class_images: return while True: UpperCamelCase = client.query(text=__UpperCamelCase ) if len(__UpperCamelCase ) >= factor * num_class_images or num_images > 1E4: break else: UpperCamelCase = int(factor * num_images ) UpperCamelCase = ClipClient( url="""https://knn.laion.ai/knn-service""" , indice_name="""laion_400m""" , num_images=__UpperCamelCase , aesthetic_weight=0.1 , ) UpperCamelCase = 0 UpperCamelCase = 0 UpperCamelCase = tqdm(desc="""downloading real regularization images""" , total=__UpperCamelCase ) with open(F"{class_data_dir}/caption.txt" , """w""" ) as fa, open(F"{class_data_dir}/urls.txt" , """w""" ) as fa, open( F"{class_data_dir}/images.txt" , """w""" ) as fa: while total < num_class_images: UpperCamelCase = class_images[count] count += 1 try: UpperCamelCase = requests.get(images["""url"""] ) if img.status_code == 200: UpperCamelCase = Image.open(BytesIO(img.content ) ) with open(F"{class_data_dir}/images/{total}.jpg" , """wb""" ) as f: f.write(img.content ) fa.write(images["""caption"""] + """\n""" ) fa.write(images["""url"""] + """\n""" ) fa.write(F"{class_data_dir}/images/{total}.jpg" + """\n""" ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowercase__ ( )-> str: UpperCamelCase = argparse.ArgumentParser("""""" , add_help=__UpperCamelCase ) parser.add_argument("""--class_prompt""" , help="""text prompt to retrieve images""" , required=__UpperCamelCase , type=__UpperCamelCase ) parser.add_argument("""--class_data_dir""" , help="""path to save images""" , required=__UpperCamelCase , type=__UpperCamelCase ) parser.add_argument("""--num_class_images""" , help="""number of images to download""" , default=200 , type=__UpperCamelCase ) return parser.parse_args() if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
321
1
def lowerCamelCase__ ( _A , _A ): a : Any = 0 a : List[str] = len(lowerCAmelCase__ ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None a : str = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(lowerCAmelCase__ ): return None a : Union[str, Any] = sorted_collection[point] if current_item == item: return point else: if point < left: a : str = left a : Tuple = point elif point > right: a : Union[str, Any] = right a : Optional[Any] = point else: if item < current_item: a : Optional[int] = point - 1 else: a : Union[str, Any] = point + 1 return None def lowerCamelCase__ ( _A , _A , _A , _A ): # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None a : List[Any] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(lowerCAmelCase__ ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) elif point > right: return interpolation_search_by_recursion(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , point - 1 ) else: return interpolation_search_by_recursion( lowerCAmelCase__ , lowerCAmelCase__ , point + 1 , lowerCAmelCase__ ) def lowerCamelCase__ ( _A ): if collection != sorted(lowerCAmelCase__ ): raise ValueError('Collection must be ascending sorted' ) return True if __name__ == "__main__": import sys lowerCAmelCase: Tuple = 0 if debug == 1: lowerCAmelCase: int = [1_0, 3_0, 4_0, 4_5, 5_0, 6_6, 7_7, 9_3] try: __assert_sorted(collection) except ValueError: sys.exit('Sequence must be ascending sorted to apply interpolation search') lowerCAmelCase: int = 6_7 lowerCAmelCase: Tuple = interpolation_search(collection, target) if result is not None: print(F"{target} found at positions: {result}") else: print('Not found')
361
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a__( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): lowercase__ = CycleDiffusionPipeline lowercase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { """negative_prompt""", """height""", """width""", """negative_prompt_embeds""", } lowercase__ = PipelineTesterMixin.required_optional_params - {"""latents"""} lowercase__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""source_prompt"""} ) lowercase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase_ ( self : Any ): torch.manual_seed(0 ) a : Union[str, Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) a : str = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='scaled_linear' , num_train_timesteps=10_00 , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) a : List[str] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) a : int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) a : List[str] = CLIPTextModel(__snake_case ) a : int = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) a : Tuple = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase_ ( self : Optional[int] , __snake_case : Dict , __snake_case : Any=0 ): a : str = floats_tensor((1, 3, 32, 32) , rng=random.Random(__snake_case ) ).to(__snake_case ) a : Optional[Any] = image / 2 + 0.5 if str(__snake_case ).startswith('mps' ): a : List[str] = torch.manual_seed(__snake_case ) else: a : Union[str, Any] = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) a : List[Any] = { 'prompt': 'An astronaut riding an elephant', 'source_prompt': 'An astronaut riding a horse', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'eta': 0.1, 'strength': 0.8, 'guidance_scale': 3, 'source_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase_ ( self : Optional[int] ): a : Optional[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator a : int = self.get_dummy_components() a : str = CycleDiffusionPipeline(**__snake_case ) a : List[str] = pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) a : Dict = self.get_dummy_inputs(__snake_case ) a : Union[str, Any] = pipe(**__snake_case ) a : List[Any] = output.images a : Optional[Any] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) a : Tuple = np.array([0.4459, 0.4943, 0.4544, 0.6643, 0.5474, 0.4327, 0.5701, 0.5959, 0.5179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU' ) def lowercase_ ( self : int ): a : List[Any] = self.get_dummy_components() for name, module in components.items(): if hasattr(__snake_case , 'half' ): a : Any = module.half() a : Tuple = CycleDiffusionPipeline(**__snake_case ) a : Any = pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) a : str = self.get_dummy_inputs(__snake_case ) a : int = pipe(**__snake_case ) a : Optional[int] = output.images a : Tuple = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) a : int = np.array([0.3506, 0.4543, 0.446, 0.4575, 0.5195, 0.4155, 0.5273, 0.518, 0.4116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowercase_ ( self : List[Any] ): return super().test_save_load_local() @unittest.skip('non-deterministic pipeline' ) def lowercase_ ( self : Dict ): return super().test_inference_batch_single_identical() @skip_mps def lowercase_ ( self : int ): return super().test_dict_tuple_outputs_equivalent() @skip_mps def lowercase_ ( self : Dict ): return super().test_save_load_optional_components() @skip_mps def lowercase_ ( self : List[Any] ): return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class a__( unittest.TestCase ): def lowercase_ ( self : Tuple ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self : Optional[int] ): a : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) a : Optional[int] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy' ) a : List[str] = init_image.resize((5_12, 5_12) ) a : Dict = 'CompVis/stable-diffusion-v1-4' a : List[str] = DDIMScheduler.from_pretrained(__snake_case , subfolder='scheduler' ) a : Any = CycleDiffusionPipeline.from_pretrained( __snake_case , scheduler=__snake_case , safety_checker=__snake_case , torch_dtype=torch.floataa , revision='fp16' ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) pipe.enable_attention_slicing() a : Union[str, Any] = 'A black colored car' a : Optional[Any] = 'A blue colored car' a : int = torch.manual_seed(0 ) a : Optional[Any] = pipe( prompt=__snake_case , source_prompt=__snake_case , image=__snake_case , num_inference_steps=1_00 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=__snake_case , output_type='np' , ) a : Dict = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5e-1 def lowercase_ ( self : int ): a : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) a : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy' ) a : str = init_image.resize((5_12, 5_12) ) a : Optional[int] = 'CompVis/stable-diffusion-v1-4' a : Union[str, Any] = DDIMScheduler.from_pretrained(__snake_case , subfolder='scheduler' ) a : str = CycleDiffusionPipeline.from_pretrained(__snake_case , scheduler=__snake_case , safety_checker=__snake_case ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) pipe.enable_attention_slicing() a : Tuple = 'A black colored car' a : Tuple = 'A blue colored car' a : List[str] = torch.manual_seed(0 ) a : str = pipe( prompt=__snake_case , source_prompt=__snake_case , image=__snake_case , num_inference_steps=1_00 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=__snake_case , output_type='np' , ) a : Tuple = output.images assert np.abs(image - expected_image ).max() < 2e-2
96
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class snake_case__ ( snake_case_, unittest.TestCase ): _snake_case : Optional[Any] = KandinskyVaaImgaImgPipeline _snake_case : List[Any] = ["""image_embeds""", """negative_image_embeds""", """image"""] _snake_case : List[Any] = [ """image_embeds""", """negative_image_embeds""", """image""", ] _snake_case : Union[str, Any] = [ """generator""", """height""", """width""", """strength""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] _snake_case : Tuple = False @property def a__ ( self ): return 32 @property def a__ ( self ): return 32 @property def a__ ( self ): return self.time_input_dim @property def a__ ( self ): return self.time_input_dim * 4 @property def a__ ( self ): return 100 @property def a__ ( self ): torch.manual_seed(0 ) __a = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } __a = UNetaDConditionModel(**lowerCamelCase ) return model @property def a__ ( self ): return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def a__ ( self ): torch.manual_seed(0 ) __a = VQModel(**self.dummy_movq_kwargs ) return model def a__ ( self ): __a = self.dummy_unet __a = self.dummy_movq __a = { "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.0_0085, "beta_end": 0.012, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } __a = DDIMScheduler(**lowerCamelCase ) __a = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def a__ ( self , lowerCamelCase , lowerCamelCase=0 ): __a = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) __a = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( lowerCamelCase ) # create init_image __a = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) __a = image.cpu().permute(0 , 2 , 3 , 1 )[0] __a = Image.fromarray(np.uinta(lowerCamelCase ) ).convert("RGB" ).resize((256, 256) ) if str(lowerCamelCase ).startswith("mps" ): __a = torch.manual_seed(lowerCamelCase ) else: __a = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) __a = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def a__ ( self ): __a = "cpu" __a = self.get_dummy_components() __a = self.pipeline_class(**lowerCamelCase ) __a = pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) __a = pipe(**self.get_dummy_inputs(lowerCamelCase ) ) __a = output.images __a = pipe( **self.get_dummy_inputs(lowerCamelCase ) , return_dict=lowerCamelCase , )[0] __a = image[0, -3:, -3:, -1] __a = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __a = np.array( [0.619_9778, 0.6398_4406, 0.4614_5785, 0.6294_4984, 0.562_2215, 0.4730_6132, 0.4744_1456, 0.460_7606, 0.4871_9263] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class snake_case__ ( unittest.TestCase ): def a__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def a__ ( self ): __a = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_img2img_frog.npy" ) __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) __a = "A red cartoon frog, 4k" __a = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(lowerCamelCase ) __a = KandinskyVaaImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) __a = pipeline.to(lowerCamelCase ) pipeline.set_progress_bar_config(disable=lowerCamelCase ) __a = torch.Generator(device="cpu" ).manual_seed(0 ) __a , __a = pipe_prior( lowerCamelCase , generator=lowerCamelCase , num_inference_steps=5 , negative_prompt="" , ).to_tuple() __a = pipeline( image=lowerCamelCase , image_embeds=lowerCamelCase , negative_image_embeds=lowerCamelCase , generator=lowerCamelCase , num_inference_steps=100 , height=768 , width=768 , strength=0.2 , output_type="np" , ) __a = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowerCamelCase , lowerCamelCase )
261
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class snake_case__ ( snake_case_, snake_case_ ): @register_to_config def __init__( self , lowerCamelCase = 768 , ): super().__init__() __a = nn.Parameter(torch.zeros(1 , lowerCamelCase ) ) __a = nn.Parameter(torch.ones(1 , lowerCamelCase ) ) def a__ ( self , lowerCamelCase = None , lowerCamelCase = None , ): __a = nn.Parameter(self.mean.to(lowerCamelCase ).to(lowerCamelCase ) ) __a = nn.Parameter(self.std.to(lowerCamelCase ).to(lowerCamelCase ) ) return self def a__ ( self , lowerCamelCase ): __a = (embeds - self.mean) * 1.0 / self.std return embeds def a__ ( self , lowerCamelCase ): __a = (embeds * self.std) + self.mean return embeds
261
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) def _snake_case( SCREAMING_SNAKE_CASE__ : Tuple ) -> Union[str, Any]: '''simple docstring''' A__ = SwinConfig( embed_dim=192 , depths=(2, 2, 18, 2) , num_heads=(6, 12, 24, 48) , window_size=12 , out_features=['stage2', 'stage3', 'stage4'] , ) A__ = DetaConfig( backbone_config=SCREAMING_SNAKE_CASE__ , num_queries=900 , encoder_ffn_dim=2048 , decoder_ffn_dim=2048 , num_feature_levels=5 , assign_first_stage=SCREAMING_SNAKE_CASE__ , with_box_refine=SCREAMING_SNAKE_CASE__ , two_stage=SCREAMING_SNAKE_CASE__ , ) # set labels A__ = 'huggingface/label-files' if "o365" in model_name: A__ = 366 A__ = 'object365-id2label.json' else: A__ = 91 A__ = 'coco-detection-id2label.json' A__ = num_labels A__ = json.load(open(cached_download(hf_hub_url(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type='dataset' ) ) , 'r' ) ) A__ = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} A__ = idalabel A__ = {v: k for k, v in idalabel.items()} return config def _snake_case( SCREAMING_SNAKE_CASE__ : str ) -> int: '''simple docstring''' A__ = [] # stem # fmt: off rename_keys.append(('backbone.0.body.patch_embed.proj.weight', 'model.backbone.model.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('backbone.0.body.patch_embed.proj.bias', 'model.backbone.model.embeddings.patch_embeddings.projection.bias') ) rename_keys.append(('backbone.0.body.patch_embed.norm.weight', 'model.backbone.model.embeddings.norm.weight') ) rename_keys.append(('backbone.0.body.patch_embed.norm.bias', 'model.backbone.model.embeddings.norm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm1.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm1.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_bias_table', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_index', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.proj.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.attn.proj.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm2.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.norm2.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.weight', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.bias', f'model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.bias') ) if i < 3: rename_keys.append((f'backbone.0.body.layers.{i}.downsample.reduction.weight', f'model.backbone.model.encoder.layers.{i}.downsample.reduction.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.downsample.norm.weight', f'model.backbone.model.encoder.layers.{i}.downsample.norm.weight') ) rename_keys.append((f'backbone.0.body.layers.{i}.downsample.norm.bias', f'model.backbone.model.encoder.layers.{i}.downsample.norm.bias') ) rename_keys.append(('backbone.0.body.norm1.weight', 'model.backbone.model.hidden_states_norms.stage2.weight') ) rename_keys.append(('backbone.0.body.norm1.bias', 'model.backbone.model.hidden_states_norms.stage2.bias') ) rename_keys.append(('backbone.0.body.norm2.weight', 'model.backbone.model.hidden_states_norms.stage3.weight') ) rename_keys.append(('backbone.0.body.norm2.bias', 'model.backbone.model.hidden_states_norms.stage3.bias') ) rename_keys.append(('backbone.0.body.norm3.weight', 'model.backbone.model.hidden_states_norms.stage4.weight') ) rename_keys.append(('backbone.0.body.norm3.bias', 'model.backbone.model.hidden_states_norms.stage4.bias') ) # transformer encoder for i in range(config.encoder_layers ): rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.sampling_offsets.weight', f'model.encoder.layers.{i}.self_attn.sampling_offsets.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.sampling_offsets.bias', f'model.encoder.layers.{i}.self_attn.sampling_offsets.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.attention_weights.weight', f'model.encoder.layers.{i}.self_attn.attention_weights.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.attention_weights.bias', f'model.encoder.layers.{i}.self_attn.attention_weights.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.value_proj.weight', f'model.encoder.layers.{i}.self_attn.value_proj.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.value_proj.bias', f'model.encoder.layers.{i}.self_attn.value_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.output_proj.weight', f'model.encoder.layers.{i}.self_attn.output_proj.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.self_attn.output_proj.bias', f'model.encoder.layers.{i}.self_attn.output_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.weight', f'model.encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'model.encoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'model.encoder.layers.{i}.fc1.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'model.encoder.layers.{i}.fc1.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'model.encoder.layers.{i}.fc2.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'model.encoder.layers.{i}.fc2.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'model.encoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'model.encoder.layers.{i}.final_layer_norm.bias') ) # transformer decoder for i in range(config.decoder_layers ): rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.sampling_offsets.weight', f'model.decoder.layers.{i}.encoder_attn.sampling_offsets.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.sampling_offsets.bias', f'model.decoder.layers.{i}.encoder_attn.sampling_offsets.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.attention_weights.weight', f'model.decoder.layers.{i}.encoder_attn.attention_weights.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.attention_weights.bias', f'model.decoder.layers.{i}.encoder_attn.attention_weights.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.value_proj.weight', f'model.decoder.layers.{i}.encoder_attn.value_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.value_proj.bias', f'model.decoder.layers.{i}.encoder_attn.value_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.output_proj.weight', f'model.decoder.layers.{i}.encoder_attn.output_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.cross_attn.output_proj.bias', f'model.decoder.layers.{i}.encoder_attn.output_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.weight', f'model.decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'model.decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'model.decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'model.decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm2.weight', f'model.decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm2.bias', f'model.decoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'model.decoder.layers.{i}.fc1.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'model.decoder.layers.{i}.fc1.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'model.decoder.layers.{i}.fc2.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'model.decoder.layers.{i}.fc2.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'model.decoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'model.decoder.layers.{i}.final_layer_norm.bias') ) # fmt: on return rename_keys def _snake_case( SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : str ) -> Dict: '''simple docstring''' A__ = dct.pop(SCREAMING_SNAKE_CASE__ ) A__ = val def _snake_case( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : str ) -> Optional[Any]: '''simple docstring''' A__ = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): A__ = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) A__ = state_dict.pop(f'backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.weight' ) A__ = state_dict.pop(f'backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ = in_proj_weight[:dim, :] A__ = in_proj_bias[: dim] A__ = in_proj_weight[ dim : dim * 2, : ] A__ = in_proj_bias[ dim : dim * 2 ] A__ = in_proj_weight[ -dim :, : ] A__ = in_proj_bias[-dim :] # fmt: on def _snake_case( SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Any ) -> Dict: '''simple docstring''' A__ = config.d_model for i in range(config.decoder_layers ): # read in weights + bias of input projection layer of self-attention A__ = state_dict.pop(f'transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) A__ = state_dict.pop(f'transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict A__ = in_proj_weight[:hidden_size, :] A__ = in_proj_bias[:hidden_size] A__ = in_proj_weight[ hidden_size : hidden_size * 2, : ] A__ = in_proj_bias[hidden_size : hidden_size * 2] A__ = in_proj_weight[-hidden_size:, :] A__ = in_proj_bias[-hidden_size:] def _snake_case( ) -> Dict: '''simple docstring''' A__ = 'http://images.cocodataset.org/val2017/000000039769.jpg' A__ = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) return im @torch.no_grad() def _snake_case( SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Tuple ) -> List[str]: '''simple docstring''' A__ = get_deta_config(SCREAMING_SNAKE_CASE__ ) # load original state dict if model_name == "deta-swin-large": A__ = hf_hub_download(repo_id='nielsr/deta-checkpoints' , filename='adet_swin_ft.pth' ) elif model_name == "deta-swin-large-o365": A__ = hf_hub_download(repo_id='jozhang97/deta-swin-l-o365' , filename='deta_swin_pt_o365.pth' ) else: raise ValueError(f'Model name {model_name} not supported' ) A__ = torch.load(SCREAMING_SNAKE_CASE__ , map_location='cpu' )['model'] # original state dict for name, param in state_dict.items(): print(SCREAMING_SNAKE_CASE__ , param.shape ) # rename keys A__ = create_rename_keys(SCREAMING_SNAKE_CASE__ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) read_in_swin_q_k_v(SCREAMING_SNAKE_CASE__ , config.backbone_config ) read_in_decoder_q_k_v(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # fix some prefixes for key in state_dict.copy().keys(): if "transformer.decoder.class_embed" in key or "transformer.decoder.bbox_embed" in key: A__ = state_dict.pop(SCREAMING_SNAKE_CASE__ ) A__ = val if "input_proj" in key: A__ = state_dict.pop(SCREAMING_SNAKE_CASE__ ) A__ = val if "level_embed" in key or "pos_trans" in key or "pix_trans" in key or "enc_output" in key: A__ = state_dict.pop(SCREAMING_SNAKE_CASE__ ) A__ = val # finally, create HuggingFace model and load state dict A__ = DetaForObjectDetection(SCREAMING_SNAKE_CASE__ ) model.load_state_dict(SCREAMING_SNAKE_CASE__ ) model.eval() A__ = 'cuda' if torch.cuda.is_available() else 'cpu' model.to(SCREAMING_SNAKE_CASE__ ) # load image processor A__ = DetaImageProcessor(format='coco_detection' ) # verify our conversion on image A__ = prepare_img() A__ = processor(images=SCREAMING_SNAKE_CASE__ , return_tensors='pt' ) A__ = encoding['pixel_values'] A__ = model(pixel_values.to(SCREAMING_SNAKE_CASE__ ) ) # verify logits print('Logits:' , outputs.logits[0, :3, :3] ) print('Boxes:' , outputs.pred_boxes[0, :3, :3] ) if model_name == "deta-swin-large": A__ = torch.tensor( [[-7.6308, -2.8485, -5.3737], [-7.2037, -4.5505, -4.8027], [-7.2943, -4.2611, -4.6617]] ) A__ = torch.tensor([[0.4987, 0.4969, 0.9999], [0.2549, 0.5498, 0.4805], [0.5498, 0.2757, 0.0569]] ) elif model_name == "deta-swin-large-o365": A__ = torch.tensor( [[-8.0122, -3.5720, -4.9717], [-8.1547, -3.6886, -4.6389], [-7.6610, -3.6194, -5.0134]] ) A__ = torch.tensor([[0.2523, 0.5549, 0.4881], [0.7715, 0.4149, 0.4601], [0.5503, 0.2753, 0.0575]] ) assert torch.allclose(outputs.logits[0, :3, :3] , expected_logits.to(SCREAMING_SNAKE_CASE__ ) , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , expected_boxes.to(SCREAMING_SNAKE_CASE__ ) , atol=1E-4 ) print('Everything ok!' ) if pytorch_dump_folder_path: # Save model and processor logger.info(f'Saving PyTorch model and processor to {pytorch_dump_folder_path}...' ) Path(SCREAMING_SNAKE_CASE__ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) # Push to hub if push_to_hub: print('Pushing model and processor to hub...' ) model.push_to_hub(f'jozhang97/{model_name}' ) processor.push_to_hub(f'jozhang97/{model_name}' ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() parser.add_argument( "--model_name", type=str, default="deta-swin-large", choices=["deta-swin-large", "deta-swin-large-o365"], help="Name of the model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model.", ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) lowercase_ = parser.parse_args() convert_deta_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
282
def _snake_case( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : list[list[int]] ) -> int: '''simple docstring''' def update_area_of_max_square(SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 A__ = update_area_of_max_square(SCREAMING_SNAKE_CASE__ , col + 1 ) A__ = update_area_of_max_square(row + 1 , col + 1 ) A__ = update_area_of_max_square(row + 1 , SCREAMING_SNAKE_CASE__ ) if mat[row][col]: A__ = 1 + min([right, diagonal, down] ) A__ = max(largest_square_area[0] , SCREAMING_SNAKE_CASE__ ) return sub_problem_sol else: return 0 A__ = [0] update_area_of_max_square(0 , 0 ) return largest_square_area[0] def _snake_case( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : list[list[int]] ) -> int: '''simple docstring''' def update_area_of_max_square_using_dp_array( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : list[list[int]] ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] A__ = update_area_of_max_square_using_dp_array(SCREAMING_SNAKE_CASE__ , col + 1 , SCREAMING_SNAKE_CASE__ ) A__ = update_area_of_max_square_using_dp_array(row + 1 , col + 1 , SCREAMING_SNAKE_CASE__ ) A__ = update_area_of_max_square_using_dp_array(row + 1 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if mat[row][col]: A__ = 1 + min([right, diagonal, down] ) A__ = max(largest_square_area[0] , SCREAMING_SNAKE_CASE__ ) A__ = sub_problem_sol return sub_problem_sol else: return 0 A__ = [0] A__ = [[-1] * cols for _ in range(SCREAMING_SNAKE_CASE__ )] update_area_of_max_square_using_dp_array(0 , 0 , SCREAMING_SNAKE_CASE__ ) return largest_square_area[0] def _snake_case( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : list[list[int]] ) -> int: '''simple docstring''' A__ = [[0] * (cols + 1) for _ in range(rows + 1 )] A__ = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): A__ = dp_array[row][col + 1] A__ = dp_array[row + 1][col + 1] A__ = dp_array[row + 1][col] if mat[row][col] == 1: A__ = 1 + min(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) A__ = max(dp_array[row][col] , SCREAMING_SNAKE_CASE__ ) else: A__ = 0 return largest_square_area def _snake_case( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : list[list[int]] ) -> int: '''simple docstring''' A__ = [0] * (cols + 1) A__ = [0] * (cols + 1) A__ = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): A__ = current_row[col + 1] A__ = next_row[col + 1] A__ = next_row[col] if mat[row][col] == 1: A__ = 1 + min(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) A__ = max(current_row[col] , SCREAMING_SNAKE_CASE__ ) else: A__ = 0 A__ = current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
282
1
def _a ( a :int = 50 ) -> int: a = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f"""{solution() = }""")
0
import argparse import math import traceback import dateutil.parser as date_parser import requests def a ( A__ : Dict ) -> str: """simple docstring""" _lowercase ={} _lowercase =job['started_at'] _lowercase =job['completed_at'] _lowercase =date_parser.parse(A__ ) _lowercase =date_parser.parse(A__ ) _lowercase =round((end_datetime - start_datetime).total_seconds() / 60.0 ) _lowercase =start _lowercase =end _lowercase =duration_in_min return job_info def a ( A__ : Dict , A__ : str=None ) -> Tuple: """simple docstring""" _lowercase =None if token is not None: _lowercase ={'Accept': 'application/vnd.github+json', 'Authorization': F'''Bearer {token}'''} _lowercase =F'''https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100''' _lowercase =requests.get(A__ , headers=A__ ).json() _lowercase ={} try: job_time.update({job['name']: extract_time_from_single_job(A__ ) for job in result['jobs']} ) _lowercase =math.ceil((result['total_count'] - 100) / 100 ) for i in range(A__ ): _lowercase =requests.get(url + F'''&page={i + 2}''' , headers=A__ ).json() job_time.update({job['name']: extract_time_from_single_job(A__ ) for job in result['jobs']} ) return job_time except Exception: print(F'''Unknown error, could not fetch links:\n{traceback.format_exc()}''' ) return {} if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') lowercase_ = parser.parse_args() lowercase_ = get_job_time(args.workflow_run_id) lowercase_ = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(f"{k}: {v['duration']}")
205
0
"""simple docstring""" import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def __lowercase ( _UpperCamelCase ) ->List[str]: """simple docstring""" lowercase : int = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''_float_tensor''', '''decoder.output_projection.weight''', ] for k in ignore_keys: state_dict.pop(UpperCAmelCase_, UpperCAmelCase_ ) def __lowercase ( _UpperCamelCase ) ->Dict: """simple docstring""" lowercase , lowercase : List[str] = emb.weight.shape lowercase : Optional[Any] = nn.Linear(UpperCAmelCase_, UpperCAmelCase_, bias=UpperCAmelCase_ ) lowercase : Dict = emb.weight.data return lin_layer def __lowercase ( _UpperCamelCase, _UpperCamelCase="facebook/mbart-large-en-ro", _UpperCamelCase=False, _UpperCamelCase=False ) ->Dict: """simple docstring""" lowercase : List[str] = torch.load(UpperCAmelCase_, map_location='''cpu''' )['''model'''] remove_ignore_keys_(UpperCAmelCase_ ) lowercase : Union[str, Any] = state_dict['''encoder.embed_tokens.weight'''].shape[0] lowercase : Any = MBartConfig.from_pretrained(UpperCAmelCase_, vocab_size=UpperCAmelCase_ ) if mbart_aa and finetuned: lowercase : Optional[Any] = '''relu''' lowercase : str = state_dict['''decoder.embed_tokens.weight'''] lowercase : Dict = MBartForConditionalGeneration(UpperCAmelCase_ ) model.model.load_state_dict(UpperCAmelCase_ ) if finetuned: lowercase : int = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default='''facebook/mbart-large-cc25''', type=str, help='''Which huggingface architecture to use: mbart-large''', ) parser.add_argument('''--mbart_50''', action='''store_true''', help='''whether the model is mMART-50 checkpoint''') parser.add_argument('''--finetuned''', action='''store_true''', help='''whether the model is a fine-tuned checkpoint''') __a = parser.parse_args() __a = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
353
import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def __lowerCamelCase ( self ): lowercase : int = FlaxXLMRobertaModel.from_pretrained('''xlm-roberta-base''' ) lowercase : List[Any] = AutoTokenizer.from_pretrained('''xlm-roberta-base''' ) lowercase : Optional[int] = '''The dog is cute and lives in the garden house''' lowercase : List[str] = jnp.array([tokenizer.encode(SCREAMING_SNAKE_CASE__ )] ) lowercase : Any = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim lowercase : Union[str, Any] = jnp.array( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) lowercase : Any = model(SCREAMING_SNAKE_CASE__ )['''last_hidden_state'''] self.assertEqual(output.shape , SCREAMING_SNAKE_CASE__ ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , SCREAMING_SNAKE_CASE__ , atol=1E-3 ) )
173
0
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class _snake_case ( _snake_case ): SCREAMING_SNAKE_CASE__ = ['image_processor'] SCREAMING_SNAKE_CASE__ = 'SamImageProcessor' def __init__( self , _lowerCamelCase ): super().__init__(_lowerCamelCase ) a :Dict = self.image_processor a :str = -10 a :List[str] = self.image_processor.size['''longest_edge'''] def __call__( self , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase = None , **_lowerCamelCase , ): a :List[Any] = self.image_processor( _lowerCamelCase , return_tensors=_lowerCamelCase , **_lowerCamelCase , ) # pop arguments that are not used in the foward but used nevertheless a :Any = encoding_image_processor['''original_sizes'''] if hasattr(_lowerCamelCase , '''numpy''' ): # Checks if Torch or TF tensor a :Union[str, Any] = original_sizes.numpy() a , a , a :Optional[Any] = self._check_and_preprocess_points( input_points=_lowerCamelCase , input_labels=_lowerCamelCase , input_boxes=_lowerCamelCase , ) a :Optional[Any] = self._normalize_and_convert( _lowerCamelCase , _lowerCamelCase , input_points=_lowerCamelCase , input_labels=_lowerCamelCase , input_boxes=_lowerCamelCase , return_tensors=_lowerCamelCase , ) return encoding_image_processor def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase="pt" , ): if input_points is not None: if len(_lowerCamelCase ) != len(_lowerCamelCase ): a :Tuple = [ self._normalize_coordinates(self.target_size , _lowerCamelCase , original_sizes[0] ) for point in input_points ] else: a :str = [ self._normalize_coordinates(self.target_size , _lowerCamelCase , _lowerCamelCase ) for point, original_size in zip(_lowerCamelCase , _lowerCamelCase ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: a , a :Tuple = self._pad_points_and_labels(_lowerCamelCase , _lowerCamelCase ) a :List[str] = np.array(_lowerCamelCase ) if input_labels is not None: a :Tuple = np.array(_lowerCamelCase ) if input_boxes is not None: if len(_lowerCamelCase ) != len(_lowerCamelCase ): a :Dict = [ self._normalize_coordinates(self.target_size , _lowerCamelCase , original_sizes[0] , is_bounding_box=_lowerCamelCase ) for box in input_boxes ] else: a :str = [ self._normalize_coordinates(self.target_size , _lowerCamelCase , _lowerCamelCase , is_bounding_box=_lowerCamelCase ) for box, original_size in zip(_lowerCamelCase , _lowerCamelCase ) ] a :Union[str, Any] = np.array(_lowerCamelCase ) if input_boxes is not None: if return_tensors == "pt": a :Optional[Any] = torch.from_numpy(_lowerCamelCase ) # boxes batch size of 1 by default a :Optional[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": a :Dict = tf.convert_to_tensor(_lowerCamelCase ) # boxes batch size of 1 by default a :Tuple = tf.expand_dims(_lowerCamelCase , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({'''input_boxes''': input_boxes} ) if input_points is not None: if return_tensors == "pt": a :List[Any] = torch.from_numpy(_lowerCamelCase ) # point batch size of 1 by default a :Any = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": a :Tuple = tf.convert_to_tensor(_lowerCamelCase ) # point batch size of 1 by default a :int = tf.expand_dims(_lowerCamelCase , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({'''input_points''': input_points} ) if input_labels is not None: if return_tensors == "pt": a :Dict = torch.from_numpy(_lowerCamelCase ) # point batch size of 1 by default a :int = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": a :Union[str, Any] = tf.convert_to_tensor(_lowerCamelCase ) # point batch size of 1 by default a :Dict = tf.expand_dims(_lowerCamelCase , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({'''input_labels''': input_labels} ) return encoding_image_processor def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase ): a :List[Any] = max([point.shape[0] for point in input_points] ) a :Union[str, Any] = [] for i, point in enumerate(_lowerCamelCase ): if point.shape[0] != expected_nb_points: a :int = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) a :Union[str, Any] = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(_lowerCamelCase ) a :Any = processed_input_points return input_points, input_labels def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ): a , a :str = original_size a , a :Optional[Any] = self.image_processor._get_preprocess_shape(_lowerCamelCase , longest_edge=_lowerCamelCase ) a :List[Any] = deepcopy(_lowerCamelCase ).astype(_lowerCamelCase ) if is_bounding_box: a :Dict = coords.reshape(-1 , 2 , 2 ) a :Tuple = coords[..., 0] * (new_w / old_w) a :Dict = coords[..., 1] * (new_h / old_h) if is_bounding_box: a :Any = coords.reshape(-1 , 4 ) return coords def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase=None , ): if input_points is not None: if hasattr(_lowerCamelCase , '''numpy''' ): # Checks for TF or Torch tensor a :int = input_points.numpy().tolist() if not isinstance(_lowerCamelCase , _lowerCamelCase ) or not isinstance(input_points[0] , _lowerCamelCase ): raise ValueError('''Input points must be a list of list of floating points.''' ) a :Dict = [np.array(_lowerCamelCase ) for input_point in input_points] else: a :List[str] = None if input_labels is not None: if hasattr(_lowerCamelCase , '''numpy''' ): a :Optional[int] = input_labels.numpy().tolist() if not isinstance(_lowerCamelCase , _lowerCamelCase ) or not isinstance(input_labels[0] , _lowerCamelCase ): raise ValueError('''Input labels must be a list of list integers.''' ) a :List[Any] = [np.array(_lowerCamelCase ) for label in input_labels] else: a :Optional[Any] = None if input_boxes is not None: if hasattr(_lowerCamelCase , '''numpy''' ): a :int = input_boxes.numpy().tolist() if ( not isinstance(_lowerCamelCase , _lowerCamelCase ) or not isinstance(input_boxes[0] , _lowerCamelCase ) or not isinstance(input_boxes[0][0] , _lowerCamelCase ) ): raise ValueError('''Input boxes must be a list of list of list of floating points.''' ) a :Optional[int] = [np.array(_lowerCamelCase ).astype(np.floataa ) for box in input_boxes] else: a :Optional[int] = None return input_points, input_labels, input_boxes @property def SCREAMING_SNAKE_CASE__ ( self ): a :List[str] = self.image_processor.model_input_names return list(dict.fromkeys(_lowerCamelCase ) ) def SCREAMING_SNAKE_CASE__ ( self , *_lowerCamelCase , **_lowerCamelCase ): return self.image_processor.post_process_masks(*_lowerCamelCase , **_lowerCamelCase )
94
'''simple docstring''' from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch("""socket.socket""" ) @patch("""builtins.open""" ) def _UpperCAmelCase ( _lowerCamelCase : Dict , _lowerCamelCase : Optional[int] ) -> Union[str, Any]: # ===== initialization ===== _lowerCAmelCase : Tuple = Mock() _lowerCAmelCase : Any = conn, Mock() _lowerCAmelCase : Optional[Any] = iter([1, None] ) _lowerCAmelCase : str = lambda _lowerCamelCase : next(_lowerCamelCase ) # ===== invoke ===== send_file(filename="""mytext.txt""" , testing=_lowerCamelCase ) # ===== ensurance ===== sock.assert_called_once() sock.return_value.bind.assert_called_once() sock.return_value.listen.assert_called_once() sock.return_value.accept.assert_called_once() conn.recv.assert_called_once() file.return_value.__enter__.assert_called_once() file.return_value.__enter__.return_value.read.assert_called() conn.send.assert_called_once() conn.close.assert_called_once() sock.return_value.shutdown.assert_called_once() sock.return_value.close.assert_called_once()
309
0
from typing import Dict from transformers import EvalPrediction, HfArgumentParser, TrainingArguments, is_torch_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, get_torch_dist_unique_port, require_torch_multi_gpu, require_torch_neuroncore, ) from transformers.training_args import ParallelMode from transformers.utils import logging lowerCamelCase = logging.get_logger(__name__) if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset from transformers import Trainer class _a ( _lowercase): def __init__( self : Tuple , _SCREAMING_SNAKE_CASE : int = 101 )-> Tuple: lowerCAmelCase__ : Any = length def __len__( self : List[str] )-> Dict: return self.length def __getitem__( self : Union[str, Any] , _SCREAMING_SNAKE_CASE : Optional[Any] )-> int: return i class _a : def __call__( self : List[Any] , _SCREAMING_SNAKE_CASE : str )-> Tuple: return {"input_ids": torch.tensor(_SCREAMING_SNAKE_CASE ), "labels": torch.tensor(_SCREAMING_SNAKE_CASE )} class _a ( nn.Module): def __init__( self : str )-> str: super().__init__() # Add some (unused) params otherwise DDP will complain. lowerCAmelCase__ : str = nn.Linear(120 , 80 ) def UpperCAmelCase__( self : Union[str, Any] , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Tuple=None )-> Union[str, Any]: if labels is not None: return torch.tensor(0.0 , device=input_ids.device ), input_ids else: return input_ids class _a ( _lowercase): @require_torch_neuroncore def UpperCAmelCase__( self : str )-> Any: lowerCAmelCase__ : Tuple = F'--nproc_per_node=2\n --master_port={get_torch_dist_unique_port()}\n {self.test_file_dir}/test_trainer_distributed.py\n '.split() lowerCAmelCase__ : Any = self.get_auto_remove_tmp_dir() lowerCAmelCase__ : Optional[Any] = F'--output_dir {output_dir}'.split() lowerCAmelCase__ : int = ['''torchrun'''] + distributed_args + args execute_subprocess_async(_SCREAMING_SNAKE_CASE , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call class _a ( _lowercase): @require_torch_multi_gpu def UpperCAmelCase__( self : Optional[Any] )-> Tuple: lowerCAmelCase__ : Union[str, Any] = F'--nproc_per_node={torch.cuda.device_count()}\n --master_port={get_torch_dist_unique_port()}\n {self.test_file_dir}/test_trainer_distributed.py\n '.split() lowerCAmelCase__ : List[Any] = self.get_auto_remove_tmp_dir() lowerCAmelCase__ : List[str] = F'--output_dir {output_dir}'.split() lowerCAmelCase__ : str = ['''torchrun'''] + distributed_args + args execute_subprocess_async(_SCREAMING_SNAKE_CASE , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call if __name__ == "__main__": # The script below is meant to be run under torch.distributed, on a machine with multiple GPUs: # # PYTHONPATH="src" python -m torch.distributed.run --nproc_per_node 2 --output_dir output_dir ./tests/test_trainer_distributed.py lowerCamelCase = HfArgumentParser((TrainingArguments,)) lowerCamelCase = parser.parse_args_into_dataclasses()[0] logger.warning( f'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}, ''' f'''distributed training: {training_args.parallel_mode != ParallelMode.NOT_DISTRIBUTED}''' ) # Essentially, what we want to verify in the distributed case is that we get all samples back, # in the right order. (this is crucial for prediction for instance) for dataset_length in [101, 40, 7]: lowerCamelCase = DummyDataset(dataset_length) def lowerCamelCase_ ( _a ): """simple docstring""" lowerCAmelCase__ : Dict = list(range(len(_a ) ) ) lowerCAmelCase__ : Union[str, Any] = p.predictions.tolist() == sequential and p.label_ids.tolist() == sequential if not success and training_args.local_rank == 0: logger.warning( '''Predictions and/or labels do not match expected results:\n - predictions: ''' f'{p.predictions.tolist()}\n - labels: {p.label_ids.tolist()}\n - expected: {sequential}' ) return {"success": success} lowerCamelCase = Trainer( model=DummyModel(), args=training_args, data_collator=DummyDataCollator(), eval_dataset=dataset, compute_metrics=compute_metrics, ) lowerCamelCase = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) lowerCamelCase = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) lowerCamelCase = 2 lowerCamelCase = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) lowerCamelCase = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) lowerCamelCase = None
368
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def lowerCamelCase_ ( _a ): """simple docstring""" lowerCAmelCase__ : Any = filter(lambda _a : p.requires_grad , model.parameters() ) lowerCAmelCase__ : str = sum([np.prod(p.size() ) for p in model_parameters] ) return params lowerCamelCase = logging.getLogger(__name__) def lowerCamelCase_ ( _a , _a ): """simple docstring""" if metric == "rouge2": lowerCAmelCase__ : Optional[int] = '''{val_avg_rouge2:.4f}-{step_count}''' elif metric == "bleu": lowerCAmelCase__ : Optional[int] = '''{val_avg_bleu:.4f}-{step_count}''' elif metric == "em": lowerCAmelCase__ : List[Any] = '''{val_avg_em:.4f}-{step_count}''' else: raise NotImplementedError( f'seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this' ''' function.''' ) lowerCAmelCase__ : Dict = ModelCheckpoint( dirpath=_a , filename=_a , monitor=f'val_{metric}' , mode='''max''' , save_top_k=3 , every_n_epochs=1 , ) return checkpoint_callback def lowerCamelCase_ ( _a , _a ): """simple docstring""" return EarlyStopping( monitor=f'val_{metric}' , mode='''min''' if '''loss''' in metric else '''max''' , patience=_a , verbose=_a , ) class _a ( pl.Callback): def UpperCAmelCase__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Any )-> Optional[int]: lowerCAmelCase__ : Dict = {F'lr_group_{i}': param['''lr'''] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(_SCREAMING_SNAKE_CASE ) @rank_zero_only def UpperCAmelCase__( self : List[str] , _SCREAMING_SNAKE_CASE : pl.Trainer , _SCREAMING_SNAKE_CASE : pl.LightningModule , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : List[str]=True )-> None: logger.info(F'***** {type_path} results at step {trainer.global_step:05d} *****' ) lowerCAmelCase__ : List[Any] = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['''log''', '''progress_bar''', '''preds''']} ) # Log results lowerCAmelCase__ : List[Any] = Path(pl_module.hparams.output_dir ) if type_path == "test": lowerCAmelCase__ : Optional[int] = od / '''test_results.txt''' lowerCAmelCase__ : Tuple = od / '''test_generations.txt''' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. lowerCAmelCase__ : int = od / F'{type_path}_results/{trainer.global_step:05d}.txt' lowerCAmelCase__ : int = od / F'{type_path}_generations/{trainer.global_step:05d}.txt' results_file.parent.mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) generations_file.parent.mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE , '''a+''' ) as writer: for key in sorted(_SCREAMING_SNAKE_CASE ): if key in ["log", "progress_bar", "preds"]: continue lowerCAmelCase__ : Optional[int] = metrics[key] if isinstance(_SCREAMING_SNAKE_CASE , torch.Tensor ): lowerCAmelCase__ : List[str] = val.item() lowerCAmelCase__ : List[str] = F'{key}: {val:.6f}\n' writer.write(_SCREAMING_SNAKE_CASE ) if not save_generations: return if "preds" in metrics: lowerCAmelCase__ : Dict = '''\n'''.join(metrics['''preds'''] ) generations_file.open('''w+''' ).write(_SCREAMING_SNAKE_CASE ) @rank_zero_only def UpperCAmelCase__( self : str , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Optional[int] )-> Optional[int]: try: lowerCAmelCase__ : Tuple = pl_module.model.model.num_parameters() except AttributeError: lowerCAmelCase__ : Optional[Any] = pl_module.model.num_parameters() lowerCAmelCase__ : Dict = count_trainable_parameters(_SCREAMING_SNAKE_CASE ) # mp stands for million parameters trainer.logger.log_metrics({'''n_params''': npars, '''mp''': npars / 1E6, '''grad_mp''': n_trainable_pars / 1E6} ) @rank_zero_only def UpperCAmelCase__( self : List[Any] , _SCREAMING_SNAKE_CASE : pl.Trainer , _SCREAMING_SNAKE_CASE : pl.LightningModule )-> Optional[Any]: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , '''test''' ) @rank_zero_only def UpperCAmelCase__( self : Optional[int] , _SCREAMING_SNAKE_CASE : pl.Trainer , _SCREAMING_SNAKE_CASE : List[Any] )-> List[Any]: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
211
0
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BlipaProcessor, BlipImageProcessor, GPTaTokenizer, PreTrainedTokenizerFast @require_vision class snake_case ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : List[str] ): '''simple docstring''' __A = tempfile.mkdtemp() __A = BlipImageProcessor() __A = GPTaTokenizer.from_pretrained('''hf-internal-testing/tiny-random-GPT2Model''' ) __A = BlipaProcessor(_lowerCamelCase, _lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def _SCREAMING_SNAKE_CASE ( self : Tuple, **_lowerCamelCase : List[Any] ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname, **_lowerCamelCase ).tokenizer def _SCREAMING_SNAKE_CASE ( self : Union[str, Any], **_lowerCamelCase : str ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname, **_lowerCamelCase ).image_processor def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def _SCREAMING_SNAKE_CASE ( self : int ): '''simple docstring''' __A = [np.random.randint(2_55, size=(3, 30, 4_00), dtype=np.uinta )] __A = [Image.fromarray(np.moveaxis(_lowerCamelCase, 0, -1 ) ) for x in image_inputs] return image_inputs def _SCREAMING_SNAKE_CASE ( self : List[str] ): '''simple docstring''' __A = BlipaProcessor(tokenizer=self.get_tokenizer(), image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __A = self.get_tokenizer(bos_token='''(BOS)''', eos_token='''(EOS)''' ) __A = self.get_image_processor(do_normalize=_lowerCamelCase, padding_value=1.0 ) __A = BlipaProcessor.from_pretrained( self.tmpdirname, bos_token='''(BOS)''', eos_token='''(EOS)''', do_normalize=_lowerCamelCase, padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer, _lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, _lowerCamelCase ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): '''simple docstring''' __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipaProcessor(tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase ) __A = self.prepare_image_inputs() __A = image_processor(_lowerCamelCase, return_tensors='''np''' ) __A = processor(images=_lowerCamelCase, return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum(), input_processor[key].sum(), delta=1e-2 ) def _SCREAMING_SNAKE_CASE ( self : Tuple ): '''simple docstring''' __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipaProcessor(tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase ) __A = '''lower newer''' __A = processor(text=_lowerCamelCase ) __A = tokenizer(_lowerCamelCase, return_token_type_ids=_lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key], encoded_processor[key] ) def _SCREAMING_SNAKE_CASE ( self : str ): '''simple docstring''' __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipaProcessor(tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase ) __A = '''lower newer''' __A = self.prepare_image_inputs() __A = processor(text=_lowerCamelCase, images=_lowerCamelCase ) self.assertListEqual(list(inputs.keys() ), ['''pixel_values''', '''input_ids''', '''attention_mask'''] ) # test if it raises when no input is passed with pytest.raises(_lowerCamelCase ): processor() def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): '''simple docstring''' __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipaProcessor(tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase ) __A = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __A = processor.batch_decode(_lowerCamelCase ) __A = tokenizer.batch_decode(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase, _lowerCamelCase ) def _SCREAMING_SNAKE_CASE ( self : Dict ): '''simple docstring''' __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipaProcessor(tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase ) __A = '''lower newer''' __A = self.prepare_image_inputs() __A = processor(text=_lowerCamelCase, images=_lowerCamelCase ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ), ['''pixel_values''', '''input_ids''', '''attention_mask'''] )
266
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class snake_case ( unittest.TestCase ): '''simple docstring''' def __init__( self : List[Any], _lowerCamelCase : int, _lowerCamelCase : List[Any]=7, _lowerCamelCase : int=3, _lowerCamelCase : Optional[Any]=18, _lowerCamelCase : Any=30, _lowerCamelCase : str=4_00, _lowerCamelCase : int=True, _lowerCamelCase : Union[str, Any]=None, _lowerCamelCase : str=True, ): '''simple docstring''' __A = size if size is not None else {'''height''': 18, '''width''': 18} __A = parent __A = batch_size __A = num_channels __A = image_size __A = min_resolution __A = max_resolution __A = do_resize __A = size __A = apply_ocr def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): '''simple docstring''' return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class snake_case ( _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' A_ : Optional[int] = LayoutLMvaImageProcessor if is_pytesseract_available() else None def _SCREAMING_SNAKE_CASE ( self : Dict ): '''simple docstring''' __A = LayoutLMvaImageProcessingTester(self ) @property def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def _SCREAMING_SNAKE_CASE ( self : int ): '''simple docstring''' __A = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase, '''do_resize''' ) ) self.assertTrue(hasattr(_lowerCamelCase, '''size''' ) ) self.assertTrue(hasattr(_lowerCamelCase, '''apply_ocr''' ) ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): '''simple docstring''' __A = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size, {'''height''': 18, '''width''': 18} ) __A = self.image_processing_class.from_dict(self.image_processor_dict, size=42 ) self.assertEqual(image_processor.size, {'''height''': 42, '''width''': 42} ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' pass def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' # Initialize image_processing __A = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __A = prepare_image_inputs(self.image_processor_tester, equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase, Image.Image ) # Test not batched input __A = image_processing(image_inputs[0], return_tensors='''pt''' ) self.assertEqual( encoding.pixel_values.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ), ) self.assertIsInstance(encoding.words, _lowerCamelCase ) self.assertIsInstance(encoding.boxes, _lowerCamelCase ) # Test batched __A = image_processing(_lowerCamelCase, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ), ) def _SCREAMING_SNAKE_CASE ( self : Any ): '''simple docstring''' # Initialize image_processing __A = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __A = prepare_image_inputs(self.image_processor_tester, equal_resolution=_lowerCamelCase, numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase, np.ndarray ) # Test not batched input __A = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ), ) # Test batched __A = image_processing(_lowerCamelCase, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ), ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' # Initialize image_processing __A = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __A = prepare_image_inputs(self.image_processor_tester, equal_resolution=_lowerCamelCase, torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase, torch.Tensor ) # Test not batched input __A = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ), ) # Test batched __A = image_processing(_lowerCamelCase, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ), ) def _SCREAMING_SNAKE_CASE ( self : List[str] ): '''simple docstring''' # with apply_OCR = True __A = LayoutLMvaImageProcessor() from datasets import load_dataset __A = load_dataset('''hf-internal-testing/fixtures_docvqa''', split='''test''' ) __A = Image.open(ds[0]['''file'''] ).convert('''RGB''' ) __A = image_processing(_lowerCamelCase, return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape, (1, 3, 2_24, 2_24) ) self.assertEqual(len(encoding.words ), len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 __A = [['''11:14''', '''to''', '''11:39''', '''a.m''', '''11:39''', '''to''', '''11:44''', '''a.m.''', '''11:44''', '''a.m.''', '''to''', '''12:25''', '''p.m.''', '''12:25''', '''to''', '''12:58''', '''p.m.''', '''12:58''', '''to''', '''4:00''', '''p.m.''', '''2:00''', '''to''', '''5:00''', '''p.m.''', '''Coffee''', '''Break''', '''Coffee''', '''will''', '''be''', '''served''', '''for''', '''men''', '''and''', '''women''', '''in''', '''the''', '''lobby''', '''adjacent''', '''to''', '''exhibit''', '''area.''', '''Please''', '''move''', '''into''', '''exhibit''', '''area.''', '''(Exhibits''', '''Open)''', '''TRRF''', '''GENERAL''', '''SESSION''', '''(PART''', '''|)''', '''Presiding:''', '''Lee''', '''A.''', '''Waller''', '''TRRF''', '''Vice''', '''President''', '''“Introductory''', '''Remarks”''', '''Lee''', '''A.''', '''Waller,''', '''TRRF''', '''Vice''', '''Presi-''', '''dent''', '''Individual''', '''Interviews''', '''with''', '''TRRF''', '''Public''', '''Board''', '''Members''', '''and''', '''Sci-''', '''entific''', '''Advisory''', '''Council''', '''Mem-''', '''bers''', '''Conducted''', '''by''', '''TRRF''', '''Treasurer''', '''Philip''', '''G.''', '''Kuehn''', '''to''', '''get''', '''answers''', '''which''', '''the''', '''public''', '''refrigerated''', '''warehousing''', '''industry''', '''is''', '''looking''', '''for.''', '''Plus''', '''questions''', '''from''', '''the''', '''floor.''', '''Dr.''', '''Emil''', '''M.''', '''Mrak,''', '''University''', '''of''', '''Cal-''', '''ifornia,''', '''Chairman,''', '''TRRF''', '''Board;''', '''Sam''', '''R.''', '''Cecil,''', '''University''', '''of''', '''Georgia''', '''College''', '''of''', '''Agriculture;''', '''Dr.''', '''Stanley''', '''Charm,''', '''Tufts''', '''University''', '''School''', '''of''', '''Medicine;''', '''Dr.''', '''Robert''', '''H.''', '''Cotton,''', '''ITT''', '''Continental''', '''Baking''', '''Company;''', '''Dr.''', '''Owen''', '''Fennema,''', '''University''', '''of''', '''Wis-''', '''consin;''', '''Dr.''', '''Robert''', '''E.''', '''Hardenburg,''', '''USDA.''', '''Questions''', '''and''', '''Answers''', '''Exhibits''', '''Open''', '''Capt.''', '''Jack''', '''Stoney''', '''Room''', '''TRRF''', '''Scientific''', '''Advisory''', '''Council''', '''Meeting''', '''Ballroom''', '''Foyer''']] # noqa: E231 __A = [[[1_41, 57, 2_14, 69], [2_28, 58, 2_52, 69], [1_41, 75, 2_16, 88], [2_30, 79, 2_80, 88], [1_42, 2_60, 2_18, 2_73], [2_30, 2_61, 2_55, 2_73], [1_43, 2_79, 2_18, 2_90], [2_31, 2_82, 2_90, 2_91], [1_43, 3_42, 2_18, 3_54], [2_31, 3_45, 2_89, 3_55], [2_02, 3_62, 2_27, 3_73], [1_43, 3_79, 2_20, 3_92], [2_31, 3_82, 2_91, 3_94], [1_44, 7_14, 2_20, 7_26], [2_31, 7_15, 2_56, 7_26], [1_44, 7_32, 2_20, 7_45], [2_32, 7_36, 2_91, 7_47], [1_44, 7_69, 2_18, 7_82], [2_31, 7_70, 2_56, 7_82], [1_41, 7_88, 2_02, 8_01], [2_15, 7_91, 2_74, 8_04], [1_43, 8_26, 2_04, 8_38], [2_15, 8_26, 2_40, 8_38], [1_42, 8_44, 2_02, 8_57], [2_15, 8_47, 2_74, 8_59], [3_34, 57, 4_27, 69], [4_40, 57, 5_22, 69], [3_69, 75, 4_61, 88], [4_69, 75, 5_16, 88], [5_28, 76, 5_62, 88], [5_70, 76, 6_67, 88], [6_75, 75, 7_11, 87], [7_21, 79, 7_78, 88], [7_89, 75, 8_40, 88], [3_69, 97, 4_70, 1_07], [4_84, 94, 5_07, 1_06], [5_18, 94, 5_62, 1_07], [5_76, 94, 6_55, 1_10], [6_68, 94, 7_92, 1_09], [8_04, 95, 8_29, 1_07], [3_69, 1_13, 4_65, 1_25], [4_77, 1_16, 5_47, 1_25], [5_62, 1_13, 6_58, 1_25], [6_71, 1_16, 7_48, 1_25], [7_61, 1_13, 8_11, 1_25], [3_69, 1_31, 4_65, 1_43], [4_77, 1_33, 5_48, 1_43], [5_63, 1_30, 6_98, 1_45], [7_10, 1_30, 8_02, 1_46], [3_36, 1_71, 4_12, 1_83], [4_23, 1_71, 5_72, 1_83], [5_82, 1_70, 7_16, 1_84], [7_28, 1_71, 8_17, 1_87], [8_29, 1_71, 8_44, 1_86], [3_38, 1_97, 4_82, 2_12], [5_07, 1_96, 5_57, 2_09], [5_69, 1_96, 5_95, 2_08], [6_10, 1_96, 7_02, 2_09], [5_05, 2_14, 5_83, 2_26], [5_95, 2_14, 6_56, 2_27], [6_70, 2_15, 8_07, 2_27], [3_35, 2_59, 5_43, 2_74], [5_56, 2_59, 7_08, 2_72], [3_72, 2_79, 4_22, 2_91], [4_35, 2_79, 4_60, 2_91], [4_74, 2_79, 5_74, 2_92], [5_87, 2_78, 6_64, 2_91], [6_76, 2_78, 7_38, 2_91], [7_51, 2_79, 8_34, 2_91], [3_72, 2_98, 4_34, 3_10], [3_35, 3_41, 4_83, 3_54], [4_97, 3_41, 6_55, 3_54], [6_67, 3_41, 7_28, 3_54], [7_40, 3_41, 8_25, 3_54], [3_35, 3_60, 4_30, 3_72], [4_42, 3_60, 5_34, 3_72], [5_45, 3_59, 6_87, 3_72], [6_97, 3_60, 7_54, 3_72], [7_65, 3_60, 8_23, 3_73], [3_34, 3_78, 4_28, 3_91], [4_40, 3_78, 5_77, 3_94], [5_90, 3_78, 7_05, 3_91], [7_20, 3_78, 8_01, 3_91], [3_34, 3_97, 4_00, 4_09], [3_70, 4_16, 5_29, 4_29], [5_44, 4_16, 5_76, 4_32], [5_87, 4_16, 6_65, 4_28], [6_77, 4_16, 8_14, 4_29], [3_72, 4_35, 4_52, 4_50], [4_65, 4_34, 4_95, 4_47], [5_11, 4_34, 6_00, 4_47], [6_11, 4_36, 6_37, 4_47], [6_49, 4_36, 6_94, 4_51], [7_05, 4_38, 8_24, 4_47], [3_69, 4_53, 4_52, 4_66], [4_64, 4_54, 5_09, 4_66], [5_22, 4_53, 6_11, 4_69], [6_25, 4_53, 7_92, 4_69], [3_70, 4_72, 5_56, 4_88], [5_70, 4_72, 6_84, 4_87], [6_97, 4_72, 7_18, 4_85], [7_32, 4_72, 8_35, 4_88], [3_69, 4_90, 4_11, 5_03], [4_25, 4_90, 4_84, 5_03], [4_96, 4_90, 6_35, 5_06], [6_45, 4_90, 7_07, 5_03], [7_18, 4_91, 7_61, 5_03], [7_71, 4_90, 8_40, 5_03], [3_36, 5_10, 3_74, 5_21], [3_88, 5_10, 4_47, 5_22], [4_60, 5_10, 4_89, 5_21], [5_03, 5_10, 5_80, 5_22], [5_92, 5_09, 7_36, 5_25], [7_45, 5_09, 7_70, 5_22], [7_81, 5_09, 8_40, 5_22], [3_38, 5_28, 4_34, 5_41], [4_48, 5_28, 5_96, 5_41], [6_09, 5_27, 6_87, 5_40], [7_00, 5_28, 7_92, 5_41], [3_36, 5_46, 3_97, 5_59], [4_07, 5_46, 4_31, 5_59], [4_43, 5_46, 5_25, 5_60], [5_37, 5_46, 6_80, 5_62], [6_88, 5_46, 7_14, 5_59], [7_22, 5_46, 8_37, 5_62], [3_36, 5_65, 4_49, 5_81], [4_61, 5_65, 4_85, 5_77], [4_97, 5_65, 6_65, 5_81], [6_81, 5_65, 7_18, 5_77], [7_32, 5_65, 8_37, 5_80], [3_37, 5_84, 4_38, 5_97], [4_52, 5_83, 5_21, 5_96], [5_35, 5_84, 6_77, 5_99], [6_90, 5_83, 7_87, 5_96], [8_01, 5_83, 8_25, 5_96], [3_38, 6_02, 4_78, 6_15], [4_92, 6_02, 5_30, 6_14], [5_43, 6_02, 6_38, 6_15], [6_50, 6_02, 6_76, 6_14], [6_88, 6_02, 7_88, 6_15], [8_02, 6_02, 8_43, 6_14], [3_37, 6_21, 5_02, 6_33], [5_16, 6_21, 6_15, 6_37], [6_29, 6_21, 7_74, 6_36], [7_89, 6_21, 8_27, 6_33], [3_37, 6_39, 4_18, 6_52], [4_32, 6_40, 5_71, 6_53], [5_87, 6_39, 7_31, 6_55], [7_43, 6_39, 7_69, 6_52], [7_80, 6_39, 8_41, 6_52], [3_38, 6_58, 4_40, 6_73], [4_55, 6_58, 4_91, 6_70], [5_08, 6_58, 6_02, 6_71], [6_16, 6_58, 6_38, 6_70], [6_54, 6_58, 8_35, 6_74], [3_37, 6_77, 4_29, 6_89], [3_37, 7_14, 4_82, 7_26], [4_95, 7_14, 5_48, 7_26], [5_61, 7_14, 6_83, 7_26], [3_38, 7_70, 4_61, 7_82], [4_74, 7_69, 5_54, 7_85], [4_89, 7_88, 5_62, 8_03], [5_76, 7_88, 6_43, 8_01], [6_56, 7_87, 7_51, 8_04], [7_64, 7_88, 8_44, 8_01], [3_34, 8_25, 4_21, 8_38], [4_30, 8_24, 5_74, 8_38], [5_84, 8_24, 7_23, 8_41], [3_35, 8_44, 4_50, 8_57], [4_64, 8_43, 5_83, 8_60], [6_28, 8_62, 7_55, 8_75], [7_69, 8_61, 8_48, 8_78]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words, _lowerCamelCase ) self.assertListEqual(encoding.boxes, _lowerCamelCase ) # with apply_OCR = False __A = LayoutLMvaImageProcessor(apply_ocr=_lowerCamelCase ) __A = image_processing(_lowerCamelCase, return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape, (1, 3, 2_24, 2_24) )
266
1
import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def UpperCAmelCase_ ( _A , _A , _A , _A , _A ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE_ ) as metadata_file: SCREAMING_SNAKE_CASE__ = json.load(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE__ = LukeConfig(use_entity_aware_attention=SCREAMING_SNAKE_CASE_ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path SCREAMING_SNAKE_CASE__ = torch.load(SCREAMING_SNAKE_CASE_ , map_location='''cpu''' )['''module'''] # Load the entity vocab file SCREAMING_SNAKE_CASE__ = load_original_entity_vocab(SCREAMING_SNAKE_CASE_ ) # add an entry for [MASK2] SCREAMING_SNAKE_CASE__ = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 SCREAMING_SNAKE_CASE__ = XLMRobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks SCREAMING_SNAKE_CASE__ = AddedToken('''<ent>''' , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE__ = AddedToken('''<ent2>''' , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'''Saving tokenizer to {pytorch_dump_folder_path}''' ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE_ ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , '''tokenizer_config.json''' ) , '''r''' ) as f: SCREAMING_SNAKE_CASE__ = json.load(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE__ = '''MLukeTokenizer''' with open(os.path.join(SCREAMING_SNAKE_CASE_ , '''tokenizer_config.json''' ) , '''w''' ) as f: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , MLukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE__ = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) # Initialize the embeddings of the special tokens SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(['''@'''] )[0] SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(['''#'''] )[0] SCREAMING_SNAKE_CASE__ = state_dict['''embeddings.word_embeddings.weight'''] SCREAMING_SNAKE_CASE__ = word_emb[ent_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE__ = word_emb[enta_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE__ = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: SCREAMING_SNAKE_CASE__ = state_dict[bias_name] SCREAMING_SNAKE_CASE__ = decoder_bias[ent_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE__ = decoder_bias[enta_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE__ = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: SCREAMING_SNAKE_CASE__ = F'''encoder.layer.{layer_index}.attention.self.''' SCREAMING_SNAKE_CASE__ = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE__ = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE__ = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks SCREAMING_SNAKE_CASE__ = state_dict['''entity_embeddings.entity_embeddings.weight'''] SCREAMING_SNAKE_CASE__ = entity_emb[entity_vocab['''[MASK]''']].unsqueeze(0 ) SCREAMING_SNAKE_CASE__ = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' SCREAMING_SNAKE_CASE__ = state_dict['''entity_predictions.bias'''] SCREAMING_SNAKE_CASE__ = entity_prediction_bias[entity_vocab['''[MASK]''']].unsqueeze(0 ) SCREAMING_SNAKE_CASE__ = torch.cat([entity_prediction_bias, entity_mask_bias] ) SCREAMING_SNAKE_CASE__ = LukeForMaskedLM(config=SCREAMING_SNAKE_CASE_ ).eval() state_dict.pop('''entity_predictions.decoder.weight''' ) state_dict.pop('''lm_head.decoder.weight''' ) state_dict.pop('''lm_head.decoder.bias''' ) SCREAMING_SNAKE_CASE__ = OrderedDict() for key, value in state_dict.items(): if not (key.startswith('''lm_head''' ) or key.startswith('''entity_predictions''' )): SCREAMING_SNAKE_CASE__ = state_dict[key] else: SCREAMING_SNAKE_CASE__ = state_dict[key] SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = model.load_state_dict(SCREAMING_SNAKE_CASE_ , strict=SCREAMING_SNAKE_CASE_ ) if set(SCREAMING_SNAKE_CASE_ ) != {"luke.embeddings.position_ids"}: raise ValueError(F'''Unexpected unexpected_keys: {unexpected_keys}''' ) if set(SCREAMING_SNAKE_CASE_ ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(F'''Unexpected missing_keys: {missing_keys}''' ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs SCREAMING_SNAKE_CASE__ = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ , task='''entity_classification''' ) SCREAMING_SNAKE_CASE__ = '''ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan).''' SCREAMING_SNAKE_CASE__ = (0, 9) SCREAMING_SNAKE_CASE__ = tokenizer(SCREAMING_SNAKE_CASE_ , entity_spans=[span] , return_tensors='''pt''' ) SCREAMING_SNAKE_CASE__ = model(**SCREAMING_SNAKE_CASE_ ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base SCREAMING_SNAKE_CASE__ = torch.Size((1, 33, 7_68) ) SCREAMING_SNAKE_CASE__ = torch.tensor([[0.0_8_9_2, 0.0_5_9_6, -0.2_8_1_9], [0.0_1_3_4, 0.1_1_9_9, 0.0_5_7_3], [-0.0_1_6_9, 0.0_9_2_7, 0.0_6_4_4]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}''' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base SCREAMING_SNAKE_CASE__ = torch.Size((1, 1, 7_68) ) SCREAMING_SNAKE_CASE__ = torch.tensor([[-0.1_4_8_2, 0.0_6_0_9, 0.0_3_2_2]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( F'''Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is''' F''' {expected_shape}''' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=1e-4 ): raise ValueError # Verify masked word/entity prediction SCREAMING_SNAKE_CASE__ = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE__ = '''Tokyo is the capital of <mask>.''' SCREAMING_SNAKE_CASE__ = (24, 30) SCREAMING_SNAKE_CASE__ = tokenizer(SCREAMING_SNAKE_CASE_ , entity_spans=[span] , return_tensors='''pt''' ) SCREAMING_SNAKE_CASE__ = model(**SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE__ = encoding['''input_ids'''][0].tolist() SCREAMING_SNAKE_CASE__ = input_ids.index(tokenizer.convert_tokens_to_ids('''<mask>''' ) ) SCREAMING_SNAKE_CASE__ = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE__ = outputs.entity_logits[0][0].argmax().item() SCREAMING_SNAKE_CASE__ = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith('''en:''' )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(SCREAMING_SNAKE_CASE_ ) ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase_ ( _A ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = ['''[MASK]''', '''[PAD]''', '''[UNK]'''] SCREAMING_SNAKE_CASE__ = [json.loads(SCREAMING_SNAKE_CASE_ ) for line in open(SCREAMING_SNAKE_CASE_ )] SCREAMING_SNAKE_CASE__ = {} for entry in data: SCREAMING_SNAKE_CASE__ = entry['''id'''] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: SCREAMING_SNAKE_CASE__ = entity_id break SCREAMING_SNAKE_CASE__ = F'''{language}:{entity_name}''' SCREAMING_SNAKE_CASE__ = entity_id return new_mapping if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) _SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
366
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Tuple = { '''transfo-xl-wt103''': '''https://huggingface.co/transfo-xl-wt103/resolve/main/config.json''', } class UpperCAmelCase__ ( A__ ): """simple docstring""" a = "transfo-xl" a = ["mems"] a = { "n_token": "vocab_size", "hidden_size": "d_model", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Any , __lowerCamelCase : int=26_7735 , __lowerCamelCase : Any=[2_0000, 4_0000, 20_0000] , __lowerCamelCase : Dict=1024 , __lowerCamelCase : Optional[int]=1024 , __lowerCamelCase : List[Any]=16 , __lowerCamelCase : Union[str, Any]=64 , __lowerCamelCase : Dict=4096 , __lowerCamelCase : int=4 , __lowerCamelCase : Dict=False , __lowerCamelCase : Tuple=18 , __lowerCamelCase : Optional[int]=1600 , __lowerCamelCase : str=1000 , __lowerCamelCase : Any=True , __lowerCamelCase : Dict=True , __lowerCamelCase : List[Any]=0 , __lowerCamelCase : int=-1 , __lowerCamelCase : Tuple=True , __lowerCamelCase : List[Any]=0.1 , __lowerCamelCase : int=0.0 , __lowerCamelCase : int=True , __lowerCamelCase : str="normal" , __lowerCamelCase : List[str]=0.01 , __lowerCamelCase : Any=0.01 , __lowerCamelCase : List[str]=0.02 , __lowerCamelCase : List[str]=1e-5 , __lowerCamelCase : Union[str, Any]=0 , **__lowerCamelCase : int , ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = [] self.cutoffs.extend(__lowerCamelCase ) if proj_share_all_but_first: SCREAMING_SNAKE_CASE__ = [False] + [True] * len(self.cutoffs ) else: SCREAMING_SNAKE_CASE__ = [False] + [False] * len(self.cutoffs ) SCREAMING_SNAKE_CASE__ = d_model SCREAMING_SNAKE_CASE__ = d_embed SCREAMING_SNAKE_CASE__ = d_head SCREAMING_SNAKE_CASE__ = d_inner SCREAMING_SNAKE_CASE__ = div_val SCREAMING_SNAKE_CASE__ = pre_lnorm SCREAMING_SNAKE_CASE__ = n_layer SCREAMING_SNAKE_CASE__ = n_head SCREAMING_SNAKE_CASE__ = mem_len SCREAMING_SNAKE_CASE__ = same_length SCREAMING_SNAKE_CASE__ = attn_type SCREAMING_SNAKE_CASE__ = clamp_len SCREAMING_SNAKE_CASE__ = sample_softmax SCREAMING_SNAKE_CASE__ = adaptive SCREAMING_SNAKE_CASE__ = dropout SCREAMING_SNAKE_CASE__ = dropatt SCREAMING_SNAKE_CASE__ = untie_r SCREAMING_SNAKE_CASE__ = init SCREAMING_SNAKE_CASE__ = init_range SCREAMING_SNAKE_CASE__ = proj_init_std SCREAMING_SNAKE_CASE__ = init_std SCREAMING_SNAKE_CASE__ = layer_norm_epsilon super().__init__(eos_token_id=__lowerCamelCase , **__lowerCamelCase ) @property def lowercase_ ( self : str ) -> Dict: # Message copied from Transformer-XL documentation logger.info(f'''The model {self.model_type} is one of the few models that has no sequence length limit.''' ) return -1 @max_position_embeddings.setter def lowercase_ ( self : List[str] , __lowerCamelCase : Any ) -> List[Any]: # Message copied from Transformer-XL documentation raise NotImplementedError( f'''The model {self.model_type} is one of the few models that has no sequence length limit.''' )
218
0
'''simple docstring''' import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class A ( __snake_case , __snake_case , unittest.TestCase ): __magic_name__ = IFPipeline __magic_name__ = TEXT_TO_IMAGE_PARAMS - {'''width''', '''height''', '''latents'''} __magic_name__ = TEXT_TO_IMAGE_BATCH_PARAMS __magic_name__ = PipelineTesterMixin.required_optional_params - {'''latents'''} def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" return self._get_dummy_components() def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=0 ) -> Optional[int]: """simple docstring""" if str(SCREAMING_SNAKE_CASE ).startswith('''mps''' ): A : str = torch.manual_seed(SCREAMING_SNAKE_CASE ) else: A : List[Any] = torch.Generator(device=SCREAMING_SNAKE_CASE ).manual_seed(SCREAMING_SNAKE_CASE ) A : Any = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1e-1 ) def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" self._test_save_load_local() def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1e-2 , ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) @slow @require_torch_gpu class A ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self ) -> str: """simple docstring""" A : Optional[int] = IFPipeline.from_pretrained('''DeepFloyd/IF-I-XL-v1.0''' , variant='''fp16''' , torch_dtype=torch.floataa ) A : str = IFSuperResolutionPipeline.from_pretrained( '''DeepFloyd/IF-II-L-v1.0''' , variant='''fp16''' , torch_dtype=torch.floataa , text_encoder=SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to('''cuda''' ) A, A : Optional[Any] = pipe_a.encode_prompt('''anime turtle''' , device='''cuda''' ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() A : List[Any] = None A : List[Any] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img A : Optional[Any] = IFImgaImgPipeline(**pipe_a.components ) A : Union[str, Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting A : str = IFInpaintingPipeline(**pipe_a.components ) A : Any = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" _start_torch_memory_measurement() A : Any = torch.Generator(device='''cpu''' ).manual_seed(0 ) A : Dict = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE , negative_prompt_embeds=SCREAMING_SNAKE_CASE , num_inference_steps=2 , generator=SCREAMING_SNAKE_CASE , output_type='''np''' , ) A : Any = output.images[0] assert image.shape == (64, 64, 3) A : Union[str, Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 A : Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy''' ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # pipeline 2 _start_torch_memory_measurement() A : str = torch.Generator(device='''cpu''' ).manual_seed(0 ) A : Tuple = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE ) A : List[str] = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE , negative_prompt_embeds=SCREAMING_SNAKE_CASE , image=SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE , num_inference_steps=2 , output_type='''np''' , ) A : List[str] = output.images[0] assert image.shape == (256, 256, 3) A : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 A : int = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy''' ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[str]: """simple docstring""" _start_torch_memory_measurement() A : Optional[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE ) A : Optional[Any] = torch.Generator(device='''cpu''' ).manual_seed(0 ) A : Any = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE , negative_prompt_embeds=SCREAMING_SNAKE_CASE , image=SCREAMING_SNAKE_CASE , num_inference_steps=2 , generator=SCREAMING_SNAKE_CASE , output_type='''np''' , ) A : List[str] = output.images[0] assert image.shape == (64, 64, 3) A : Any = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 A : Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy''' ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # pipeline 2 _start_torch_memory_measurement() A : Any = torch.Generator(device='''cpu''' ).manual_seed(0 ) A : Union[str, Any] = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE ) A : str = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE ) A : Tuple = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE , negative_prompt_embeds=SCREAMING_SNAKE_CASE , image=SCREAMING_SNAKE_CASE , original_image=SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE , num_inference_steps=2 , output_type='''np''' , ) A : Any = output.images[0] assert image.shape == (256, 256, 3) A : Union[str, Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 A : List[str] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy''' ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" _start_torch_memory_measurement() A : List[str] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE ) A : Dict = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(SCREAMING_SNAKE_CASE ) A : Optional[Any] = torch.Generator(device='''cpu''' ).manual_seed(0 ) A : List[Any] = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE , negative_prompt_embeds=SCREAMING_SNAKE_CASE , image=SCREAMING_SNAKE_CASE , mask_image=SCREAMING_SNAKE_CASE , num_inference_steps=2 , generator=SCREAMING_SNAKE_CASE , output_type='''np''' , ) A : Any = output.images[0] assert image.shape == (64, 64, 3) A : Optional[Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 A : int = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy''' ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # pipeline 2 _start_torch_memory_measurement() A : int = torch.Generator(device='''cpu''' ).manual_seed(0 ) A : Union[str, Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE ) A : Any = floats_tensor((1, 3, 256, 256) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE ) A : Optional[int] = floats_tensor((1, 3, 256, 256) , rng=random.Random(1 ) ).to(SCREAMING_SNAKE_CASE ) A : Dict = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE , negative_prompt_embeds=SCREAMING_SNAKE_CASE , image=SCREAMING_SNAKE_CASE , mask_image=SCREAMING_SNAKE_CASE , original_image=SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE , num_inference_steps=2 , output_type='''np''' , ) A : Optional[int] = output.images[0] assert image.shape == (256, 256, 3) A : List[Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 A : Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy''' ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( ): '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
3
from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time A : Dict = Lock() def __lowerCamelCase ( __a :Dict , __a :List[str] , __a :Optional[int] , __a :Optional[int] , __a :Optional[Any] , __a :Optional[int] , __a :int ) -> Dict: """simple docstring""" global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 1_0 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(__a ) process_lock.release() # receive your right neighbor's value process_lock.acquire() A__ = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left A__ = min(__a , __a ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(__a ) process_lock.release() # receive your left neighbor's value process_lock.acquire() A__ = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right A__ = max(__a , __a ) # after all swaps are performed, send the values back to main result_pipe[1].send(__a ) def __lowerCamelCase ( __a :List[str] ) -> int: """simple docstring""" A__ = [] A__ = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop A__ = Pipe() A__ = Pipe() process_array_.append( Process( target=__a , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) A__ = temp_rs A__ = temp_rr for i in range(1 , len(__a ) - 1 ): A__ = Pipe() A__ = Pipe() process_array_.append( Process( target=__a , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) A__ = temp_rs A__ = temp_rr process_array_.append( Process( target=__a , args=( len(__a ) - 1, arr[len(__a ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(__a ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(__a ) ): A__ = result_pipe[p][0].recv() process_array_[p].join() return arr def __lowerCamelCase ( ) -> str: """simple docstring""" A__ = list(range(1_0 , 0 , -1 ) ) print("""Initial List""" ) print(*__a ) A__ = odd_even_transposition(__a ) print("""Sorted List\n""" ) print(*__a ) if __name__ == "__main__": main()
274
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: Union[List[PIL.Image.Image], np.ndarray] UpperCAmelCase__: Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('>=', '4.25.0')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('>=', '4.26.0')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version('>=', '0.0.12') ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: np.ndarray UpperCAmelCase__: List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
371
import requests A_ : List[Any] = 'YOUR API KEY' def UpperCamelCase (lowercase_: str , lowercase_: str = giphy_api_key ) -> list: A__ : Dict = """+""".join(query.split() ) A__ : Optional[int] = f"""https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}""" A__ : Any = requests.get(lowercase_ ).json()["""data"""] return [gif["url"] for gif in gifs] if __name__ == "__main__": print('\n'.join(get_gifs('space ship')))
141
0
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'facebook/timesformer': 'https://huggingface.co/facebook/timesformer/resolve/main/config.json', } class _lowercase ( _UpperCAmelCase ): lowercase = 'timesformer' def __init__( self : List[Any] , snake_case : int=2_2_4 , snake_case : Optional[Any]=1_6 , snake_case : Dict=3 , snake_case : Optional[Any]=8 , snake_case : str=7_6_8 , snake_case : List[Any]=1_2 , snake_case : List[Any]=1_2 , snake_case : int=3_0_7_2 , snake_case : Tuple="gelu" , snake_case : int=0.0 , snake_case : Any=0.0 , snake_case : str=0.02 , snake_case : int=1e-6 , snake_case : List[str]=True , snake_case : Any="divided_space_time" , snake_case : List[Any]=0 , **snake_case : Union[str, Any] , ) -> int: """simple docstring""" super().__init__(**lowercase_ ) UpperCamelCase_ : int = image_size UpperCamelCase_ : Optional[Any] = patch_size UpperCamelCase_ : List[Any] = num_channels UpperCamelCase_ : Optional[int] = num_frames UpperCamelCase_ : Any = hidden_size UpperCamelCase_ : int = num_hidden_layers UpperCamelCase_ : int = num_attention_heads UpperCamelCase_ : Any = intermediate_size UpperCamelCase_ : List[Any] = hidden_act UpperCamelCase_ : Optional[int] = hidden_dropout_prob UpperCamelCase_ : Any = attention_probs_dropout_prob UpperCamelCase_ : Optional[int] = initializer_range UpperCamelCase_ : Tuple = layer_norm_eps UpperCamelCase_ : Tuple = qkv_bias UpperCamelCase_ : Union[str, Any] = attention_type UpperCamelCase_ : int = drop_path_rate
175
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { "s-JoL/Open-Llama-V1": "https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json", } class A ( _UpperCAmelCase ): """simple docstring""" lowerCamelCase = 'open-llama' def __init__( self : Any,lowercase_ : Optional[int]=1_0_0_0_0_0,lowercase_ : Union[str, Any]=4_0_9_6,lowercase_ : Dict=1_1_0_0_8,lowercase_ : Dict=3_2,lowercase_ : Optional[int]=3_2,lowercase_ : Dict="silu",lowercase_ : Union[str, Any]=2_0_4_8,lowercase_ : Optional[int]=0.02,lowercase_ : Dict=1E-6,lowercase_ : Dict=True,lowercase_ : List[Any]=0,lowercase_ : Optional[int]=1,lowercase_ : str=2,lowercase_ : str=False,lowercase_ : str=True,lowercase_ : int=0.1,lowercase_ : List[Any]=0.1,lowercase_ : List[Any]=True,lowercase_ : Union[str, Any]=True,lowercase_ : Any=None,**lowercase_ : List[Any],)-> Tuple: '''simple docstring''' A__ = vocab_size A__ = max_position_embeddings A__ = hidden_size A__ = intermediate_size A__ = num_hidden_layers A__ = num_attention_heads A__ = hidden_act A__ = initializer_range A__ = rms_norm_eps A__ = use_cache A__ = kwargs.pop( 'use_memorry_efficient_attention',lowercase_ ) A__ = hidden_dropout_prob A__ = attention_dropout_prob A__ = use_stable_embedding A__ = shared_input_output_embedding A__ = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=lowercase_,bos_token_id=lowercase_,eos_token_id=lowercase_,tie_word_embeddings=lowercase_,**lowercase_,) def snake_case__ ( self : str )-> str: '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling,lowercase_ ) or len(self.rope_scaling ) != 2: raise ValueError( '`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ' F'got {self.rope_scaling}' ) A__ = self.rope_scaling.get('type',lowercase_ ) A__ = self.rope_scaling.get('factor',lowercase_ ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F'`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}' ) if rope_scaling_factor is None or not isinstance(lowercase_,lowercase_ ) or rope_scaling_factor <= 1.0: raise ValueError(F'`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}' )
7
0
'''simple docstring''' import numpy as np from numpy import ndarray from scipy.optimize import Bounds, LinearConstraint, minimize def UpperCamelCase_ ( A__ : ndarray ): '''simple docstring''' return np.dot(_a , _a ) class __snake_case : """simple docstring""" def __init__( self : List[str] , *, lowerCamelCase : Optional[Any] = np.inf , lowerCamelCase : List[Any] = "linear" , lowerCamelCase : Optional[Any] = 0.0 , ) -> None: lowerCAmelCase_ : Optional[int] = regularization lowerCAmelCase_ : List[Any] = gamma if kernel == "linear": lowerCAmelCase_ : int = self.__linear elif kernel == "rbf": if self.gamma == 0: raise ValueError("""rbf kernel requires gamma""" ) if not isinstance(self.gamma , (float, int) ): raise ValueError("""gamma must be float or int""" ) if not self.gamma > 0: raise ValueError("""gamma must be > 0""" ) lowerCAmelCase_ : Optional[Any] = self.__rbf # in the future, there could be a default value like in sklearn # sklear: def_gamma = 1/(n_features * X.var()) (wiki) # previously it was 1/(n_features) else: lowerCAmelCase_ : List[Any] = F'Unknown kernel: {kernel}' raise ValueError(_SCREAMING_SNAKE_CASE ) def __lowercase ( self : Optional[Any] , lowerCamelCase : Dict , lowerCamelCase : Union[str, Any] ) -> float: return np.dot(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __lowercase ( self : List[str] , lowerCamelCase : Dict , lowerCamelCase : Any ) -> float: return np.exp(-(self.gamma * norm_squared(vectora - vectora )) ) def __lowercase ( self : Optional[Any] , lowerCamelCase : Tuple , lowerCamelCase : List[str] ) -> None: lowerCAmelCase_ : int = observations lowerCAmelCase_ : str = classes # using Wolfe's Dual to calculate w. # Primal problem: minimize 1/2*norm_squared(w) # constraint: yn(w . xn + b) >= 1 # # With l a vector # Dual problem: maximize sum_n(ln) - # 1/2 * sum_n(sum_m(ln*lm*yn*ym*xn . xm)) # constraint: self.C >= ln >= 0 # and sum_n(ln*yn) = 0 # Then we get w using w = sum_n(ln*yn*xn) # At the end we can get b ~= mean(yn - w . xn) # # Since we use kernels, we only need l_star to calculate b # and to classify observations (lowerCAmelCase_ ) : Any = np.shape(_SCREAMING_SNAKE_CASE ) def to_minimize(lowerCamelCase : str ) -> float: lowerCAmelCase_ : Dict = 0 (lowerCAmelCase_ ) : str = np.shape(_SCREAMING_SNAKE_CASE ) for i in range(_SCREAMING_SNAKE_CASE ): for j in range(_SCREAMING_SNAKE_CASE ): s += ( candidate[i] * candidate[j] * classes[i] * classes[j] * self.kernel(observations[i] , observations[j] ) ) return 1 / 2 * s - sum(_SCREAMING_SNAKE_CASE ) lowerCAmelCase_ : List[Any] = LinearConstraint(_SCREAMING_SNAKE_CASE , 0 , 0 ) lowerCAmelCase_ : Union[str, Any] = Bounds(0 , self.regularization ) lowerCAmelCase_ : Tuple = minimize( _SCREAMING_SNAKE_CASE , np.ones(_SCREAMING_SNAKE_CASE ) , bounds=_SCREAMING_SNAKE_CASE , constraints=[ly_contraint] ).x lowerCAmelCase_ : Optional[int] = l_star # calculating mean offset of separation plane to points lowerCAmelCase_ : Optional[Any] = 0 for i in range(_SCREAMING_SNAKE_CASE ): for j in range(_SCREAMING_SNAKE_CASE ): s += classes[i] - classes[i] * self.optimum[i] * self.kernel( observations[i] , observations[j] ) lowerCAmelCase_ : int = s / n def __lowercase ( self : Union[str, Any] , lowerCamelCase : Optional[int] ) -> int: lowerCAmelCase_ : List[str] = sum( self.optimum[n] * self.classes[n] * self.kernel(self.observations[n] , _SCREAMING_SNAKE_CASE ) for n in range(len(self.classes ) ) ) return 1 if s + self.offset >= 0 else -1 if __name__ == "__main__": import doctest doctest.testmod()
364
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __snake_case ( unittest.TestCase): """simple docstring""" @property def __lowercase ( self : str ) -> List[Any]: torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) return model def __lowercase ( self : Tuple ) -> Optional[Any]: lowerCAmelCase_ : Optional[int] = self.dummy_uncond_unet lowerCAmelCase_ : Tuple = PNDMScheduler() lowerCAmelCase_ : List[Any] = PNDMPipeline(unet=lowerCamelCase , scheduler=lowerCamelCase ) pndm.to(lowerCamelCase ) pndm.set_progress_bar_config(disable=lowerCamelCase ) lowerCAmelCase_ : Dict = torch.manual_seed(0 ) lowerCAmelCase_ : List[Any] = pndm(generator=lowerCamelCase , num_inference_steps=20 , output_type="""numpy""" ).images lowerCAmelCase_ : str = torch.manual_seed(0 ) lowerCAmelCase_ : int = pndm(generator=lowerCamelCase , num_inference_steps=20 , output_type="""numpy""" , return_dict=lowerCamelCase )[0] lowerCAmelCase_ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase_ : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase_ : int = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class __snake_case ( unittest.TestCase): """simple docstring""" def __lowercase ( self : str ) -> Tuple: lowerCAmelCase_ : str = """google/ddpm-cifar10-32""" lowerCAmelCase_ : Dict = UNetaDModel.from_pretrained(lowerCamelCase ) lowerCAmelCase_ : Dict = PNDMScheduler() lowerCAmelCase_ : Union[str, Any] = PNDMPipeline(unet=lowerCamelCase , scheduler=lowerCamelCase ) pndm.to(lowerCamelCase ) pndm.set_progress_bar_config(disable=lowerCamelCase ) lowerCAmelCase_ : Any = torch.manual_seed(0 ) lowerCAmelCase_ : Union[str, Any] = pndm(generator=lowerCamelCase , output_type="""numpy""" ).images lowerCAmelCase_ : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase_ : List[Any] = np.array([0.1_564, 0.14_645, 0.1_406, 0.14_715, 0.12_425, 0.14_045, 0.13_115, 0.12_175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
89
0
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore __UpperCamelCase : Optional[int] = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" __UpperCamelCase : Optional[Any] = [file for file in filepaths if file != file.lower()] if upper_files: print(f"{len(upper_files)} files contain uppercase characters:") print("\n".join(upper_files) + "\n") __UpperCamelCase : Optional[int] = [file for file in filepaths if " " in file] if space_files: print(f"{len(space_files)} files contain space characters:") print("\n".join(space_files) + "\n") __UpperCamelCase : Optional[Any] = [file for file in filepaths if "-" in file] if hyphen_files: print(f"{len(hyphen_files)} files contain hyphen characters:") print("\n".join(hyphen_files) + "\n") __UpperCamelCase : Any = [file for file in filepaths if os.sep not in file] if nodir_files: print(f"{len(nodir_files)} files are not in a directory:") print("\n".join(nodir_files) + "\n") __UpperCamelCase : List[str] = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
146
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__lowerCAmelCase) , "Tatoeba directory does not exist.") class __magic_name__ ( unittest.TestCase): @cached_property def UpperCAmelCase__ ( self : str ) -> List[Any]: '''simple docstring''' UpperCamelCase__ : Tuple = tempfile.mkdtemp() return TatoebaConverter(save_dir=lowerCamelCase__ ) @slow def UpperCAmelCase__ ( self : Optional[int] ) -> Tuple: '''simple docstring''' self.resolver.convert_models(['''heb-eng'''] ) @slow def UpperCAmelCase__ ( self : Dict ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ : Dict = self.resolver.write_model_card('''opus-mt-he-en''' , dry_run=lowerCamelCase__ ) assert mmeta["long_pair"] == "heb-eng"
146
1
def lowerCamelCase_ ( _a : int , _a : int ): '''simple docstring''' if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) UpperCAmelCase_ : Union[str, Any] = str(bin(lowercase__ ) )[2:] # remove the leading "0b" UpperCAmelCase_ : Dict = str(bin(lowercase__ ) )[2:] UpperCAmelCase_ : List[Any] = max(len(lowercase__ ) , len(lowercase__ ) ) return "0b" + "".join( str(int("""1""" in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(lowercase__ ) , b_binary.zfill(lowercase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
351
from scipy.stats import spearmanr import datasets UpperCamelCase_ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' UpperCamelCase_ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' UpperCamelCase_ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): '''simple docstring''' def A__ ( self: int ) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) ,reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"""] ,) def A__ ( self: int ,lowerCamelCase_: int ,lowerCamelCase_: List[Any] ,lowerCamelCase_: List[str]=False ) -> Dict: UpperCAmelCase_ : List[str] = spearmanr(lowerCamelCase_ ,lowerCamelCase_ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
59
0
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def _UpperCAmelCase ( ): """simple docstring""" _lowerCAmelCase = ArgumentParser( description=( """PyTorch TPU distributed training launch """ """helper utility that will spawn up """ """multiple distributed processes""" ) ) # Optional arguments for the launch helper parser.add_argument("""--num_cores""" , type=snake_case , default=1 , help="""Number of TPU cores to use (1 or 8).""" ) # positional parser.add_argument( """training_script""" , type=snake_case , help=( """The full path to the single TPU training """ """program/script to be launched in parallel, """ """followed by all the arguments for the """ """training script""" ) , ) # rest from the training program parser.add_argument("""training_script_args""" , nargs=snake_case ) return parser.parse_args() def _UpperCAmelCase ( ): """simple docstring""" _lowerCAmelCase = parse_args() # Import training_script as a module. _lowerCAmelCase = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) _lowerCAmelCase = script_fpath.stem _lowerCAmelCase = importlib.import_module(snake_case ) # Patch sys.argv _lowerCAmelCase = [args.training_script] + args.training_script_args + ["""--tpu_num_cores""", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
82
'''simple docstring''' import argparse import os import re lowerCAmelCase_ : Any = 'src/transformers/models/auto' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict lowerCAmelCase_ : List[str] = re.compile(R'[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict') # re pattern that matches identifiers in mappings lowerCAmelCase_ : Tuple = re.compile(R'\s*\(\s*"(\S[^"]+)"') def _lowerCamelCase ( lowercase : Any , lowercase : bool = False ) -> Optional[Any]: with open(lowercase , "r" , encoding="utf-8" ) as f: _a = f.read() _a = content.split("\n" ) _a = [] _a = 0 while line_idx < len(lowercase ): if _re_intro_mapping.search(lines[line_idx] ) is not None: _a = len(re.search(r"^(\s*)\S" , lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(" " * indent + "(" ): new_lines.append(lines[line_idx] ) line_idx += 1 _a = [] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": _a = line_idx while not lines[line_idx].startswith(" " * indent + ")" ): line_idx += 1 blocks.append("\n".join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers _a = sorted(lowercase , key=lambda lowercase : _re_identifier.search(lowercase ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(lowercase , "w" , encoding="utf-8" ) as f: f.write("\n".join(lowercase ) ) elif "\n".join(lowercase ) != content: return True def _lowerCamelCase ( lowercase : bool = False ) -> List[str]: _a = [os.path.join(lowercase , lowercase ) for f in os.listdir(lowercase ) if f.endswith(".py" )] _a = [sort_auto_mapping(lowercase , overwrite=lowercase ) for fname in fnames] if not overwrite and any(lowercase ): _a = [f for f, d in zip(lowercase , lowercase ) if d] raise ValueError( F'The following files have auto mappings that need sorting: {", ".join(lowercase )}. Run `make style` to fix' " this." ) if __name__ == "__main__": lowerCAmelCase_ : Any = argparse.ArgumentParser() parser.add_argument('--check_only', action='store_true', help='Whether to only check or fix style.') lowerCAmelCase_ : Optional[int] = parser.parse_args() sort_all_auto_mappings(not args.check_only)
63
0
import math def a__ ( _SCREAMING_SNAKE_CASE : int = 1_00 ) -> int: """simple docstring""" UpperCAmelCase_ : str = sum(i * i for i in range(1 , n + 1 ) ) UpperCAmelCase_ : Union[str, Any] = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(f"""{solution() = }""")
354
'''simple docstring''' from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def a__ ( ) -> tuple[list[int], int]: """simple docstring""" UpperCAmelCase_ : Tuple = [randint(-10_00 , 10_00 ) for i in range(10 )] UpperCAmelCase_ : str = randint(-50_00 , 50_00 ) return (arr, r) _lowerCamelCase = make_dataset() def a__ ( _SCREAMING_SNAKE_CASE : list[int] , _SCREAMING_SNAKE_CASE : int ) -> tuple[int, ...]: """simple docstring""" for triplet in permutations(_SCREAMING_SNAKE_CASE , 3 ): if sum(_SCREAMING_SNAKE_CASE ) == target: return tuple(sorted(_SCREAMING_SNAKE_CASE ) ) return (0, 0, 0) def a__ ( _SCREAMING_SNAKE_CASE : list[int] , _SCREAMING_SNAKE_CASE : int ) -> tuple[int, int, int]: """simple docstring""" arr.sort() UpperCAmelCase_ : Optional[int] = len(_SCREAMING_SNAKE_CASE ) for i in range(n - 1 ): UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def a__ ( ) -> tuple[float, float]: """simple docstring""" UpperCAmelCase_ : Tuple = "\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n" UpperCAmelCase_ : Optional[Any] = "\ntriplet_sum1(*dataset)\n" UpperCAmelCase_ : str = "\ntriplet_sum2(*dataset)\n" UpperCAmelCase_ : Dict = repeat(setup=_SCREAMING_SNAKE_CASE , stmt=_SCREAMING_SNAKE_CASE , repeat=5 , number=1_00_00 ) UpperCAmelCase_ : Dict = repeat(setup=_SCREAMING_SNAKE_CASE , stmt=_SCREAMING_SNAKE_CASE , repeat=5 , number=1_00_00 ) return (min(_SCREAMING_SNAKE_CASE ), min(_SCREAMING_SNAKE_CASE )) if __name__ == "__main__": from doctest import testmod testmod() _lowerCamelCase = solution_times() print(f"""The time for naive implementation is {times[0]}.""") print(f"""The time for optimized implementation is {times[1]}.""")
67
0
from __future__ import annotations import requests lowerCamelCase : str = set( '''approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports'''.split() ) def snake_case_ ( lowerCAmelCase_ : Tuple , lowerCAmelCase_ : List[str] = 1 , lowerCAmelCase_ : Optional[int] = "new" , lowerCAmelCase_ : Optional[int] = None ): __lowercase : List[str] = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(__UpperCamelCase ) - valid_terms ) ): __lowercase : Any = F"Invalid search term: {invalid_search_terms}" raise ValueError(__UpperCamelCase ) __lowercase : Any = requests.get( F"https://reddit.com/r/{subreddit}/{age}.json?limit={limit}" , headers={"""User-agent""": """A random string"""} , ) if response.status_code == 429: raise requests.HTTPError __lowercase : Optional[Any] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(__UpperCamelCase )} __lowercase : int = {} for id_ in range(__UpperCamelCase ): __lowercase : List[str] = { item: data["data"]["children"][id_]["data"][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('''learnpython''', wanted_data=['''title''', '''url''', '''selftext''']))
233
"""simple docstring""" from collections.abc import Iterable from typing import Generic, TypeVar lowercase__ = TypeVar("""_T""") class __lowerCamelCase ( Generic[_T] ): '''simple docstring''' def __init__( self : Optional[int] , a_ : Iterable[_T] | None = None ): lowerCAmelCase_ : list[_T] = list(iterable or [] ) lowerCAmelCase_ : list[_T] = [] def __len__( self : str ): return len(self._stacka ) + len(self._stacka ) def __repr__( self : List[Any] ): return f'''Queue({tuple(self._stacka[::-1] + self._stacka )})''' def lowerCamelCase ( self : List[str] , a_ : _T ): self._stacka.append(a_ ) def lowerCamelCase ( self : List[str] ): lowerCAmelCase_ : int = self._stacka.pop lowerCAmelCase_ : Any = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("Queue is empty" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
241
0
"""simple docstring""" from argparse import ArgumentParser from . import BaseTransformersCLICommand def lowercase__ ( snake_case_ :Dict ): return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class _UpperCAmelCase ( _lowerCAmelCase ): @staticmethod def a ( _lowercase : ArgumentParser ): __UpperCAmelCase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' , type=_lowercase , default=_lowercase , help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' , action='''store_true''' , help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' , action='''store_true''' , help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' , ) download_parser.add_argument('''model''' , type=_lowercase , help='''Name of the model to download''' ) download_parser.set_defaults(func=_lowercase ) def __init__( self : Dict , _lowercase : str , _lowercase : str , _lowercase : bool , _lowercase : bool ): __UpperCAmelCase = model __UpperCAmelCase = cache __UpperCAmelCase = force __UpperCAmelCase = trust_remote_code def a ( self : Dict ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
371
"""simple docstring""" from __future__ import annotations import bisect def lowercase__ ( snake_case_ :list[int] , snake_case_ :int , snake_case_ :int = 0 , snake_case_ :int = -1 ): if hi < 0: __UpperCAmelCase = len(snake_case_ ) while lo < hi: __UpperCAmelCase = lo + (hi - lo) // 2 if sorted_collection[mid] < item: __UpperCAmelCase = mid + 1 else: __UpperCAmelCase = mid return lo def lowercase__ ( snake_case_ :list[int] , snake_case_ :int , snake_case_ :int = 0 , snake_case_ :int = -1 ): if hi < 0: __UpperCAmelCase = len(snake_case_ ) while lo < hi: __UpperCAmelCase = lo + (hi - lo) // 2 if sorted_collection[mid] <= item: __UpperCAmelCase = mid + 1 else: __UpperCAmelCase = mid return lo def lowercase__ ( snake_case_ :list[int] , snake_case_ :int , snake_case_ :int = 0 , snake_case_ :int = -1 ): sorted_collection.insert(bisect_left(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) , snake_case_ ) def lowercase__ ( snake_case_ :list[int] , snake_case_ :int , snake_case_ :int = 0 , snake_case_ :int = -1 ): sorted_collection.insert(bisect_right(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) , snake_case_ ) def lowercase__ ( snake_case_ :list[int] , snake_case_ :int ): __UpperCAmelCase = 0 __UpperCAmelCase = len(snake_case_ ) - 1 while left <= right: __UpperCAmelCase = left + (right - left) // 2 __UpperCAmelCase = sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: __UpperCAmelCase = midpoint - 1 else: __UpperCAmelCase = midpoint + 1 return None def lowercase__ ( snake_case_ :list[int] , snake_case_ :int ): __UpperCAmelCase = bisect.bisect_left(snake_case_ , snake_case_ ) if index != len(snake_case_ ) and sorted_collection[index] == item: return index return None def lowercase__ ( snake_case_ :list[int] , snake_case_ :int , snake_case_ :int , snake_case_ :int ): if right < left: return None __UpperCAmelCase = left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(snake_case_ , snake_case_ , snake_case_ , midpoint - 1 ) else: return binary_search_by_recursion(snake_case_ , snake_case_ , midpoint + 1 , snake_case_ ) if __name__ == "__main__": _lowercase : Optional[Any] = input('Enter numbers separated by comma:\n').strip() _lowercase : Optional[int] = sorted(int(item) for item in user_input.split(',')) _lowercase : Optional[Any] = int(input('Enter a single number to be found in the list:\n')) _lowercase : int = binary_search(collection, target) if result is None: print(f"""{target} was not found in {collection}.""") else: print(f"""{target} was found at position {result} in {collection}.""")
86
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class UpperCAmelCase__ ( __UpperCamelCase ,unittest.TestCase ): '''simple docstring''' UpperCamelCase = ShapEPipeline UpperCamelCase = ["""prompt"""] UpperCamelCase = ["""prompt"""] UpperCamelCase = [ """num_images_per_prompt""", """num_inference_steps""", """generator""", """latents""", """guidance_scale""", """frame_size""", """output_type""", """return_dict""", ] UpperCamelCase = False @property def snake_case__ ( self : List[Any] ): '''simple docstring''' return 32 @property def snake_case__ ( self : Union[str, Any] ): '''simple docstring''' return 32 @property def snake_case__ ( self : Optional[Any] ): '''simple docstring''' return self.time_input_dim * 4 @property def snake_case__ ( self : str ): '''simple docstring''' return 8 @property def snake_case__ ( self : Union[str, Any] ): '''simple docstring''' __UpperCAmelCase : List[Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def snake_case__ ( self : List[Any] ): '''simple docstring''' torch.manual_seed(0 ) __UpperCAmelCase : int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModelWithProjection(a_ ) @property def snake_case__ ( self : List[Any] ): '''simple docstring''' torch.manual_seed(0 ) __UpperCAmelCase : Any = { '''num_attention_heads''': 2, '''attention_head_dim''': 16, '''embedding_dim''': self.time_input_dim, '''num_embeddings''': 32, '''embedding_proj_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''num_layers''': 1, '''clip_embed_dim''': self.time_input_dim * 2, '''additional_embeddings''': 0, '''time_embed_act_fn''': '''gelu''', '''norm_in_type''': '''layer''', '''encoder_hid_proj_type''': None, '''added_emb_type''': None, } __UpperCAmelCase : Tuple = PriorTransformer(**a_ ) return model @property def snake_case__ ( self : Any ): '''simple docstring''' torch.manual_seed(0 ) __UpperCAmelCase : int = { '''param_shapes''': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), '''d_latent''': self.time_input_dim, '''d_hidden''': self.renderer_dim, '''n_output''': 12, '''background''': ( 0.1, 0.1, 0.1, ), } __UpperCAmelCase : List[Any] = ShapERenderer(**a_ ) return model def snake_case__ ( self : List[str] ): '''simple docstring''' __UpperCAmelCase : Any = self.dummy_prior __UpperCAmelCase : Tuple = self.dummy_text_encoder __UpperCAmelCase : Dict = self.dummy_tokenizer __UpperCAmelCase : List[Any] = self.dummy_renderer __UpperCAmelCase : Any = HeunDiscreteScheduler( beta_schedule='''exp''' , num_train_timesteps=10_24 , prediction_type='''sample''' , use_karras_sigmas=a_ , clip_sample=a_ , clip_sample_range=1.0 , ) __UpperCAmelCase : List[str] = { '''prior''': prior, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''renderer''': renderer, '''scheduler''': scheduler, } return components def snake_case__ ( self : int , a_ : List[Any] , a_ : int=0 ): '''simple docstring''' if str(a_ ).startswith('''mps''' ): __UpperCAmelCase : List[Any] = torch.manual_seed(a_ ) else: __UpperCAmelCase : Any = torch.Generator(device=a_ ).manual_seed(a_ ) __UpperCAmelCase : Union[str, Any] = { '''prompt''': '''horse''', '''generator''': generator, '''num_inference_steps''': 1, '''frame_size''': 32, '''output_type''': '''np''', } return inputs def snake_case__ ( self : Optional[Any] ): '''simple docstring''' __UpperCAmelCase : List[str] = '''cpu''' __UpperCAmelCase : Union[str, Any] = self.get_dummy_components() __UpperCAmelCase : str = self.pipeline_class(**a_ ) __UpperCAmelCase : List[str] = pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __UpperCAmelCase : List[Any] = pipe(**self.get_dummy_inputs(a_ ) ) __UpperCAmelCase : int = output.images[0] __UpperCAmelCase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __UpperCAmelCase : Any = np.array( [ 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, 0.0_0_0_3_9_2_1_6, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self : Dict ): '''simple docstring''' self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def snake_case__ ( self : str ): '''simple docstring''' __UpperCAmelCase : int = torch_device == '''cpu''' __UpperCAmelCase : int = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=a_ , relax_max_difference=a_ , ) def snake_case__ ( self : int ): '''simple docstring''' __UpperCAmelCase : Optional[int] = self.get_dummy_components() __UpperCAmelCase : str = self.pipeline_class(**a_ ) __UpperCAmelCase : Optional[Any] = pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __UpperCAmelCase : List[str] = 1 __UpperCAmelCase : Dict = 2 __UpperCAmelCase : List[Any] = self.get_dummy_inputs(a_ ) for key in inputs.keys(): if key in self.batch_params: __UpperCAmelCase : Optional[int] = batch_size * [inputs[key]] __UpperCAmelCase : Tuple = pipe(**a_ , num_images_per_prompt=a_ )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class UpperCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self : str ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self : Any ): '''simple docstring''' __UpperCAmelCase : List[str] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/test_shap_e_np_out.npy''' ) __UpperCAmelCase : List[str] = ShapEPipeline.from_pretrained('''openai/shap-e''' ) __UpperCAmelCase : Tuple = pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __UpperCAmelCase : List[str] = torch.Generator(device=a_ ).manual_seed(0 ) __UpperCAmelCase : Optional[Any] = pipe( '''a shark''' , generator=a_ , guidance_scale=1_5.0 , num_inference_steps=64 , frame_size=64 , output_type='''np''' , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(a_ , a_ )
226
import qiskit def a ( _UpperCAmelCase : int , _UpperCAmelCase : int ): '''simple docstring''' __UpperCAmelCase : List[str] = qiskit.Aer.get_backend('''aer_simulator''' ) __UpperCAmelCase : Optional[Any] = qiskit.QuantumCircuit(4 , 2 ) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0 ) if bita == 1: qc_ha.x(1 ) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 , 2 ) qc_ha.cx(1 , 2 ) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 , 1 , 3 ) qc_ha.barrier() # extract outputs qc_ha.measure(2 , 0 ) # extract XOR value qc_ha.measure(3 , 1 ) # extract AND value # Execute the circuit on the qasm simulator __UpperCAmelCase : str = qiskit.execute(_UpperCAmelCase , _UpperCAmelCase , shots=10_00 ) # Return the histogram data of the results of the experiment return job.result().get_counts(_UpperCAmelCase ) if __name__ == "__main__": __A =half_adder(1, 1) print(f'''Half Adder Output Qubit Counts: {counts}''')
226
1
import math import random from typing import Any from .hill_climbing import SearchProblem def _lowerCAmelCase ( UpperCAmelCase : Any , UpperCAmelCase : bool = True , UpperCAmelCase : float = math.inf , UpperCAmelCase : float = -math.inf , UpperCAmelCase : float = math.inf , UpperCAmelCase : float = -math.inf , UpperCAmelCase : bool = False , UpperCAmelCase : float = 100 , UpperCAmelCase : float = 0.01 , UpperCAmelCase : float = 1 , ): '''simple docstring''' UpperCamelCase__ : int =False UpperCamelCase__ : Optional[int] =search_prob UpperCamelCase__ : Any =start_temperate UpperCamelCase__ : Union[str, Any] =[] UpperCamelCase__ : Union[str, Any] =0 UpperCamelCase__ : Any =None while not search_end: UpperCamelCase__ : int =current_state.score() if best_state is None or current_score > best_state.score(): UpperCamelCase__ : List[Any] =current_state scores.append(UpperCAmelCase ) iterations += 1 UpperCamelCase__ : Optional[int] =None UpperCamelCase__ : List[str] =current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to UpperCamelCase__ : Tuple =random.randint(0 , len(UpperCAmelCase ) - 1 ) # picking a random neighbor UpperCamelCase__ : Optional[Any] =neighbors.pop(UpperCAmelCase ) UpperCamelCase__ : Tuple =picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: UpperCamelCase__ : List[Any] =change * -1 # in case we are finding minimum if change > 0: # improves the solution UpperCamelCase__ : Any =picked_neighbor else: UpperCamelCase__ : str =(math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability UpperCamelCase__ : Optional[int] =picked_neighbor UpperCamelCase__ : Any =current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor UpperCamelCase__ : Optional[Any] =True else: UpperCamelCase__ : Tuple =next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(UpperCAmelCase ) , UpperCAmelCase ) plt.xlabel('''Iterations''' ) plt.ylabel('''Function values''' ) plt.show() return best_state if __name__ == "__main__": def _lowerCAmelCase ( UpperCAmelCase : Any , UpperCAmelCase : Union[str, Any] ): '''simple docstring''' return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) _SCREAMING_SNAKE_CASE : Tuple = SearchProblem(x=1_2, y=4_7, step_size=1, function_to_optimize=test_fa) _SCREAMING_SNAKE_CASE : Union[str, Any] = simulated_annealing( prob, find_max=False, max_x=1_0_0, min_x=5, max_y=5_0, min_y=-5, visualization=True ) print( """The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 """ F'''and 50 > y > - 5 found via hill climbing: {local_min.score()}''' ) # starting the problem with initial coordinates (12, 47) _SCREAMING_SNAKE_CASE : Dict = SearchProblem(x=1_2, y=4_7, step_size=1, function_to_optimize=test_fa) _SCREAMING_SNAKE_CASE : List[str] = simulated_annealing( prob, find_max=True, max_x=1_0_0, min_x=5, max_y=5_0, min_y=-5, visualization=True ) print( """The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 """ F'''and 50 > y > - 5 found via hill climbing: {local_min.score()}''' ) def _lowerCAmelCase ( UpperCAmelCase : Optional[int] , UpperCAmelCase : List[Any] ): '''simple docstring''' return (3 * x**2) - (6 * y) _SCREAMING_SNAKE_CASE : List[str] = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) _SCREAMING_SNAKE_CASE : Any = simulated_annealing(prob, find_max=False, visualization=True) print( """The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: """ F'''{local_min.score()}''' ) _SCREAMING_SNAKE_CASE : str = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) _SCREAMING_SNAKE_CASE : Union[str, Any] = simulated_annealing(prob, find_max=True, visualization=True) print( """The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: """ F'''{local_min.score()}''' )
353
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class __a ( unittest.TestCase ): """simple docstring""" def _lowerCAmelCase ( self : Tuple ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowerCAmelCase ( self : Tuple ): UpperCamelCase__ : List[str] =1 UpperCamelCase__ : List[str] =3 UpperCamelCase__ : Optional[Any] =(32, 32) UpperCamelCase__ : Tuple =floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowercase_ ) return image @property def _lowerCAmelCase ( self : Union[str, Any] ): torch.manual_seed(0 ) UpperCamelCase__ : Dict =UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) return model @property def _lowerCAmelCase ( self : str ): torch.manual_seed(0 ) UpperCamelCase__ : Optional[int] =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def _lowerCAmelCase ( self : Union[str, Any] ): torch.manual_seed(0 ) UpperCamelCase__ : Union[str, Any] =RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5006 , ) return RobertaSeriesModelWithTransformation(lowercase_ ) @property def _lowerCAmelCase ( self : Optional[Any] ): def extract(*lowercase_ : Dict , **lowercase_ : List[Any] ): class __a : """simple docstring""" def __init__( self : Optional[Any] ): UpperCamelCase__ : Dict =torch.ones([0] ) def _lowerCAmelCase ( self : Union[str, Any] , lowercase_ : Optional[int] ): self.pixel_values.to(lowercase_ ) return self return Out() return extract def _lowerCAmelCase ( self : Optional[Any] ): UpperCamelCase__ : str ='''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCamelCase__ : Any =self.dummy_cond_unet UpperCamelCase__ : Tuple =PNDMScheduler(skip_prk_steps=lowercase_ ) UpperCamelCase__ : Optional[Any] =self.dummy_vae UpperCamelCase__ : List[str] =self.dummy_text_encoder UpperCamelCase__ : List[str] =XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) UpperCamelCase__ : Tuple =77 UpperCamelCase__ : int =self.dummy_image.to(lowercase_ ) UpperCamelCase__ : Tuple =init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk UpperCamelCase__ : Union[str, Any] =AltDiffusionImgaImgPipeline( unet=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , safety_checker=lowercase_ , feature_extractor=self.dummy_extractor , ) UpperCamelCase__ : List[Any] =VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowercase_ ) UpperCamelCase__ : Union[str, Any] =alt_pipe.to(lowercase_ ) alt_pipe.set_progress_bar_config(disable=lowercase_ ) UpperCamelCase__ : Tuple ='''A painting of a squirrel eating a burger''' UpperCamelCase__ : str =torch.Generator(device=lowercase_ ).manual_seed(0 ) UpperCamelCase__ : str =alt_pipe( [prompt] , generator=lowercase_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=lowercase_ , ) UpperCamelCase__ : Any =output.images UpperCamelCase__ : Tuple =torch.Generator(device=lowercase_ ).manual_seed(0 ) UpperCamelCase__ : str =alt_pipe( [prompt] , generator=lowercase_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=lowercase_ , return_dict=lowercase_ , )[0] UpperCamelCase__ : Union[str, Any] =image[0, -3:, -3:, -1] UpperCamelCase__ : int =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) UpperCamelCase__ : Optional[Any] =np.array([0.4_4_2_7, 0.3_7_3_1, 0.4_2_4_9, 0.4_9_4_1, 0.4_5_4_6, 0.4_1_4_8, 0.4_1_9_3, 0.4_6_6_6, 0.4_4_9_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5e-3 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def _lowerCAmelCase ( self : str ): UpperCamelCase__ : List[Any] =self.dummy_cond_unet UpperCamelCase__ : int =PNDMScheduler(skip_prk_steps=lowercase_ ) UpperCamelCase__ : Optional[Any] =self.dummy_vae UpperCamelCase__ : Dict =self.dummy_text_encoder UpperCamelCase__ : Optional[int] =XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) UpperCamelCase__ : List[Any] =77 UpperCamelCase__ : List[Any] =self.dummy_image.to(lowercase_ ) # put models in fp16 UpperCamelCase__ : Dict =unet.half() UpperCamelCase__ : List[str] =vae.half() UpperCamelCase__ : int =bert.half() # make sure here that pndm scheduler skips prk UpperCamelCase__ : List[str] =AltDiffusionImgaImgPipeline( unet=lowercase_ , scheduler=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , safety_checker=lowercase_ , feature_extractor=self.dummy_extractor , ) UpperCamelCase__ : Union[str, Any] =VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowercase_ ) UpperCamelCase__ : List[Any] =alt_pipe.to(lowercase_ ) alt_pipe.set_progress_bar_config(disable=lowercase_ ) UpperCamelCase__ : Dict ='''A painting of a squirrel eating a burger''' UpperCamelCase__ : Optional[Any] =torch.manual_seed(0 ) UpperCamelCase__ : str =alt_pipe( [prompt] , generator=lowercase_ , num_inference_steps=2 , output_type='''np''' , image=lowercase_ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def _lowerCAmelCase ( self : Union[str, Any] ): UpperCamelCase__ : str =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) # resize to resolution that is divisible by 8 but not 16 or 32 UpperCamelCase__ : int =init_image.resize((760, 504) ) UpperCamelCase__ : Optional[int] ='''BAAI/AltDiffusion''' UpperCamelCase__ : Union[str, Any] =AltDiffusionImgaImgPipeline.from_pretrained( lowercase_ , safety_checker=lowercase_ , ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) pipe.enable_attention_slicing() UpperCamelCase__ : Dict ='''A fantasy landscape, trending on artstation''' UpperCamelCase__ : str =torch.manual_seed(0 ) UpperCamelCase__ : Any =pipe( prompt=lowercase_ , image=lowercase_ , strength=0.7_5 , guidance_scale=7.5 , generator=lowercase_ , output_type='''np''' , ) UpperCamelCase__ : List[Any] =output.images[0] UpperCamelCase__ : int =image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) UpperCamelCase__ : Union[str, Any] =np.array([0.9_3_5_8, 0.9_3_9_7, 0.9_5_9_9, 0.9_9_0_1, 1.0_0_0_0, 1.0_0_0_0, 0.9_8_8_2, 1.0_0_0_0, 1.0_0_0_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class __a ( unittest.TestCase ): """simple docstring""" def _lowerCAmelCase ( self : Dict ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCAmelCase ( self : List[Any] ): UpperCamelCase__ : Tuple =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) UpperCamelCase__ : List[Any] =init_image.resize((768, 512) ) UpperCamelCase__ : str =load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy''' ) UpperCamelCase__ : List[str] ='''BAAI/AltDiffusion''' UpperCamelCase__ : List[str] =AltDiffusionImgaImgPipeline.from_pretrained( lowercase_ , safety_checker=lowercase_ , ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) pipe.enable_attention_slicing() UpperCamelCase__ : List[Any] ='''A fantasy landscape, trending on artstation''' UpperCamelCase__ : List[Any] =torch.manual_seed(0 ) UpperCamelCase__ : int =pipe( prompt=lowercase_ , image=lowercase_ , strength=0.7_5 , guidance_scale=7.5 , generator=lowercase_ , output_type='''np''' , ) UpperCamelCase__ : List[Any] =output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1e-2
157
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A = { "configuration_blip_2": [ "BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP", "Blip2Config", "Blip2QFormerConfig", "Blip2VisionConfig", ], "processing_blip_2": ["Blip2Processor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST", "Blip2Model", "Blip2QFormerModel", "Blip2PreTrainedModel", "Blip2ForConditionalGeneration", "Blip2VisionModel", ] if TYPE_CHECKING: from .configuration_blip_a import ( BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipaConfig, BlipaQFormerConfig, BlipaVisionConfig, ) from .processing_blip_a import BlipaProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip_a import ( BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST, BlipaForConditionalGeneration, BlipaModel, BlipaPreTrainedModel, BlipaQFormerModel, BlipaVisionModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
217
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot import BlenderbotTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __A = logging.get_logger(__name__) __A = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } __A = { "vocab_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json"}, "merges_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt"}, "tokenizer_config_file": { "facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json" }, } __A = {"facebook/blenderbot-3B": 128} class snake_case ( __snake_case ): SCREAMING_SNAKE_CASE_ : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : str = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : Optional[int] = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE_ : List[Any] = BlenderbotTokenizer def __init__( self : Any , UpperCamelCase__ : str=None , UpperCamelCase__ : Tuple=None , UpperCamelCase__ : List[Any]=None , UpperCamelCase__ : Optional[Any]="replace" , UpperCamelCase__ : int="<s>" , UpperCamelCase__ : str="</s>" , UpperCamelCase__ : List[Any]="</s>" , UpperCamelCase__ : Union[str, Any]="<s>" , UpperCamelCase__ : List[str]="<unk>" , UpperCamelCase__ : Dict="<pad>" , UpperCamelCase__ : List[Any]="<mask>" , UpperCamelCase__ : Any=False , UpperCamelCase__ : int=True , **UpperCamelCase__ : Union[str, Any] , )-> int: '''simple docstring''' super().__init__( UpperCamelCase__ , UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , errors=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ , **UpperCamelCase__ , ) __lowerCAmelCase: Any = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get("add_prefix_space" , UpperCamelCase__) != add_prefix_space: __lowerCAmelCase: Dict = getattr(UpperCamelCase__ , pre_tok_state.pop("type")) __lowerCAmelCase: int = add_prefix_space __lowerCAmelCase: str = pre_tok_class(**UpperCamelCase__) __lowerCAmelCase: List[Any] = add_prefix_space __lowerCAmelCase: Union[str, Any] = "post_processor" __lowerCAmelCase: Tuple = getattr(self.backend_tokenizer , UpperCamelCase__ , UpperCamelCase__) if tokenizer_component_instance: __lowerCAmelCase: List[Any] = json.loads(tokenizer_component_instance.__getstate__()) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: __lowerCAmelCase: str = tuple(state["sep"]) if "cls" in state: __lowerCAmelCase: Any = tuple(state["cls"]) __lowerCAmelCase: int = False if state.get("add_prefix_space" , UpperCamelCase__) != add_prefix_space: __lowerCAmelCase: Dict = add_prefix_space __lowerCAmelCase: Optional[Any] = True if state.get("trim_offsets" , UpperCamelCase__) != trim_offsets: __lowerCAmelCase: Dict = trim_offsets __lowerCAmelCase: Tuple = True if changes_to_apply: __lowerCAmelCase: Dict = getattr(UpperCamelCase__ , state.pop("type")) __lowerCAmelCase: str = component_class(**UpperCamelCase__) setattr(self.backend_tokenizer , UpperCamelCase__ , UpperCamelCase__) @property # Copied from transformers.models.roberta.tokenization_roberta_fast.RobertaTokenizerFast.mask_token with Roberta->Blenderbot, RoBERTa->Blenderbot def lowercase_ ( self : Optional[Any])-> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet.") return None return str(self._mask_token) @mask_token.setter def lowercase_ ( self : str , UpperCamelCase__ : Union[str, Any])-> Tuple: '''simple docstring''' __lowerCAmelCase: int = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__) if isinstance(UpperCamelCase__ , UpperCamelCase__) else value __lowerCAmelCase: Any = value def lowercase_ ( self : Union[str, Any] , *UpperCamelCase__ : List[str] , **UpperCamelCase__ : Optional[int])-> BatchEncoding: '''simple docstring''' __lowerCAmelCase: List[str] = kwargs.get("is_split_into_words" , UpperCamelCase__) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*UpperCamelCase__ , **UpperCamelCase__) def lowercase_ ( self : Any , *UpperCamelCase__ : Optional[int] , **UpperCamelCase__ : int)-> BatchEncoding: '''simple docstring''' __lowerCAmelCase: List[Any] = kwargs.get("is_split_into_words" , UpperCamelCase__) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*UpperCamelCase__ , **UpperCamelCase__) def lowercase_ ( self : List[str] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None)-> Tuple[str]: '''simple docstring''' __lowerCAmelCase: Optional[int] = self._tokenizer.model.save(UpperCamelCase__ , name=UpperCamelCase__) return tuple(UpperCamelCase__) def lowercase_ ( self : str , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None)-> List[int]: '''simple docstring''' __lowerCAmelCase: Tuple = [self.sep_token_id] __lowerCAmelCase: Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def lowercase_ ( self : Union[str, Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None)-> List[Any]: '''simple docstring''' return token_ids_a + [self.eos_token_id] def lowercase_ ( self : Dict , UpperCamelCase__ : "Conversation")-> List[int]: '''simple docstring''' __lowerCAmelCase: str = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(" " + text) else: # Generated responses should contain them already. inputs.append(UpperCamelCase__) __lowerCAmelCase: Optional[int] = " ".join(UpperCamelCase__) __lowerCAmelCase: Tuple = self.encode(UpperCamelCase__) if len(UpperCamelCase__) > self.model_max_length: __lowerCAmelCase: int = input_ids[-self.model_max_length :] logger.warning(f"Trimmed input from conversation as it was longer than {self.model_max_length} tokens.") return input_ids
217
1
"""simple docstring""" class lowerCAmelCase__ : '''simple docstring''' def __init__( self : List[str]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = {} def _SCREAMING_SNAKE_CASE ( self : Dict): '''simple docstring''' print(self.vertex) for i in self.vertex: print(lowercase_ , ''' -> ''' , ''' -> '''.join([str(lowercase_) for j in self.vertex[i]])) def _SCREAMING_SNAKE_CASE ( self : int , lowercase_ : int , lowercase_ : int): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(lowercase_) else: # else make a new vertex SCREAMING_SNAKE_CASE_ : Tuple = [to_vertex] def _SCREAMING_SNAKE_CASE ( self : Dict): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = [False] * len(self.vertex) # call the recursive helper function for i in range(len(self.vertex)): if not visited[i]: self.dfs_recursive(lowercase_ , lowercase_) def _SCREAMING_SNAKE_CASE ( self : Any , lowercase_ : int , lowercase_ : list): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = True print(lowercase_ , end=''' ''') # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(lowercase_ , lowercase_) if __name__ == "__main__": UpperCAmelCase_ : Union[str, Any] = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
365
"""simple docstring""" from itertools import permutations def _A (__a ) -> bool: """simple docstring""" if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False SCREAMING_SNAKE_CASE_ : List[str] = [7, 11, 13, 17] for i, test in enumerate(__a ): if (num[i + 4] * 1_00 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def _A (__a = 10 ) -> int: """simple docstring""" return sum( int(''''''.join(map(__a , __a ) ) ) for num in permutations(range(__a ) ) if is_substring_divisible(__a ) ) if __name__ == "__main__": print(f'''{solution() = }''')
318
0
"""simple docstring""" import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class SCREAMING_SNAKE_CASE__ : def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 13 , _SCREAMING_SNAKE_CASE = 64 , _SCREAMING_SNAKE_CASE = 2 , _SCREAMING_SNAKE_CASE = 3 , _SCREAMING_SNAKE_CASE = 3 , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE = 128 , _SCREAMING_SNAKE_CASE=[16, 32, 64, 128] , _SCREAMING_SNAKE_CASE = 7 , _SCREAMING_SNAKE_CASE = 4 , _SCREAMING_SNAKE_CASE = 37 , _SCREAMING_SNAKE_CASE = "gelu" , _SCREAMING_SNAKE_CASE = 0.1 , _SCREAMING_SNAKE_CASE = 0.1 , _SCREAMING_SNAKE_CASE = 10 , _SCREAMING_SNAKE_CASE = 0.02 , _SCREAMING_SNAKE_CASE = 2 , _SCREAMING_SNAKE_CASE = 1 , _SCREAMING_SNAKE_CASE = 128 , _SCREAMING_SNAKE_CASE = [2, 2, 2, 2] , _SCREAMING_SNAKE_CASE = 2 , _SCREAMING_SNAKE_CASE = 2 , ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : Tuple = parent UpperCAmelCase : Optional[int] = batch_size UpperCAmelCase : Any = image_size UpperCAmelCase : Tuple = patch_size UpperCAmelCase : int = num_channels UpperCAmelCase : List[Any] = is_training UpperCAmelCase : List[Any] = use_labels UpperCAmelCase : Dict = hidden_size UpperCAmelCase : Optional[Any] = num_hidden_layers UpperCAmelCase : Union[str, Any] = num_attention_heads UpperCAmelCase : List[str] = intermediate_size UpperCAmelCase : Dict = hidden_act UpperCAmelCase : int = hidden_dropout_prob UpperCAmelCase : Dict = attention_probs_dropout_prob UpperCAmelCase : Any = type_sequence_label_size UpperCAmelCase : List[Any] = initializer_range UpperCAmelCase : Union[str, Any] = encoder_stride UpperCAmelCase : Dict = num_attention_outputs UpperCAmelCase : List[str] = embed_dim UpperCAmelCase : int = embed_dim + 1 UpperCAmelCase : Union[str, Any] = resolution UpperCAmelCase : Any = depths UpperCAmelCase : int = hidden_sizes UpperCAmelCase : List[str] = dim UpperCAmelCase : Optional[Any] = mlp_expansion_ratio def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase : Dict = None if self.use_labels: UpperCAmelCase : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase : Any = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self ) -> List[str]: '''simple docstring''' return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : Optional[int] = TFEfficientFormerModel(config=_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Any = model(_SCREAMING_SNAKE_CASE , training=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Dict: '''simple docstring''' UpperCAmelCase : Any = self.type_sequence_label_size UpperCAmelCase : Optional[Any] = TFEfficientFormerForImageClassification(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , training=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase : List[str] = 1 UpperCAmelCase : Optional[Any] = TFEfficientFormerForImageClassification(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' UpperCAmelCase : List[str] = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Tuple = config_and_inputs UpperCAmelCase : List[str] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): __lowerCAmelCase : Optional[int] = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) __lowerCAmelCase : Optional[int] = ( { 'feature-extraction': TFEfficientFormerModel, 'image-classification': ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) __lowerCAmelCase : Any = False __lowerCAmelCase : str = False __lowerCAmelCase : Union[str, Any] = False __lowerCAmelCase : Dict = False __lowerCAmelCase : Tuple = False def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Optional[Any] = TFEfficientFormerModelTester(self ) UpperCAmelCase : List[str] = ConfigTester( self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE , hidden_size=37 ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="""EfficientFormer does not use inputs_embeds""" ) def SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' pass @unittest.skip(reason="""EfficientFormer does not support input and output embeddings""" ) def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' pass def SCREAMING_SNAKE_CASE ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase , UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase : Any = model_class(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : List[str] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase : Any = [*signature.parameters.keys()] UpperCAmelCase : int = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: '''simple docstring''' def check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): UpperCAmelCase : Any = model_class(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : int = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , training=_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase : Optional[Any] = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) if hasattr(self.model_tester , """encoder_seq_length""" ): UpperCAmelCase : str = self.model_tester.encoder_seq_length if hasattr(self.model_tester , """chunk_length""" ) and self.model_tester.chunk_length > 1: UpperCAmelCase : List[Any] = seq_length * self.model_tester.chunk_length else: UpperCAmelCase : str = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: UpperCAmelCase : Dict = outputs.decoder_hidden_states self.asseretIsInstance(_SCREAMING_SNAKE_CASE , (list, tuple) ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : Dict = getattr(self.model_tester , """seq_length""" , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : Optional[Any] = getattr(self.model_tester , """decoder_seq_length""" , _SCREAMING_SNAKE_CASE ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) UpperCAmelCase , UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase : Optional[int] = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase : Optional[int] = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) -> Dict: '''simple docstring''' UpperCAmelCase : Dict = super()._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) @unittest.skip(reason="""EfficientFormer does not implement masked image modeling yet""" ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @slow def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : Optional[Any] = TFEfficientFormerModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase , UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase : Optional[Any] = True UpperCAmelCase : Tuple = getattr(self.model_tester , """seq_length""" , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : List[Any] = getattr(self.model_tester , """encoder_seq_length""" , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : Any = getattr(self.model_tester , """key_length""" , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : int = getattr(self.model_tester , """chunk_length""" , _SCREAMING_SNAKE_CASE ) if chunk_length is not None and hasattr(self.model_tester , """num_hashes""" ): UpperCAmelCase : Dict = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: UpperCAmelCase : Optional[Any] = True UpperCAmelCase : str = False UpperCAmelCase : List[Any] = True UpperCAmelCase : Union[str, Any] = model_class(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Dict = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , training=_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Tuple = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCAmelCase : Optional[Any] = True UpperCAmelCase : Dict = model_class(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Optional[Any] = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , training=_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Union[str, Any] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' UpperCAmelCase , UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model UpperCAmelCase : Tuple = model_class(_SCREAMING_SNAKE_CASE ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes UpperCAmelCase : Optional[int] = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=_SCREAMING_SNAKE_CASE ) for key, val in model.input_signature.items() if key in model.dummy_inputs } UpperCAmelCase : Union[str, Any] = model(_SCREAMING_SNAKE_CASE ) self.assertTrue(outputs_dict is not None ) def _snake_case ( ): UpperCAmelCase : Tuple = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE ( self ) -> List[str]: '''simple docstring''' return ( EfficientFormerImageProcessor.from_pretrained("""snap-research/efficientformer-l1-300""" ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Any = TFEfficientFormerForImageClassification.from_pretrained("""snap-research/efficientformer-l1-300""" ) UpperCAmelCase : List[str] = self.default_image_processor UpperCAmelCase : Optional[Any] = prepare_img() UpperCAmelCase : str = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="""tf""" ) # forward pass UpperCAmelCase : List[Any] = model(**_SCREAMING_SNAKE_CASE , training=_SCREAMING_SNAKE_CASE ) # verify the logits UpperCAmelCase : str = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : Union[str, Any] = tf.constant([-0.0555, 0.4825, -0.0852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) ) @slow def SCREAMING_SNAKE_CASE ( self ) -> str: '''simple docstring''' UpperCAmelCase : str = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( """snap-research/efficientformer-l1-300""" ) UpperCAmelCase : str = self.default_image_processor UpperCAmelCase : List[Any] = prepare_img() UpperCAmelCase : List[str] = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors="""tf""" ) # forward pass UpperCAmelCase : List[str] = model(**_SCREAMING_SNAKE_CASE , training=_SCREAMING_SNAKE_CASE ) # verify the logits UpperCAmelCase : Dict = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) UpperCAmelCase : List[Any] = tf.constant([-0.1312, 0.4353, -1.0499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) )
109
import math import unittest def A ( _UpperCAmelCase : int ) -> bool: '''simple docstring''' assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True class __lowerCAmelCase ( unittest.TestCase ): def _lowerCamelCase ( self : Tuple) -> Union[str, Any]: """simple docstring""" self.assertTrue(is_prime(2)) self.assertTrue(is_prime(3)) self.assertTrue(is_prime(5)) self.assertTrue(is_prime(7)) self.assertTrue(is_prime(11)) self.assertTrue(is_prime(13)) self.assertTrue(is_prime(17)) self.assertTrue(is_prime(19)) self.assertTrue(is_prime(23)) self.assertTrue(is_prime(29)) def _lowerCamelCase ( self : Optional[int]) -> Any: """simple docstring""" with self.assertRaises(A): is_prime(-19) self.assertFalse( is_prime(0) , 'Zero doesn\'t have any positive factors, primes must have exactly two.' , ) self.assertFalse( is_prime(1) , 'One only has 1 positive factor, primes must have exactly two.' , ) self.assertFalse(is_prime(2 * 2)) self.assertFalse(is_prime(2 * 3)) self.assertFalse(is_prime(3 * 3)) self.assertFalse(is_prime(3 * 5)) self.assertFalse(is_prime(3 * 5 * 7)) if __name__ == "__main__": unittest.main()
339
0
'''simple docstring''' def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ): while a != 0: __a , __a : int = b % a, a return b def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ): if gcd(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) != 1: __a : List[str] = F"""mod inverse of {a!r} and {m!r} does not exist""" raise ValueError(_SCREAMING_SNAKE_CASE ) __a , __a , __a : List[Any] = 1, 0, a __a , __a , __a : List[str] = 0, 1, m while va != 0: __a : Optional[Any] = ua // va __a , __a , __a , __a , __a , __a : Optional[int] = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
294
'''simple docstring''' import re from filelock import FileLock try: import nltk __lowercase : Optional[Any] = True except (ImportError, ModuleNotFoundError): __lowercase : Dict = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def lowerCamelCase (_SCREAMING_SNAKE_CASE : str ): re.sub('<n>' , '' , _SCREAMING_SNAKE_CASE ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(_SCREAMING_SNAKE_CASE ) )
294
1
"""simple docstring""" from .integrations import ( is_optuna_available, is_ray_available, is_sigopt_available, is_wandb_available, run_hp_search_optuna, run_hp_search_ray, run_hp_search_sigopt, run_hp_search_wandb, ) from .trainer_utils import ( HPSearchBackend, default_hp_space_optuna, default_hp_space_ray, default_hp_space_sigopt, default_hp_space_wandb, ) from .utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) class lowerCAmelCase_ : '''simple docstring''' _snake_case = 42 _snake_case = None @staticmethod def A__ ( ) -> str: raise NotImplementedError def A__ ( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) -> Union[str, Any]: raise NotImplementedError def A__ ( self , snake_case_ ) -> Dict: raise NotImplementedError def A__ ( self ) -> Optional[Any]: if not self.is_available(): raise RuntimeError( f"""You picked the {self.name} backend, but it is not installed. Run {self.pip_install()}.""" ) @classmethod def A__ ( cls ) -> Tuple: return f"""`pip install {cls.pip_package or cls.name}`""" class lowerCAmelCase_ ( A__ ): '''simple docstring''' _snake_case = '''optuna''' @staticmethod def A__ ( ) -> List[str]: return is_optuna_available() def A__ ( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) -> List[Any]: return run_hp_search_optuna(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) def A__ ( self , snake_case_ ) -> List[str]: return default_hp_space_optuna(snake_case_ ) class lowerCAmelCase_ ( A__ ): '''simple docstring''' _snake_case = '''ray''' _snake_case = '''\'ray[tune]\'''' @staticmethod def A__ ( ) -> Optional[int]: return is_ray_available() def A__ ( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) -> Union[str, Any]: return run_hp_search_ray(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) def A__ ( self , snake_case_ ) -> Optional[Any]: return default_hp_space_ray(snake_case_ ) class lowerCAmelCase_ ( A__ ): '''simple docstring''' _snake_case = '''sigopt''' @staticmethod def A__ ( ) -> Optional[Any]: return is_sigopt_available() def A__ ( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) -> List[Any]: return run_hp_search_sigopt(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) def A__ ( self , snake_case_ ) -> List[Any]: return default_hp_space_sigopt(snake_case_ ) class lowerCAmelCase_ ( A__ ): '''simple docstring''' _snake_case = '''wandb''' @staticmethod def A__ ( ) -> Any: return is_wandb_available() def A__ ( self , snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) -> Tuple: return run_hp_search_wandb(snake_case_ , snake_case_ , snake_case_ , **snake_case_ ) def A__ ( self , snake_case_ ) -> str: return default_hp_space_wandb(snake_case_ ) SCREAMING_SNAKE_CASE_ = { HPSearchBackend(backend.name): backend for backend in [OptunaBackend, RayTuneBackend, SigOptBackend, WandbBackend] } def lowercase (): __lowerCAmelCase = [backend for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() if backend.is_available()] if len(_lowerCAmelCase ) > 0: __lowerCAmelCase = available_backends[0].name if len(_lowerCAmelCase ) > 1: logger.info( f"""{len(_lowerCAmelCase )} hyperparameter search backends available. Using {name} as the default.""" ) return name raise RuntimeError( """No hyperparameter search backend available.\n""" + """\n""".join( f""" - To install {backend.name} run {backend.pip_install()}""" for backend in ALL_HYPERPARAMETER_SEARCH_BACKENDS.values() ) )
301
"""simple docstring""" def lowercase (_lowerCAmelCase = 100_0000 ): __lowerCAmelCase = 1 __lowerCAmelCase = 1 __lowerCAmelCase = {1: 1} for inputa in range(2 , _lowerCAmelCase ): __lowerCAmelCase = 0 __lowerCAmelCase = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __lowerCAmelCase = (3 * number) + 1 counter += 1 if inputa not in counters: __lowerCAmelCase = counter if counter > pre_counter: __lowerCAmelCase = inputa __lowerCAmelCase = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
301
1
'''simple docstring''' import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class __UpperCamelCase : def __init__( self, lowerCAmelCase, lowerCAmelCase=100, lowerCAmelCase=13, lowerCAmelCase=30, lowerCAmelCase=2, lowerCAmelCase=3, lowerCAmelCase=True, lowerCAmelCase=True, lowerCAmelCase=32, lowerCAmelCase=4, lowerCAmelCase=4, lowerCAmelCase=37, lowerCAmelCase="gelu", lowerCAmelCase=0.1, lowerCAmelCase=0.1, lowerCAmelCase=10, lowerCAmelCase=0.0_2, lowerCAmelCase=3, lowerCAmelCase=None, lowerCAmelCase=[0, 1, 2, 3], ): """simple docstring""" lowerCamelCase_ =parent lowerCamelCase_ =100 lowerCamelCase_ =batch_size lowerCamelCase_ =image_size lowerCamelCase_ =patch_size lowerCamelCase_ =num_channels lowerCamelCase_ =is_training lowerCamelCase_ =use_labels lowerCamelCase_ =hidden_size lowerCamelCase_ =num_hidden_layers lowerCamelCase_ =num_attention_heads lowerCamelCase_ =intermediate_size lowerCamelCase_ =hidden_act lowerCamelCase_ =hidden_dropout_prob lowerCamelCase_ =attention_probs_dropout_prob lowerCamelCase_ =type_sequence_label_size lowerCamelCase_ =initializer_range lowerCamelCase_ =scope lowerCamelCase_ =out_indices lowerCamelCase_ =num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase_ =(image_size // patch_size) ** 2 lowerCamelCase_ =num_patches + 1 def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase_ =None lowerCamelCase_ =None if self.use_labels: lowerCamelCase_ =ids_tensor([self.batch_size], self.type_sequence_label_size ) lowerCamelCase_ =ids_tensor([self.batch_size, self.image_size, self.image_size], self.num_labels ) lowerCamelCase_ =self.get_config() return config, pixel_values, labels, pixel_labels def lowercase__ ( self ): """simple docstring""" return BeitConfig( vocab_size=self.vocab_size, image_size=self.image_size, patch_size=self.patch_size, num_channels=self.num_channels, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, is_decoder=lowerCAmelCase, initializer_range=self.initializer_range, out_indices=self.out_indices, ) def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =BeitModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() lowerCamelCase_ =model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =BeitForMaskedImageModeling(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() lowerCamelCase_ =model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length - 1, self.vocab_size) ) def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =self.type_sequence_label_size lowerCamelCase_ =BeitForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() lowerCamelCase_ =model(lowerCAmelCase, labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase_ =1 lowerCamelCase_ =BeitForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() lowerCamelCase_ =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase_ =model(lowerCAmelCase, labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.type_sequence_label_size) ) def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =self.num_labels lowerCamelCase_ =BeitForSemanticSegmentation(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() lowerCamelCase_ =model(lowerCAmelCase ) self.parent.assertEqual( result.logits.shape, (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) lowerCamelCase_ =model(lowerCAmelCase, labels=lowerCAmelCase ) self.parent.assertEqual( result.logits.shape, (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =self.prepare_config_and_inputs() lowerCamelCase_, lowerCamelCase_, lowerCamelCase_, lowerCamelCase_ =config_and_inputs lowerCamelCase_ ={'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __UpperCamelCase ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): lowercase : Tuple =( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) lowercase : List[str] =( { 'feature-extraction': BeitModel, 'image-classification': BeitForImageClassification, 'image-segmentation': BeitForSemanticSegmentation, } if is_torch_available() else {} ) lowercase : List[Any] =False lowercase : Optional[int] =False lowercase : List[str] =False def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =BeitModelTester(self ) lowerCamelCase_ =ConfigTester(self, config_class=lowerCAmelCase, has_text_modality=lowerCAmelCase, hidden_size=37 ) def lowercase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''BEiT does not use inputs_embeds''' ) def lowercase__ ( self ): """simple docstring""" pass @require_torch_multi_gpu @unittest.skip(reason='''BEiT has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def lowercase__ ( self ): """simple docstring""" pass def lowercase__ ( self ): """simple docstring""" lowerCamelCase_, lowerCamelCase_ =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ =model_class(lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings(), (nn.Module) ) lowerCamelCase_ =model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase, nn.Linear ) ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_, lowerCamelCase_ =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ =model_class(lowerCAmelCase ) lowerCamelCase_ =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ =[*signature.parameters.keys()] lowerCamelCase_ =['''pixel_values'''] self.assertListEqual(arg_names[:1], lowerCAmelCase ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCAmelCase ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCAmelCase ) def lowercase__ ( self ): """simple docstring""" if not self.model_tester.is_training: return lowerCamelCase_, lowerCamelCase_ =self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase_ =True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(lowerCAmelCase ), BeitForMaskedImageModeling]: continue lowerCamelCase_ =model_class(lowerCAmelCase ) model.to(lowerCAmelCase ) model.train() lowerCamelCase_ =self._prepare_for_class(lowerCAmelCase, lowerCAmelCase, return_labels=lowerCAmelCase ) lowerCamelCase_ =model(**lowerCAmelCase ).loss loss.backward() def lowercase__ ( self ): """simple docstring""" lowerCamelCase_, lowerCamelCase_ =self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return lowerCamelCase_ =False lowerCamelCase_ =True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(lowerCAmelCase ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue lowerCamelCase_ =model_class(lowerCAmelCase ) model.gradient_checkpointing_enable() model.to(lowerCAmelCase ) model.train() lowerCamelCase_ =self._prepare_for_class(lowerCAmelCase, lowerCAmelCase, return_labels=lowerCAmelCase ) lowerCamelCase_ =model(**lowerCAmelCase ).loss loss.backward() def lowercase__ ( self ): """simple docstring""" lowerCamelCase_, lowerCamelCase_ =self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase_ =_config_zero_init(lowerCAmelCase ) for model_class in self.all_model_classes: lowerCamelCase_ =model_class(config=lowerCAmelCase ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item(), [0.0, 1.0], msg=f'''Parameter {name} of model {model_class} seems not properly initialized''', ) @slow def lowercase__ ( self ): """simple docstring""" for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ =BeitModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) def a_ ( ) -> Optional[int]: """simple docstring""" lowerCamelCase_ =Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def lowercase__ ( self ): """simple docstring""" return BeitImageProcessor.from_pretrained('''microsoft/beit-base-patch16-224''' ) if is_vision_available() else None @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =BeitForMaskedImageModeling.from_pretrained('''microsoft/beit-base-patch16-224-pt22k''' ).to(lowerCAmelCase ) lowerCamelCase_ =self.default_image_processor lowerCamelCase_ =prepare_img() lowerCamelCase_ =image_processor(images=lowerCAmelCase, return_tensors='''pt''' ).pixel_values.to(lowerCAmelCase ) # prepare bool_masked_pos lowerCamelCase_ =torch.ones((1, 196), dtype=torch.bool ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): lowerCamelCase_ =model(pixel_values=lowerCAmelCase, bool_masked_pos=lowerCAmelCase ) lowerCamelCase_ =outputs.logits # verify the logits lowerCamelCase_ =torch.Size((1, 196, 8_192) ) self.assertEqual(logits.shape, lowerCAmelCase ) lowerCamelCase_ =torch.tensor( [[-3.2_4_3_7, 0.5_0_7_2, -1_3.9_1_7_4], [-3.2_4_5_6, 0.4_9_4_8, -1_3.9_4_0_1], [-3.2_0_3_3, 0.5_1_2_1, -1_3.8_5_5_0]] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3], lowerCAmelCase, atol=1e-2 ) ) @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =BeitForImageClassification.from_pretrained('''microsoft/beit-base-patch16-224''' ).to(lowerCAmelCase ) lowerCamelCase_ =self.default_image_processor lowerCamelCase_ =prepare_img() lowerCamelCase_ =image_processor(images=lowerCAmelCase, return_tensors='''pt''' ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): lowerCamelCase_ =model(**lowerCAmelCase ) lowerCamelCase_ =outputs.logits # verify the logits lowerCamelCase_ =torch.Size((1, 1_000) ) self.assertEqual(logits.shape, lowerCAmelCase ) lowerCamelCase_ =torch.tensor([-1.2_3_8_5, -1.0_9_8_7, -1.0_1_0_8] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(logits[0, :3], lowerCAmelCase, atol=1e-4 ) ) lowerCamelCase_ =281 self.assertEqual(logits.argmax(-1 ).item(), lowerCAmelCase ) @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =BeitForImageClassification.from_pretrained('''microsoft/beit-large-patch16-224-pt22k-ft22k''' ).to( lowerCAmelCase ) lowerCamelCase_ =self.default_image_processor lowerCamelCase_ =prepare_img() lowerCamelCase_ =image_processor(images=lowerCAmelCase, return_tensors='''pt''' ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): lowerCamelCase_ =model(**lowerCAmelCase ) lowerCamelCase_ =outputs.logits # verify the logits lowerCamelCase_ =torch.Size((1, 21_841) ) self.assertEqual(logits.shape, lowerCAmelCase ) lowerCamelCase_ =torch.tensor([1.6_8_8_1, -0.2_7_8_7, 0.5_9_0_1] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(logits[0, :3], lowerCAmelCase, atol=1e-4 ) ) lowerCamelCase_ =2_396 self.assertEqual(logits.argmax(-1 ).item(), lowerCAmelCase ) @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =BeitForSemanticSegmentation.from_pretrained('''microsoft/beit-base-finetuned-ade-640-640''' ) lowerCamelCase_ =model.to(lowerCAmelCase ) lowerCamelCase_ =BeitImageProcessor(do_resize=lowerCAmelCase, size=640, do_center_crop=lowerCAmelCase ) lowerCamelCase_ =load_dataset('''hf-internal-testing/fixtures_ade20k''', split='''test''' ) lowerCamelCase_ =Image.open(ds[0]['''file'''] ) lowerCamelCase_ =image_processor(images=lowerCAmelCase, return_tensors='''pt''' ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): lowerCamelCase_ =model(**lowerCAmelCase ) lowerCamelCase_ =outputs.logits # verify the logits lowerCamelCase_ =torch.Size((1, 150, 160, 160) ) self.assertEqual(logits.shape, lowerCAmelCase ) lowerCamelCase_ =version.parse(PIL.__version__ ) < version.parse('''9.0.0''' ) if is_pillow_less_than_a: lowerCamelCase_ =torch.tensor( [ [[-4.9_2_2_5, -2.3_9_5_4, -3.0_5_2_2], [-2.8_8_2_2, -1.0_0_4_6, -1.7_5_6_1], [-2.9_5_4_9, -1.3_2_2_8, -2.1_3_4_7]], [[-5.8_1_6_8, -3.4_1_2_9, -4.0_7_7_8], [-3.8_6_5_1, -2.2_2_1_4, -3.0_2_7_7], [-3.8_3_5_6, -2.4_6_4_3, -3.3_5_3_5]], [[-0.0_0_7_8, 3.9_9_5_2, 4.0_7_5_4], [2.9_8_5_6, 4.6_9_4_4, 5.0_0_3_5], [3.2_4_1_3, 4.7_8_1_3, 4.9_9_6_9]], ], device=lowerCAmelCase, ) else: lowerCamelCase_ =torch.tensor( [ [[-4.8_9_6_0, -2.3_6_8_8, -3.0_3_5_5], [-2.8_4_7_8, -0.9_8_3_6, -1.7_4_1_8], [-2.9_4_4_9, -1.3_3_3_2, -2.1_4_5_6]], [[-5.8_0_8_1, -3.4_1_2_4, -4.1_0_0_6], [-3.8_5_6_1, -2.2_0_8_1, -3.0_3_2_3], [-3.8_3_6_5, -2.4_6_0_1, -3.3_6_6_9]], [[-0.0_3_0_9, 3.9_8_6_8, 4.0_5_4_0], [2.9_6_4_0, 4.6_8_7_7, 4.9_9_7_6], [3.2_0_8_1, 4.7_6_9_0, 4.9_9_4_2]], ], device=lowerCAmelCase, ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3], lowerCAmelCase, atol=1e-4 ) ) @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =BeitForSemanticSegmentation.from_pretrained('''microsoft/beit-base-finetuned-ade-640-640''' ) lowerCamelCase_ =model.to(lowerCAmelCase ) lowerCamelCase_ =BeitImageProcessor(do_resize=lowerCAmelCase, size=640, do_center_crop=lowerCAmelCase ) lowerCamelCase_ =load_dataset('''hf-internal-testing/fixtures_ade20k''', split='''test''' ) lowerCamelCase_ =Image.open(ds[0]['''file'''] ) lowerCamelCase_ =image_processor(images=lowerCAmelCase, return_tensors='''pt''' ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): lowerCamelCase_ =model(**lowerCAmelCase ) lowerCamelCase_ =outputs.logits.detach().cpu() lowerCamelCase_ =image_processor.post_process_semantic_segmentation(outputs=lowerCAmelCase, target_sizes=[(500, 300)] ) lowerCamelCase_ =torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape, lowerCAmelCase ) lowerCamelCase_ =image_processor.post_process_semantic_segmentation(outputs=lowerCAmelCase ) lowerCamelCase_ =torch.Size((160, 160) ) self.assertEqual(segmentation[0].shape, lowerCAmelCase )
6
'''simple docstring''' import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class __UpperCamelCase ( lowerCamelCase__ ): lowercase : List[str] =['image_processor', 'tokenizer'] lowercase : Optional[int] ='AutoImageProcessor' lowercase : List[str] ='AutoTokenizer' def __init__( self, lowerCAmelCase=None, lowerCAmelCase=None, **lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''', lowerCAmelCase, ) lowerCamelCase_ =kwargs.pop('''feature_extractor''' ) lowerCamelCase_ =image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(lowerCAmelCase, lowerCAmelCase ) lowerCamelCase_ =self.image_processor lowerCamelCase_ =False def __call__( self, *lowerCAmelCase, **lowerCAmelCase ): """simple docstring""" if self._in_target_context_manager: return self.current_processor(*lowerCAmelCase, **lowerCAmelCase ) lowerCamelCase_ =kwargs.pop('''images''', lowerCAmelCase ) lowerCamelCase_ =kwargs.pop('''text''', lowerCAmelCase ) if len(lowerCAmelCase ) > 0: lowerCamelCase_ =args[0] lowerCamelCase_ =args[1:] if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: lowerCamelCase_ =self.image_processor(lowerCAmelCase, *lowerCAmelCase, **lowerCAmelCase ) if text is not None: lowerCamelCase_ =self.tokenizer(lowerCAmelCase, **lowerCAmelCase ) if text is None: return inputs elif images is None: return encodings else: lowerCamelCase_ =encodings['''input_ids'''] return inputs def lowercase__ ( self, *lowerCAmelCase, **lowerCAmelCase ): """simple docstring""" return self.tokenizer.batch_decode(*lowerCAmelCase, **lowerCAmelCase ) def lowercase__ ( self, *lowerCAmelCase, **lowerCAmelCase ): """simple docstring""" return self.tokenizer.decode(*lowerCAmelCase, **lowerCAmelCase ) @contextmanager def lowercase__ ( self ): """simple docstring""" warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your images inputs, or in a separate call.''' ) lowerCamelCase_ =True lowerCamelCase_ =self.tokenizer yield lowerCamelCase_ =self.image_processor lowerCamelCase_ =False def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase=False, lowerCAmelCase=None ): """simple docstring""" if added_vocab is None: lowerCamelCase_ =self.tokenizer.get_added_vocab() lowerCamelCase_ ={} while tokens: lowerCamelCase_ =re.search(R'''<s_(.*?)>''', lowerCAmelCase, re.IGNORECASE ) if start_token is None: break lowerCamelCase_ =start_token.group(1 ) lowerCamelCase_ =re.search(Rf'''</s_{key}>''', lowerCAmelCase, re.IGNORECASE ) lowerCamelCase_ =start_token.group() if end_token is None: lowerCamelCase_ =tokens.replace(lowerCAmelCase, '''''' ) else: lowerCamelCase_ =end_token.group() lowerCamelCase_ =re.escape(lowerCAmelCase ) lowerCamelCase_ =re.escape(lowerCAmelCase ) lowerCamelCase_ =re.search(f'''{start_token_escaped}(.*?){end_token_escaped}''', lowerCAmelCase, re.IGNORECASE ) if content is not None: lowerCamelCase_ =content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node lowerCamelCase_ =self.tokenajson(lowerCAmelCase, is_inner_value=lowerCAmelCase, added_vocab=lowerCAmelCase ) if value: if len(lowerCAmelCase ) == 1: lowerCamelCase_ =value[0] lowerCamelCase_ =value else: # leaf nodes lowerCamelCase_ =[] for leaf in content.split(R'''<sep/>''' ): lowerCamelCase_ =leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": lowerCamelCase_ =leaf[1:-2] # for categorical special tokens output[key].append(lowerCAmelCase ) if len(output[key] ) == 1: lowerCamelCase_ =output[key][0] lowerCamelCase_ =tokens[tokens.find(lowerCAmelCase ) + len(lowerCAmelCase ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:], is_inner_value=lowerCAmelCase, added_vocab=lowerCAmelCase ) if len(lowerCAmelCase ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def lowercase__ ( self ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''', lowerCAmelCase, ) return self.image_processor_class @property def lowercase__ ( self ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''', lowerCAmelCase, ) return self.image_processor
6
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class UpperCAmelCase ( A_ ): def _SCREAMING_SNAKE_CASE (self : Optional[int] ) -> List[str]: '''simple docstring''' snake_case : Optional[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(snake_case__ , "tf_padding" ) ) self.parent.assertTrue(hasattr(snake_case__ , "depth_multiplier" ) ) class UpperCAmelCase : def __init__(self : List[str] , snake_case__ : Union[str, Any] , snake_case__ : Any=13 , snake_case__ : int=3 , snake_case__ : List[str]=32 , snake_case__ : Any=0.25 , snake_case__ : List[Any]=8 , snake_case__ : List[str]=True , snake_case__ : Any=10_24 , snake_case__ : List[Any]=32 , snake_case__ : Optional[int]="relu6" , snake_case__ : List[str]=0.1 , snake_case__ : Any=0.02 , snake_case__ : Union[str, Any]=True , snake_case__ : Dict=True , snake_case__ : Tuple=10 , snake_case__ : Tuple=None , ) -> Dict: '''simple docstring''' snake_case : Optional[int] = parent snake_case : List[str] = batch_size snake_case : Any = num_channels snake_case : List[Any] = image_size snake_case : List[str] = depth_multiplier snake_case : Optional[int] = min_depth snake_case : Optional[Any] = tf_padding snake_case : Tuple = int(last_hidden_size * depth_multiplier ) snake_case : List[Any] = output_stride snake_case : Union[str, Any] = hidden_act snake_case : Optional[Any] = classifier_dropout_prob snake_case : Dict = use_labels snake_case : Union[str, Any] = is_training snake_case : Optional[int] = num_labels snake_case : Optional[Any] = initializer_range snake_case : str = scope def _SCREAMING_SNAKE_CASE (self : List[str] ) -> List[str]: '''simple docstring''' snake_case : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case : Optional[int] = None snake_case : int = None if self.use_labels: snake_case : int = ids_tensor([self.batch_size] , self.num_labels ) snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) snake_case : Union[str, Any] = self.get_config() return config, pixel_values, labels, pixel_labels def _SCREAMING_SNAKE_CASE (self : List[str] ) -> Optional[Any]: '''simple docstring''' return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def _SCREAMING_SNAKE_CASE (self : int , snake_case__ : Optional[Any] , snake_case__ : Optional[int] , snake_case__ : Optional[Any] , snake_case__ : Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case : Optional[int] = MobileNetVaModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() snake_case : int = model(snake_case__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _SCREAMING_SNAKE_CASE (self : List[Any] , snake_case__ : Any , snake_case__ : Any , snake_case__ : Dict , snake_case__ : Optional[int] ) -> Any: '''simple docstring''' snake_case : int = self.num_labels snake_case : Tuple = MobileNetVaForImageClassification(snake_case__ ) model.to(snake_case__ ) model.eval() snake_case : Union[str, Any] = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _SCREAMING_SNAKE_CASE (self : str ) -> List[str]: '''simple docstring''' snake_case : str = self.prepare_config_and_inputs() snake_case , snake_case , snake_case , snake_case : List[Any] = config_and_inputs snake_case : Tuple = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase ( A_ ,A_ ,unittest.TestCase ): A__ : str = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () A__ : str = ( {"feature-extraction": MobileNetVaModel, "image-classification": MobileNetVaForImageClassification} if is_torch_available() else {} ) A__ : List[str] = False A__ : List[str] = False A__ : Union[str, Any] = False A__ : Optional[int] = False def _SCREAMING_SNAKE_CASE (self : List[Any] ) -> Any: '''simple docstring''' snake_case : List[Any] = MobileNetVaModelTester(self ) snake_case : int = MobileNetVaConfigTester(self , config_class=snake_case__ , has_text_modality=snake_case__ ) def _SCREAMING_SNAKE_CASE (self : List[str] ) -> Tuple: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="MobileNetV1 does not use inputs_embeds" ) def _SCREAMING_SNAKE_CASE (self : int ) -> Dict: '''simple docstring''' pass @unittest.skip(reason="MobileNetV1 does not support input and output embeddings" ) def _SCREAMING_SNAKE_CASE (self : List[str] ) -> Tuple: '''simple docstring''' pass @unittest.skip(reason="MobileNetV1 does not output attentions" ) def _SCREAMING_SNAKE_CASE (self : Any ) -> Dict: '''simple docstring''' pass def _SCREAMING_SNAKE_CASE (self : Dict ) -> Any: '''simple docstring''' snake_case , snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : int = model_class(snake_case__ ) snake_case : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case : str = [*signature.parameters.keys()] snake_case : Optional[int] = ["pixel_values"] self.assertListEqual(arg_names[:1] , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Optional[Any] ) -> List[Any]: '''simple docstring''' snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> str: '''simple docstring''' def check_hidden_states_output(snake_case__ : str , snake_case__ : Dict , snake_case__ : int ): snake_case : Optional[Any] = model_class(snake_case__ ) model.to(snake_case__ ) model.eval() with torch.no_grad(): snake_case : Any = model(**self._prepare_for_class(snake_case__ , snake_case__ ) ) snake_case : Optional[int] = outputs.hidden_states snake_case : int = 26 self.assertEqual(len(snake_case__ ) , snake_case__ ) snake_case , snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : List[str] = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case : Union[str, Any] = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : List[str] ) -> Any: '''simple docstring''' snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) @slow def _SCREAMING_SNAKE_CASE (self : Any ) -> Union[str, Any]: '''simple docstring''' for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case : str = MobileNetVaModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) def UpperCamelCase ( ): snake_case : Union[str, Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): @cached_property def _SCREAMING_SNAKE_CASE (self : Optional[int] ) -> Tuple: '''simple docstring''' return ( MobileNetVaImageProcessor.from_pretrained("google/mobilenet_v1_1.0_224" ) if is_vision_available() else None ) @slow def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' snake_case : List[str] = MobileNetVaForImageClassification.from_pretrained("google/mobilenet_v1_1.0_224" ).to(snake_case__ ) snake_case : Union[str, Any] = self.default_image_processor snake_case : Dict = prepare_img() snake_case : Tuple = image_processor(images=snake_case__ , return_tensors="pt" ).to(snake_case__ ) # forward pass with torch.no_grad(): snake_case : List[Any] = model(**snake_case__ ) # verify the logits snake_case : Optional[int] = torch.Size((1, 10_01) ) self.assertEqual(outputs.logits.shape , snake_case__ ) snake_case : List[Any] = torch.tensor([-4.1739, -1.1233, 3.1205] ).to(snake_case__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case__ , atol=1e-4 ) )
59
import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor lowerCAmelCase__ = logging.get_logger(__name__) class snake_case__(_UpperCamelCase ): """simple docstring""" def __init__( self : List[Any] , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Optional[Any] ): warnings.warn( "The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use PerceiverImageProcessor instead." , SCREAMING_SNAKE_CASE , ) super().__init__(*SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE )
130
0
"""simple docstring""" from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image __SCREAMING_SNAKE_CASE : Tuple = ['''text''', '''image''', '''audio'''] def lowerCAmelCase_( lowercase_ : List[str] ) -> Tuple: _lowerCamelCase = [] for input_type in input_types: if input_type == "text": inputs.append('''Text input''' ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir('''fixtures/tests_samples/COCO''' ) ) / '''000000039769.png''' ).resize((5_12, 5_12) ) ) elif input_type == "audio": inputs.append(torch.ones(30_00 ) ) elif isinstance(lowercase_ , lowercase_ ): inputs.append(create_inputs(lowercase_ ) ) else: raise ValueError(F"""Invalid type requested: {input_type}""" ) return inputs def lowerCAmelCase_( lowercase_ : List ) -> Optional[Any]: _lowerCamelCase = [] for output in outputs: if isinstance(lowercase_ , (str, AgentText) ): output_types.append('''text''' ) elif isinstance(lowercase_ , (Image.Image, AgentImage) ): output_types.append('''image''' ) elif isinstance(lowercase_ , (torch.Tensor, AgentAudio) ): output_types.append('''audio''' ) else: raise ValueError(F"""Invalid output: {output}""" ) return output_types @is_tool_test class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self ): self.assertTrue(hasattr(self.tool , '''inputs''' ) ) self.assertTrue(hasattr(self.tool , '''outputs''' ) ) _lowerCamelCase = self.tool.inputs for _input in inputs: if isinstance(_input , lowerCamelCase__ ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) _lowerCamelCase = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def snake_case__ ( self ): _lowerCamelCase = create_inputs(self.tool.inputs ) _lowerCamelCase = self.tool(*lowerCamelCase__ ) # There is a single output if len(self.tool.outputs ) == 1: _lowerCamelCase = [outputs] self.assertListEqual(output_types(lowerCamelCase__ ) , self.tool.outputs ) def snake_case__ ( self ): self.assertTrue(hasattr(self.tool , '''description''' ) ) self.assertTrue(hasattr(self.tool , '''default_checkpoint''' ) ) self.assertTrue(self.tool.description.startswith('''This is a tool that''' ) ) def snake_case__ ( self ): _lowerCamelCase = create_inputs(self.tool.inputs ) _lowerCamelCase = self.tool(*lowerCamelCase__ ) if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = [outputs] self.assertEqual(len(lowerCamelCase__ ) , len(self.tool.outputs ) ) for output, output_type in zip(lowerCamelCase__ , self.tool.outputs ): _lowerCamelCase = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(lowerCamelCase__ , lowerCamelCase__ ) ) def snake_case__ ( self ): _lowerCamelCase = create_inputs(self.tool.inputs ) _lowerCamelCase = [] for _input, input_type in zip(lowerCamelCase__ , self.tool.inputs ): if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error _lowerCamelCase = self.tool(*lowerCamelCase__ ) if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = [outputs] self.assertEqual(len(lowerCamelCase__ ) , len(self.tool.outputs ) )
73
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''alibaba-damo/mgp-str-base''': '''https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json''', } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] = 'mgp-str' def __init__( self , lowerCamelCase__=[3_2, 1_2_8] , lowerCamelCase__=4 , lowerCamelCase__=3 , lowerCamelCase__=2_7 , lowerCamelCase__=3_8 , lowerCamelCase__=5_0_2_5_7 , lowerCamelCase__=3_0_5_2_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=4.0 , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=1e-5 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=False , lowerCamelCase__=0.0_2 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = max_token_length _lowerCamelCase = num_character_labels _lowerCamelCase = num_bpe_labels _lowerCamelCase = num_wordpiece_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = mlp_ratio _lowerCamelCase = distilled _lowerCamelCase = layer_norm_eps _lowerCamelCase = drop_rate _lowerCamelCase = qkv_bias _lowerCamelCase = attn_drop_rate _lowerCamelCase = drop_path_rate _lowerCamelCase = output_aa_attentions _lowerCamelCase = initializer_range
73
1
import math def UpperCAmelCase_( a__ ): """simple docstring""" if not isinstance(a__ , a__ ): SCREAMING_SNAKE_CASE : List[str] = F"""Input value of [number={number}] must be an integer""" raise TypeError(a__ ) if number < 1: SCREAMING_SNAKE_CASE : int = F"""Input value of [number={number}] must be > 0""" raise ValueError(a__ ) elif number == 1: return 3 elif number == 2: return 5 else: SCREAMING_SNAKE_CASE : List[str] = int(math.log(number // 3 , 2 ) ) + 2 SCREAMING_SNAKE_CASE : Optional[Any] = [3, 5] SCREAMING_SNAKE_CASE : List[Any] = 2 SCREAMING_SNAKE_CASE : Dict = 3 for block in range(1 , a__ ): for _ in range(a__ ): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1] ) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(11): a__ : Optional[int] = 0 try: a__ : Tuple = proth(number) except ValueError: print(F"ValueError: there is no {number}th Proth number") continue print(F"The {number}th Proth number: {value}")
313
import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import DataLoader, RandomSampler from transformers import GPTaLMHeadModel def UpperCAmelCase_( a__=32 , a__=10 , a__=100 , a__=1_026 , a__=True , a__="data/tokenized_stories_train_wikitext103.jbl" , a__="igf_context_pairs.jbl" , ): """simple docstring""" set_seed(3 ) # generate train_data and objective_set SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[Any] = generate_datasets( a__ , a__ , number=a__ , min_len=1_026 , trim=a__ ) # keeps model same across runs set_seed(4 ) # model, lm_optimizer, lm_scheduler = recopy_gpt2(model, device, max_steps) # store original model weights # can we train on GPU? SCREAMING_SNAKE_CASE : str = torch.device('''cuda:0''' if torch.cuda.is_available() else '''cpu''' ) # load pretrained model SCREAMING_SNAKE_CASE : Dict = load_gpta('''gpt2''' ).to(a__ ) print('''computing perplexity on objective set''' ) SCREAMING_SNAKE_CASE : int = compute_perplexity(a__ , a__ , a__ ).item() print('''perplexity on objective set:''' , a__ ) # collect igf pairs and save to file demo.jbl collect_objective_set(a__ , a__ , a__ , a__ , a__ , a__ , a__ , a__ ) # clean up, delete model and data we don't need anymore del model, train_data, objective_set torch.cuda.empty_cache() def UpperCAmelCase_( a__ , a__=15 , a__=128 , a__=100 , a__="igf_model.pt" , ): """simple docstring""" set_seed(42 ) # Load pre-trained model SCREAMING_SNAKE_CASE : List[Any] = GPTaLMHeadModel.from_pretrained('''gpt2''' ) # Initialize secondary learner to use embedding weights of model SCREAMING_SNAKE_CASE : str = SecondaryLearner(a__ ) # Train secondary learner SCREAMING_SNAKE_CASE : Union[str, Any] = train_secondary_learner( a__ , a__ , max_epochs=a__ , batch_size=a__ , eval_freq=100 , igf_model_path=a__ , ) del model, secondary_learner_train_data torch.cuda.empty_cache() return secondary_learner def UpperCAmelCase_( a__ , a__ , a__ , a__=32 , a__=1_000 , a__=16 , a__=1.0 , a__=recopy_gpta , a__=None , a__=10 , a__="gpt2_finetuned.pt" , ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = torch.device('''cuda:0''' if torch.cuda.is_available() else '''cpu''' ) SCREAMING_SNAKE_CASE : Optional[int] = RandomSampler(a__ ) SCREAMING_SNAKE_CASE : Dict = DataLoader(a__ , sampler=a__ ) SCREAMING_SNAKE_CASE : Tuple = max_steps // (len(a__ )) + 1 SCREAMING_SNAKE_CASE : int = 0 SCREAMING_SNAKE_CASE : Union[str, Any] = torch.zeros((1, context_len) , dtype=torch.long , device=a__ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : str = recopy_model(a__ , a__ , a__ ) model.train() if secondary_learner is not None: secondary_learner.to(a__ ) secondary_learner.eval() SCREAMING_SNAKE_CASE : List[str] = [] SCREAMING_SNAKE_CASE : Optional[int] = 0 SCREAMING_SNAKE_CASE : Optional[int] = [] SCREAMING_SNAKE_CASE : Tuple = [] # Compute the performance of the transformer model at the beginning SCREAMING_SNAKE_CASE : str = compute_perplexity(a__ , a__ , a__ ) test_perps.append(a__ ) print('''Test perplexity, step''' , a__ , ''':''' , a__ ) for epoch in range(int(a__ ) ): for step, example in enumerate(a__ ): torch.cuda.empty_cache() SCREAMING_SNAKE_CASE : Union[str, Any] = random.randint(0 , example.size(2 ) - context_len - 1 ) SCREAMING_SNAKE_CASE : Optional[int] = example[0, 0, start : start + context_len] lm_optimizer.zero_grad() SCREAMING_SNAKE_CASE : Optional[Any] = model(a__ , labels=a__ ) SCREAMING_SNAKE_CASE : Union[str, Any] = True if secondary_learner is not None: SCREAMING_SNAKE_CASE : List[str] = secondary_learner.forward( torch.tensor(a__ , dtype=torch.long , device=a__ ).unsqueeze(0 ) )[0].item() observed_qs.append(float(a__ ) ) # Here we implement the simple non-constant threshold for the predicted IG(X) value # We will decay the selectivity of our secondary learner filter from # 1 standard deviation above average to 1 below average after 10 batches. if global_step == 10: SCREAMING_SNAKE_CASE : Dict = -1 if predicted_q < threshold: SCREAMING_SNAKE_CASE : str = False # If we passed the filter, add the context to the batch! if do_backprop: contexts.append(np.array(context.cpu() ) ) SCREAMING_SNAKE_CASE : List[str] = outputs[0] lm_loss.backward() examples += 1 del outputs # Once the batch is filled with enough contexts, backprop on the batch. if examples == batch_size: torch.cuda.empty_cache() SCREAMING_SNAKE_CASE : Any = 0 # Do LM backprop torch.nn.utils.clip_grad_norm_(model.parameters() , 3.0 ) lm_optimizer.step() lm_scheduler.step() # Update learning rate schedule global_step += 1 # Compute the performance of the transformer model at this batch if global_step % eval_interval == 0: SCREAMING_SNAKE_CASE : str = compute_perplexity(a__ , a__ , a__ ) test_perps.append(a__ ) print('''Test perplexity, step''' , a__ , ''':''' , a__ ) # Break out of the loop after 60 batches if max_steps > 0 and global_step > 60: break if max_steps > 0 and global_step > 60: break # save finetuned transformer model torch.save(model.state_dict() , a__ ) torch.cuda.empty_cache() # Do some cleaning up so we can reinitialize for the next run of this function del lm_optimizer del lm_scheduler return model def UpperCAmelCase_( ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser(description='''Fine-tune a transformer model with IGF on a language modeling task''' ) # Required parameters parser.add_argument( '''--data_dir''' , default=a__ , type=a__ , required=a__ , help='''The input data dir. Should contain data files for WikiText.''' , ) parser.add_argument( '''--model_name_or_path''' , default=a__ , type=a__ , required=a__ , help='''Path to pretrained model or model identifier from huggingface.co/models''' , ) parser.add_argument( '''--data_file''' , type=a__ , default=a__ , help=( '''A jbl file containing tokenized data which can be split as objective dataset, ''' '''train_dataset and test_dataset.''' ) , ) parser.add_argument( '''--igf_data_file''' , type=a__ , default=a__ , help='''A jbl file containing the context and information gain pairs to train secondary learner.''' , ) parser.add_argument( '''--output_dir''' , default=a__ , type=a__ , required=a__ , help='''The output directory where the final fine-tuned model is stored.''' , ) parser.add_argument( '''--tokenizer_name''' , default=a__ , type=a__ , help='''Pretrained tokenizer name or path if not the same as model_name''' , ) parser.add_argument('''--seed''' , type=a__ , default=a__ , help='''A seed for reproducible training.''' ) parser.add_argument( '''--context_len''' , default=32 , type=a__ , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--size_objective_set''' , default=100 , type=a__ , help='''number of articles that are long enough to be used as our objective set''' , ) parser.add_argument( '''--eval_freq''' , default=100 , type=a__ , help='''secondary model evaluation is triggered at eval_freq''' ) parser.add_argument('''--max_steps''' , default=1_000 , type=a__ , help='''To calculate training epochs''' ) parser.add_argument( '''--secondary_learner_batch_size''' , default=128 , type=a__ , help='''batch size of training data for secondary learner''' , ) parser.add_argument( '''--batch_size''' , default=16 , type=a__ , help='''batch size of training data of language model(gpt2) ''' ) parser.add_argument( '''--eval_interval''' , default=10 , type=a__ , help=( '''decay the selectivity of our secondary learner filter from''' '''1 standard deviation above average to 1 below average after 10 batches''' ) , ) parser.add_argument( '''--number''' , default=100 , type=a__ , help='''The number of examples split to be used as objective_set/test_data''' ) parser.add_argument( '''--min_len''' , default=1_026 , type=a__ , help='''The minimum length of the article to be used as objective set''' ) parser.add_argument( '''--secondary_learner_max_epochs''' , default=15 , type=a__ , help='''number of epochs to train secondary learner''' ) parser.add_argument('''--trim''' , default=a__ , type=a__ , help='''truncate the example if it exceeds context length''' ) parser.add_argument( '''--threshold''' , default=1.0 , type=a__ , help=( '''The threshold value used by secondary learner to filter the train_data and allow only''' ''' informative data as input to the model''' ) , ) parser.add_argument('''--finetuned_model_name''' , default='''gpt2_finetuned.pt''' , type=a__ , help='''finetuned_model_name''' ) parser.add_argument( '''--recopy_model''' , default=a__ , type=a__ , help='''Reset the model to the original pretrained GPT-2 weights after each iteration''' , ) # function calls # Collecting *n* pairs of context and information gain(X, IG(X)) for training the secondary learner generate_n_pairs( context_len=32 , max_steps=10 , size_objective_set=100 , min_len=1_026 , trim=a__ , data_file='''data/tokenized_stories_train_wikitext103.jbl''' , igf_data_file='''igf_context_pairs.jbl''' , ) # Load train data for secondary learner SCREAMING_SNAKE_CASE : List[Any] = joblib.load('''data/IGF_values.jbl''' ) # Train secondary learner SCREAMING_SNAKE_CASE : Tuple = training_secondary_learner( a__ , secondary_learner_max_epochs=15 , secondary_learner_batch_size=128 , eval_freq=100 , igf_model_path='''igf_model.pt''' , ) # load pretrained gpt2 model SCREAMING_SNAKE_CASE : Optional[Any] = GPTaLMHeadModel.from_pretrained('''gpt2''' ) set_seed(42 ) # Generate train and test data to train and evaluate gpt2 model SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Any = generate_datasets( context_len=32 , file='''data/tokenized_stories_train_wikitext103.jbl''' , number=100 , min_len=1_026 , trim=a__ ) # fine-tuning of the gpt2 model using igf (Information Gain Filtration) finetune( a__ , a__ , a__ , context_len=32 , max_steps=1_000 , batch_size=16 , threshold=1.0 , recopy_model=a__ , secondary_learner=a__ , eval_interval=10 , finetuned_model_name='''gpt2_finetuned.pt''' , ) if __name__ == "__main__": main()
313
1
"""simple docstring""" import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def __lowerCAmelCase (_UpperCamelCase ): if "cls_token" in name: __lowerCAmelCase : Any = name.replace('cls_token' , 'vit.embeddings.cls_token' ) if "mask_token" in name: __lowerCAmelCase : Optional[Any] = name.replace('mask_token' , 'decoder.mask_token' ) if "decoder_pos_embed" in name: __lowerCAmelCase : Optional[int] = name.replace('decoder_pos_embed' , 'decoder.decoder_pos_embed' ) if "pos_embed" in name and "decoder" not in name: __lowerCAmelCase : int = name.replace('pos_embed' , 'vit.embeddings.position_embeddings' ) if "patch_embed.proj" in name: __lowerCAmelCase : Dict = name.replace('patch_embed.proj' , 'vit.embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: __lowerCAmelCase : Tuple = name.replace('patch_embed.norm' , 'vit.embeddings.norm' ) if "decoder_blocks" in name: __lowerCAmelCase : Optional[int] = name.replace('decoder_blocks' , 'decoder.decoder_layers' ) if "blocks" in name: __lowerCAmelCase : List[str] = name.replace('blocks' , 'vit.encoder.layer' ) if "attn.proj" in name: __lowerCAmelCase : Optional[Any] = name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: __lowerCAmelCase : Tuple = name.replace('attn' , 'attention.self' ) if "norm1" in name: __lowerCAmelCase : Any = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: __lowerCAmelCase : Union[str, Any] = name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: __lowerCAmelCase : Optional[int] = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: __lowerCAmelCase : Union[str, Any] = name.replace('mlp.fc2' , 'output.dense' ) if "decoder_embed" in name: __lowerCAmelCase : Optional[Any] = name.replace('decoder_embed' , 'decoder.decoder_embed' ) if "decoder_norm" in name: __lowerCAmelCase : str = name.replace('decoder_norm' , 'decoder.decoder_norm' ) if "decoder_pred" in name: __lowerCAmelCase : List[Any] = name.replace('decoder_pred' , 'decoder.decoder_pred' ) if "norm.weight" in name and "decoder" not in name: __lowerCAmelCase : Any = name.replace('norm.weight' , 'vit.layernorm.weight' ) if "norm.bias" in name and "decoder" not in name: __lowerCAmelCase : Optional[int] = name.replace('norm.bias' , 'vit.layernorm.bias' ) return name def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase ): for key in orig_state_dict.copy().keys(): __lowerCAmelCase : Optional[Any] = orig_state_dict.pop(A__ ) if "qkv" in key: __lowerCAmelCase : List[str] = key.split('.' ) __lowerCAmelCase : Dict = int(key_split[1] ) if "decoder_blocks" in key: __lowerCAmelCase : Any = config.decoder_hidden_size __lowerCAmelCase : Union[str, Any] = 'decoder.decoder_layers.' if "weight" in key: __lowerCAmelCase : Dict = val[:dim, :] __lowerCAmelCase : Dict = val[dim : dim * 2, :] __lowerCAmelCase : List[str] = val[-dim:, :] elif "bias" in key: __lowerCAmelCase : Any = val[:dim] __lowerCAmelCase : str = val[dim : dim * 2] __lowerCAmelCase : Dict = val[-dim:] else: __lowerCAmelCase : Tuple = config.hidden_size __lowerCAmelCase : Union[str, Any] = 'vit.encoder.layer.' if "weight" in key: __lowerCAmelCase : str = val[:dim, :] __lowerCAmelCase : List[str] = val[dim : dim * 2, :] __lowerCAmelCase : List[Any] = val[-dim:, :] elif "bias" in key: __lowerCAmelCase : Dict = val[:dim] __lowerCAmelCase : List[str] = val[dim : dim * 2] __lowerCAmelCase : Union[str, Any] = val[-dim:] else: __lowerCAmelCase : str = val return orig_state_dict def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : Optional[int] = ViTMAEConfig() if "large" in checkpoint_url: __lowerCAmelCase : Dict = 1024 __lowerCAmelCase : Optional[Any] = 4096 __lowerCAmelCase : Optional[int] = 24 __lowerCAmelCase : Any = 16 elif "huge" in checkpoint_url: __lowerCAmelCase : str = 14 __lowerCAmelCase : Optional[int] = 1280 __lowerCAmelCase : List[Any] = 5120 __lowerCAmelCase : Tuple = 32 __lowerCAmelCase : Any = 16 __lowerCAmelCase : List[str] = ViTMAEForPreTraining(A__ ) __lowerCAmelCase : int = torch.hub.load_state_dict_from_url(A__ , map_location='cpu' )['model'] __lowerCAmelCase : Optional[Any] = ViTMAEImageProcessor(size=config.image_size ) __lowerCAmelCase : List[str] = convert_state_dict(A__ , A__ ) model.load_state_dict(A__ ) model.eval() __lowerCAmelCase : Any = 'https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg' __lowerCAmelCase : str = Image.open(requests.get(A__ , stream=A__ ).raw ) __lowerCAmelCase : Optional[int] = ViTMAEImageProcessor(size=config.image_size ) __lowerCAmelCase : int = image_processor(images=A__ , return_tensors='pt' ) # forward pass torch.manual_seed(2 ) __lowerCAmelCase : Tuple = model(**A__ ) __lowerCAmelCase : List[Any] = outputs.logits if "large" in checkpoint_url: __lowerCAmelCase : Any = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: __lowerCAmelCase : str = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: __lowerCAmelCase : Dict = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , A__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(A__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(A__ ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth""", type=str, help="""URL of the checkpoint you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) lowerCamelCase__ = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
358
"""simple docstring""" class A__ : def __init__( self , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : List[Any] = size __lowerCAmelCase : str = [0] * size __lowerCAmelCase : Any = [0] * size @staticmethod def __lowerCamelCase ( _SCREAMING_SNAKE_CASE ): return index | (index + 1) @staticmethod def __lowerCamelCase ( _SCREAMING_SNAKE_CASE ): return (index & (index + 1)) - 1 def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Any = value while index < self.size: __lowerCAmelCase : Dict = self.get_prev(_SCREAMING_SNAKE_CASE ) + 1 if current_left_border == index: __lowerCAmelCase : Any = value else: __lowerCAmelCase : Any = max(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[Any] = self.get_next(_SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): right -= 1 # Because of right is exclusive __lowerCAmelCase : Optional[int] = 0 while left <= right: __lowerCAmelCase : Optional[int] = self.get_prev(_SCREAMING_SNAKE_CASE ) if left <= current_left: __lowerCAmelCase : Optional[Any] = max(_SCREAMING_SNAKE_CASE , self.tree[right] ) __lowerCAmelCase : Optional[Any] = current_left else: __lowerCAmelCase : List[str] = max(_SCREAMING_SNAKE_CASE , self.arr[right] ) right -= 1 return result if __name__ == "__main__": import doctest doctest.testmod()
182
0
'''simple docstring''' def __lowerCamelCase ( ) -> Union[str, Any]: return [list(range(1000 - i , -1000 - i , -1 ) ) for i in range(1000 )] __lowerCAmelCase = generate_large_matrix() __lowerCAmelCase = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def __lowerCamelCase ( lowerCAmelCase_ ) -> List[str]: assert all(row == sorted(lowerCamelCase__ , reverse=lowerCamelCase__ ) for row in grid ) assert all(list(lowerCamelCase__ ) == sorted(lowerCamelCase__ , reverse=lowerCamelCase__ ) for col in zip(*lowerCamelCase__ ) ) def __lowerCamelCase ( lowerCAmelCase_ ) -> List[Any]: _a : Tuple = 0 _a : int = len(lowerCamelCase__ ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _a : List[Any] = (left + right) // 2 _a : Optional[int] = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _a : str = mid + 1 else: _a : List[Any] = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(lowerCamelCase__ ) def __lowerCamelCase ( lowerCAmelCase_ ) -> Union[str, Any]: _a : Tuple = 0 _a : List[str] = len(grid[0] ) for i in range(len(lowerCamelCase__ ) ): _a : str = find_negative_index(grid[i][:bound] ) total += bound return (len(lowerCamelCase__ ) * len(grid[0] )) - total def __lowerCamelCase ( lowerCAmelCase_ ) -> List[str]: return len([number for row in grid for number in row if number < 0] ) def __lowerCamelCase ( lowerCAmelCase_ ) -> List[Any]: _a : str = 0 for row in grid: for i, number in enumerate(lowerCamelCase__ ): if number < 0: total += len(lowerCamelCase__ ) - i break return total def __lowerCamelCase ( ) -> Tuple: from timeit import timeit print('Running benchmarks' ) _a : str = ( 'from __main__ import count_negatives_binary_search, ' 'count_negatives_brute_force, count_negatives_brute_force_with_break, grid' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _a : str = timeit(f"""{func}(grid=grid)""" , setup=lowerCamelCase__ , number=500 ) print(f"""{func}() took {time:0.4f} seconds""" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
89
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class __lowercase ( unittest.TestCase ): """simple docstring""" def __init__( self , A , A=7 , A=3 , A=30 , A=4_00 , A=True , A=None , A=True , A=[0.5, 0.5, 0.5] , A=[0.5, 0.5, 0.5] , A=True , A=1 / 2_55 , A=True , ) -> str: '''simple docstring''' lowerCamelCase = size if size is not None else {"""shortest_edge""": 18, """longest_edge""": 13_33} lowerCamelCase = parent lowerCamelCase = batch_size lowerCamelCase = num_channels lowerCamelCase = min_resolution lowerCamelCase = max_resolution lowerCamelCase = do_resize lowerCamelCase = size lowerCamelCase = do_normalize lowerCamelCase = image_mean lowerCamelCase = image_std lowerCamelCase = do_rescale lowerCamelCase = rescale_factor lowerCamelCase = do_pad def __A ( self ) -> List[Any]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __A ( self , A , A=False ) -> List[Any]: '''simple docstring''' if not batched: lowerCamelCase = image_inputs[0] if isinstance(A , Image.Image ): lowerCamelCase , lowerCamelCase = image.size else: lowerCamelCase , lowerCamelCase = image.shape[1], image.shape[2] if w < h: lowerCamelCase = int(self.size["""shortest_edge"""] * h / w ) lowerCamelCase = self.size["""shortest_edge"""] elif w > h: lowerCamelCase = self.size["""shortest_edge"""] lowerCamelCase = int(self.size["""shortest_edge"""] * w / h ) else: lowerCamelCase = self.size["""shortest_edge"""] lowerCamelCase = self.size["""shortest_edge"""] else: lowerCamelCase = [] for image in image_inputs: lowerCamelCase , lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCamelCase = max(A , key=lambda A : item[0] )[0] lowerCamelCase = max(A , key=lambda A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class __lowercase ( a_ , unittest.TestCase ): """simple docstring""" UpperCamelCase : Union[str, Any] = YolosImageProcessor if is_vision_available() else None def __A ( self ) -> Union[str, Any]: '''simple docstring''' lowerCamelCase = YolosImageProcessingTester(self ) @property def __A ( self ) -> List[Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def __A ( self ) -> str: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A , """image_mean""" ) ) self.assertTrue(hasattr(A , """image_std""" ) ) self.assertTrue(hasattr(A , """do_normalize""" ) ) self.assertTrue(hasattr(A , """do_resize""" ) ) self.assertTrue(hasattr(A , """size""" ) ) def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18, """longest_edge""": 13_33} ) self.assertEqual(image_processor.do_pad , A ) lowerCamelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=A ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42, """longest_edge""": 84} ) self.assertEqual(image_processor.do_pad , A ) def __A ( self ) -> Union[str, Any]: '''simple docstring''' pass def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A , Image.Image ) # Test not batched input lowerCamelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A , batched=A ) lowerCamelCase = image_processing(A , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __A ( self ) -> List[str]: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , numpify=A ) for image in image_inputs: self.assertIsInstance(A , np.ndarray ) # Test not batched input lowerCamelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase = image_processing(A , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test not batched input lowerCamelCase = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCamelCase = image_processing(A , return_tensors="""pt""" ).pixel_values lowerCamelCase , lowerCamelCase = self.image_processor_tester.get_expected_values(A , batched=A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __A ( self ) -> Any: '''simple docstring''' lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) lowerCamelCase = self.image_processing_class(do_resize=A , do_normalize=A , do_rescale=A ) # create random PyTorch tensors lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=A , torchify=A ) for image in image_inputs: self.assertIsInstance(A , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors lowerCamelCase = image_processing_a.pad(A , return_tensors="""pt""" ) lowerCamelCase = image_processing_a(A , return_tensors="""pt""" ) self.assertTrue( torch.allclose(encoded_images_with_method["""pixel_values"""] , encoded_images["""pixel_values"""] , atol=1e-4 ) ) @slow def __A ( self ) -> List[Any]: '''simple docstring''' lowerCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_annotations.txt""" , """r""" ) as f: lowerCamelCase = json.loads(f.read() ) lowerCamelCase = {"""image_id""": 3_97_69, """annotations""": target} # encode them lowerCamelCase = YolosImageProcessor.from_pretrained("""hustvl/yolos-small""" ) lowerCamelCase = image_processing(images=A , annotations=A , return_tensors="""pt""" ) # verify pixel values lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["""pixel_values"""].shape , A ) lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowerCamelCase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , A ) ) # verify boxes lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , A ) lowerCamelCase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , A , atol=1e-3 ) ) # verify image_id lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , A ) ) # verify is_crowd lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , A ) ) # verify class_labels lowerCamelCase = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , A ) ) # verify orig_size lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , A ) ) # verify size lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , A ) ) @slow def __A ( self ) -> List[Any]: '''simple docstring''' lowerCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt""" , """r""" ) as f: lowerCamelCase = json.loads(f.read() ) lowerCamelCase = {"""file_name""": """000000039769.png""", """image_id""": 3_97_69, """segments_info""": target} lowerCamelCase = pathlib.Path("""./tests/fixtures/tests_samples/COCO/coco_panoptic""" ) # encode them lowerCamelCase = YolosImageProcessor(format="""coco_panoptic""" ) lowerCamelCase = image_processing(images=A , annotations=A , masks_path=A , return_tensors="""pt""" ) # verify pixel values lowerCamelCase = torch.Size([1, 3, 8_00, 10_66] ) self.assertEqual(encoding["""pixel_values"""].shape , A ) lowerCamelCase = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , A , atol=1e-4 ) ) # verify area lowerCamelCase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , A ) ) # verify boxes lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , A ) lowerCamelCase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , A , atol=1e-3 ) ) # verify image_id lowerCamelCase = torch.tensor([3_97_69] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , A ) ) # verify is_crowd lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , A ) ) # verify class_labels lowerCamelCase = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , A ) ) # verify masks lowerCamelCase = 82_28_73 self.assertEqual(encoding["""labels"""][0]["""masks"""].sum().item() , A ) # verify orig_size lowerCamelCase = torch.tensor([4_80, 6_40] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , A ) ) # verify size lowerCamelCase = torch.tensor([8_00, 10_66] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , A ) )
252
0
'''simple docstring''' import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=32 , _lowerCamelCase=3 , _lowerCamelCase=4 , _lowerCamelCase=[10, 20, 30, 40] , _lowerCamelCase=[2, 2, 3, 2] , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=10 , _lowerCamelCase=0.02 , _lowerCamelCase=["stage2", "stage3", "stage4"] , _lowerCamelCase=[2, 3, 4] , _lowerCamelCase=None , ) -> Optional[Any]: A_ : Dict = parent A_ : Optional[int] = batch_size A_ : List[Any] = image_size A_ : Optional[Any] = num_channels A_ : List[str] = num_stages A_ : List[str] = hidden_sizes A_ : List[str] = depths A_ : str = is_training A_ : Tuple = use_labels A_ : Any = intermediate_size A_ : List[str] = hidden_act A_ : str = num_labels A_ : Union[str, Any] = initializer_range A_ : Optional[int] = out_features A_ : Optional[Any] = out_indices A_ : Any = scope def UpperCAmelCase_ ( self ) -> str: A_ : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Tuple = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.num_labels ) A_ : List[str] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> List[Any]: return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : Tuple = ConvNextVaModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : List[str] = model(_lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Optional[Any]: A_ : int = ConvNextVaForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Union[str, Any] = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Any: A_ : Union[str, Any] = ConvNextVaBackbone(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Any = model(_lowerCamelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None A_ : Union[str, Any] = None A_ : Any = ConvNextVaBackbone(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : List[str] = model(_lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : int = self.prepare_config_and_inputs() A_ , A_ , A_ : Tuple = config_and_inputs A_ : int = {"""pixel_values""": pixel_values} return config, inputs_dict def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[int] = self.prepare_config_and_inputs() A_ , A_ , A_ : Dict = config_and_inputs A_ : List[str] = {"""pixel_values""": pixel_values, """labels""": labels} return config, inputs_dict @require_torch class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) lowerCamelCase = ( {'''feature-extraction''': ConvNextVaModel, '''image-classification''': ConvNextVaForImageClassification} if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> str: A_ : List[str] = ConvNextVaModelTester(self ) A_ : Union[str, Any] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self ) -> List[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase_ ( self ) -> str: return @unittest.skip(reason="""ConvNextV2 does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Dict: pass @unittest.skip(reason="""ConvNextV2 does not support input and output embeddings""" ) def UpperCAmelCase_ ( self ) -> List[Any]: pass @unittest.skip(reason="""ConvNextV2 does not use feedforward chunking""" ) def UpperCAmelCase_ ( self ) -> List[Any]: pass def UpperCAmelCase_ ( self ) -> Optional[int]: if not self.model_tester.is_training: return for model_class in self.all_model_classes: A_ , A_ : int = self.model_tester.prepare_config_and_inputs_with_labels() A_ : int = True if model_class.__name__ in [ *get_values(_lowerCamelCase ), *get_values(_lowerCamelCase ), ]: continue A_ : Optional[Any] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Any = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> Optional[int]: if not self.model_tester.is_training: return for model_class in self.all_model_classes: A_ , A_ : Optional[int] = self.model_tester.prepare_config_and_inputs_with_labels() A_ : Union[str, Any] = False A_ : List[Any] = True if ( model_class.__name__ in [*get_values(_lowerCamelCase ), *get_values(_lowerCamelCase )] or not model_class.supports_gradient_checkpointing ): continue A_ : Optional[Any] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.gradient_checkpointing_enable() model.train() A_ : Dict = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : str = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> Optional[int]: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Dict = model_class(_lowerCamelCase ) A_ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : List[str] = [*signature.parameters.keys()] A_ : Optional[int] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: def check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): A_ : Tuple = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() with torch.no_grad(): A_ : int = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) ) A_ : Optional[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : str = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Optional[int] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : int = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def UpperCAmelCase_ ( self ) -> Dict: for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Union[str, Any] = ConvNextVaModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> List[str]: """simple docstring""" A_ : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> Dict: return AutoImageProcessor.from_pretrained("""facebook/convnextv2-tiny-1k-224""" ) if is_vision_available() else None @slow def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : Any = ConvNextVaForImageClassification.from_pretrained("""facebook/convnextv2-tiny-1k-224""" ).to(_lowerCamelCase ) A_ : str = self.default_image_processor A_ : str = prepare_img() A_ : Optional[int] = preprocessor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) # forward pass with torch.no_grad(): A_ : Union[str, Any] = model(**_lowerCamelCase ) # verify the logits A_ : List[Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : Optional[Any] = torch.tensor([0.9996, 0.1966, -0.4386] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 ) )
164
'''simple docstring''' UpperCamelCase__ : Optional[Any] = [ 999, 800, 799, 600, 599, 500, 400, 399, 377, 355, 333, 311, 288, 266, 244, 222, 200, 199, 177, 155, 133, 111, 88, 66, 44, 22, 0, ] UpperCamelCase__ : List[Any] = [ 999, 976, 952, 928, 905, 882, 858, 857, 810, 762, 715, 714, 572, 429, 428, 286, 285, 238, 190, 143, 142, 118, 95, 71, 47, 24, 0, ] UpperCamelCase__ : Optional[Any] = [ 999, 988, 977, 966, 955, 944, 933, 922, 911, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 350, 300, 299, 266, 233, 200, 199, 179, 159, 140, 120, 100, 99, 88, 77, 66, 55, 44, 33, 22, 11, 0, ] UpperCamelCase__ : List[Any] = [ 999, 995, 992, 989, 985, 981, 978, 975, 971, 967, 964, 961, 957, 956, 951, 947, 942, 937, 933, 928, 923, 919, 914, 913, 908, 903, 897, 892, 887, 881, 876, 871, 870, 864, 858, 852, 846, 840, 834, 828, 827, 820, 813, 806, 799, 792, 785, 784, 777, 770, 763, 756, 749, 742, 741, 733, 724, 716, 707, 699, 698, 688, 677, 666, 656, 655, 645, 634, 623, 613, 612, 598, 584, 570, 569, 555, 541, 527, 526, 505, 484, 483, 462, 440, 439, 396, 395, 352, 351, 308, 307, 264, 263, 220, 219, 176, 132, 88, 44, 0, ] UpperCamelCase__ : Optional[Any] = [ 999, 997, 995, 992, 990, 988, 986, 984, 981, 979, 977, 975, 972, 970, 968, 966, 964, 961, 959, 957, 956, 954, 951, 949, 946, 944, 941, 939, 936, 934, 931, 929, 926, 924, 921, 919, 916, 914, 913, 910, 907, 905, 902, 899, 896, 893, 891, 888, 885, 882, 879, 877, 874, 871, 870, 867, 864, 861, 858, 855, 852, 849, 846, 843, 840, 837, 834, 831, 828, 827, 824, 821, 817, 814, 811, 808, 804, 801, 798, 795, 791, 788, 785, 784, 780, 777, 774, 770, 766, 763, 760, 756, 752, 749, 746, 742, 741, 737, 733, 730, 726, 722, 718, 714, 710, 707, 703, 699, 698, 694, 690, 685, 681, 677, 673, 669, 664, 660, 656, 655, 650, 646, 641, 636, 632, 627, 622, 618, 613, 612, 607, 602, 596, 591, 586, 580, 575, 570, 569, 563, 557, 551, 545, 539, 533, 527, 526, 519, 512, 505, 498, 491, 484, 483, 474, 466, 457, 449, 440, 439, 428, 418, 407, 396, 395, 381, 366, 352, 351, 330, 308, 307, 286, 264, 263, 242, 220, 219, 176, 175, 132, 131, 88, 44, 0, ] UpperCamelCase__ : str = [ 999, 991, 982, 974, 966, 958, 950, 941, 933, 925, 916, 908, 900, 899, 874, 850, 825, 800, 799, 700, 600, 500, 400, 300, 200, 100, 0, ] UpperCamelCase__ : int = [ 999, 992, 985, 978, 971, 964, 957, 949, 942, 935, 928, 921, 914, 907, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 300, 299, 200, 199, 100, 99, 0, ] UpperCamelCase__ : List[Any] = [ 999, 996, 992, 989, 985, 982, 979, 975, 972, 968, 965, 961, 958, 955, 951, 948, 944, 941, 938, 934, 931, 927, 924, 920, 917, 914, 910, 907, 903, 900, 899, 891, 884, 876, 869, 861, 853, 846, 838, 830, 823, 815, 808, 800, 799, 788, 777, 766, 755, 744, 733, 722, 711, 700, 699, 688, 677, 666, 655, 644, 633, 622, 611, 600, 599, 585, 571, 557, 542, 528, 514, 500, 499, 485, 471, 457, 442, 428, 414, 400, 399, 379, 359, 340, 320, 300, 299, 279, 259, 240, 220, 200, 199, 166, 133, 100, 99, 66, 33, 0, ]
164
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) _a : Dict = {'configuration_beit': ['BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BeitConfig', 'BeitOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Union[str, Any] = ['BeitFeatureExtractor'] _a : Dict = ['BeitImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Tuple = [ 'BEIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BeitForImageClassification', 'BeitForMaskedImageModeling', 'BeitForSemanticSegmentation', 'BeitModel', 'BeitPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[str] = [ 'FlaxBeitForImageClassification', 'FlaxBeitForMaskedImageModeling', 'FlaxBeitModel', 'FlaxBeitPreTrainedModel', ] if TYPE_CHECKING: from .configuration_beit import BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, BeitConfig, BeitOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_beit import BeitFeatureExtractor from .image_processing_beit import BeitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_beit import ( BEIT_PRETRAINED_MODEL_ARCHIVE_LIST, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, BeitPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_beit import ( FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel, FlaxBeitPreTrainedModel, ) else: import sys _a : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
44
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _snake_case : '''simple docstring''' def __init__( self: Tuple ,lowerCamelCase_: List[str] ,lowerCamelCase_: int=13 ,lowerCamelCase_: int=32 ,lowerCamelCase_: Optional[int]=2 ,lowerCamelCase_: Any=3 ,lowerCamelCase_: str=16 ,lowerCamelCase_: Optional[Any]=[1, 2, 1] ,lowerCamelCase_: Tuple=[2, 2, 4] ,lowerCamelCase_: int=2 ,lowerCamelCase_: List[Any]=2.0 ,lowerCamelCase_: str=True ,lowerCamelCase_: Optional[int]=0.0 ,lowerCamelCase_: List[Any]=0.0 ,lowerCamelCase_: List[str]=0.1 ,lowerCamelCase_: Tuple="gelu" ,lowerCamelCase_: Union[str, Any]=False ,lowerCamelCase_: Union[str, Any]=True ,lowerCamelCase_: Optional[int]=0.0_2 ,lowerCamelCase_: int=1e-5 ,lowerCamelCase_: Optional[int]=True ,lowerCamelCase_: Union[str, Any]=None ,lowerCamelCase_: Union[str, Any]=True ,lowerCamelCase_: Optional[int]=10 ,lowerCamelCase_: Tuple=8 ,) -> List[Any]: UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = batch_size UpperCAmelCase_ : int = image_size UpperCAmelCase_ : Union[str, Any] = patch_size UpperCAmelCase_ : Optional[Any] = num_channels UpperCAmelCase_ : int = embed_dim UpperCAmelCase_ : Union[str, Any] = depths UpperCAmelCase_ : List[str] = num_heads UpperCAmelCase_ : int = window_size UpperCAmelCase_ : List[str] = mlp_ratio UpperCAmelCase_ : Tuple = qkv_bias UpperCAmelCase_ : Tuple = hidden_dropout_prob UpperCAmelCase_ : str = attention_probs_dropout_prob UpperCAmelCase_ : Tuple = drop_path_rate UpperCAmelCase_ : List[str] = hidden_act UpperCAmelCase_ : int = use_absolute_embeddings UpperCAmelCase_ : Any = patch_norm UpperCAmelCase_ : Optional[int] = layer_norm_eps UpperCAmelCase_ : Tuple = initializer_range UpperCAmelCase_ : Optional[Any] = is_training UpperCAmelCase_ : Dict = scope UpperCAmelCase_ : int = use_labels UpperCAmelCase_ : Optional[Any] = type_sequence_label_size UpperCAmelCase_ : List[str] = encoder_stride def A__ ( self: Any ) -> int: UpperCAmelCase_ : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : List[Any] = None if self.use_labels: UpperCAmelCase_ : Optional[int] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase_ : str = self.get_config() return config, pixel_values, labels def A__ ( self: List[Any] ) -> Union[str, Any]: return SwinvaConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,depths=self.depths ,num_heads=self.num_heads ,window_size=self.window_size ,mlp_ratio=self.mlp_ratio ,qkv_bias=self.qkv_bias ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,drop_path_rate=self.drop_path_rate ,hidden_act=self.hidden_act ,use_absolute_embeddings=self.use_absolute_embeddings ,path_norm=self.patch_norm ,layer_norm_eps=self.layer_norm_eps ,initializer_range=self.initializer_range ,encoder_stride=self.encoder_stride ,) def A__ ( self: Dict ,lowerCamelCase_: Tuple ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: List[str] ) -> str: UpperCAmelCase_ : str = SwinvaModel(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Optional[Any] = model(lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) UpperCAmelCase_ : List[Any] = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, expected_seq_len, expected_dim) ) def A__ ( self: List[Any] ,lowerCamelCase_: List[Any] ,lowerCamelCase_: int ,lowerCamelCase_: int ) -> int: UpperCAmelCase_ : Any = SwinvaForMaskedImageModeling(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Union[str, Any] = model(lowerCamelCase_ ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase_ : str = 1 UpperCAmelCase_ : Optional[Any] = SwinvaForMaskedImageModeling(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ : int = model(lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, 1, self.image_size, self.image_size) ) def A__ ( self: int ,lowerCamelCase_: int ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Optional[Any] ) -> int: UpperCAmelCase_ : Union[str, Any] = self.type_sequence_label_size UpperCAmelCase_ : int = SwinvaForImageClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Optional[int] = model(lowerCamelCase_ ,labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def A__ ( self: str ) -> Union[str, Any]: UpperCAmelCase_ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = config_and_inputs UpperCAmelCase_ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _snake_case ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' A__ : Tuple = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) A__ : Optional[Any] = ( {"feature-extraction": SwinvaModel, "image-classification": SwinvaForImageClassification} if is_torch_available() else {} ) A__ : List[Any] = False A__ : Tuple = False A__ : int = False A__ : Union[str, Any] = False def A__ ( self: List[str] ) -> Optional[Any]: UpperCAmelCase_ : Any = SwinvaModelTester(self ) UpperCAmelCase_ : str = ConfigTester(self ,config_class=lowerCamelCase_ ,embed_dim=37 ) def A__ ( self: Optional[int] ) -> List[Any]: self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A__ ( self: Any ) -> Dict: UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) @unittest.skip(reason="""Got `CUDA error: misaligned address` with PyTorch 2.0.0.""" ) def A__ ( self: int ) -> Dict: pass @unittest.skip(reason="""Swinv2 does not use inputs_embeds""" ) def A__ ( self: Tuple ) -> List[str]: pass def A__ ( self: str ) -> List[Any]: UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : int = model_class(lowerCamelCase_ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) UpperCAmelCase_ : Tuple = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase_ ,nn.Linear ) ) def A__ ( self: Optional[Any] ) -> Optional[int]: UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Dict = model_class(lowerCamelCase_ ) UpperCAmelCase_ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : int = [*signature.parameters.keys()] UpperCAmelCase_ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,lowerCamelCase_ ) def A__ ( self: Union[str, Any] ) -> Optional[Any]: UpperCAmelCase_ , UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Any = True for model_class in self.all_model_classes: UpperCAmelCase_ : Optional[Any] = True UpperCAmelCase_ : Union[str, Any] = False UpperCAmelCase_ : str = True UpperCAmelCase_ : List[Any] = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Optional[int] = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) UpperCAmelCase_ : Optional[Any] = outputs.attentions UpperCAmelCase_ : List[str] = len(self.model_tester.depths ) self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCAmelCase_ : str = True UpperCAmelCase_ : Optional[Any] = config.window_size**2 UpperCAmelCase_ : Optional[int] = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Optional[Any] = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) UpperCAmelCase_ : List[Any] = outputs.attentions self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_heads[0], window_size_squared, window_size_squared] ,) UpperCAmelCase_ : Optional[Any] = len(lowerCamelCase_ ) # Check attention is always last and order is fine UpperCAmelCase_ : Tuple = True UpperCAmelCase_ : List[Any] = True UpperCAmelCase_ : Tuple = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Union[str, Any] = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) if hasattr(self.model_tester ,"""num_hidden_states_types""" ): UpperCAmelCase_ : List[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states UpperCAmelCase_ : List[str] = 2 self.assertEqual(out_len + added_hidden_states ,len(lowerCamelCase_ ) ) UpperCAmelCase_ : Any = outputs.attentions self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) ,[self.model_tester.num_heads[0], window_size_squared, window_size_squared] ,) def A__ ( self: List[str] ,lowerCamelCase_: Dict ,lowerCamelCase_: Tuple ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: Optional[int] ) -> List[Any]: UpperCAmelCase_ : str = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : int = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) UpperCAmelCase_ : List[str] = outputs.hidden_states UpperCAmelCase_ : Optional[Any] = getattr( self.model_tester ,"""expected_num_hidden_layers""" ,len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) # Swinv2 has a different seq_length UpperCAmelCase_ : Optional[Any] = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCAmelCase_ : int = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[num_patches, self.model_tester.embed_dim] ,) UpperCAmelCase_ : Optional[int] = outputs.reshaped_hidden_states self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = reshaped_hidden_states[0].shape UpperCAmelCase_ : Optional[Any] = ( reshaped_hidden_states[0].view(lowerCamelCase_ ,lowerCamelCase_ ,height * width ).permute(0 ,2 ,1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) ,[num_patches, self.model_tester.embed_dim] ,) def A__ ( self: Any ) -> int: UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Dict = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: UpperCAmelCase_ : Any = True self.check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ : str = True self.check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) def A__ ( self: List[str] ) -> Dict: UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Union[str, Any] = 3 UpperCAmelCase_ : Optional[int] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) UpperCAmelCase_ : List[str] = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCAmelCase_ : List[Any] = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) UpperCAmelCase_ : Optional[Any] = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: UpperCAmelCase_ : Optional[Any] = True self.check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,(padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ : List[str] = True self.check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,(padded_height, padded_width) ) def A__ ( self: Optional[int] ) -> str: UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase_ ) def A__ ( self: Union[str, Any] ) -> Dict: UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase_ ) @slow def A__ ( self: str ) -> Tuple: for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Dict = SwinvaModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) def A__ ( self: Any ) -> int: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : List[str] = _config_zero_init(lowerCamelCase_ ) for model_class in self.all_model_classes: UpperCAmelCase_ : int = model_class(config=lowerCamelCase_ ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() ,[0.0, 1.0] ,msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' ,) @require_vision @require_torch class _snake_case ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self: Dict ) -> Optional[Any]: return ( AutoImageProcessor.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ) if is_vision_available() else None ) @slow def A__ ( self: str ) -> List[Any]: UpperCAmelCase_ : Tuple = SwinvaForImageClassification.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ).to( lowerCamelCase_ ) UpperCAmelCase_ : Any = self.default_image_processor UpperCAmelCase_ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) UpperCAmelCase_ : Optional[int] = image_processor(images=lowerCamelCase_ ,return_tensors="""pt""" ).to(lowerCamelCase_ ) # forward pass with torch.no_grad(): UpperCAmelCase_ : Optional[Any] = model(**lowerCamelCase_ ) # verify the logits UpperCAmelCase_ : Dict = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape ,lowerCamelCase_ ) UpperCAmelCase_ : Any = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(lowerCamelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowerCamelCase_ ,atol=1e-4 ) )
345
0
"""simple docstring""" import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class _SCREAMING_SNAKE_CASE ( ctypes.Structure ): UpperCAmelCase_ :List[Any] = [("""size""", ctypes.c_int), ("""visible""", ctypes.c_byte)] def _snake_case ( ) -> str: '''simple docstring''' if os.name == "nt": lowerCAmelCase_ :Union[str, Any] = CursorInfo() lowerCAmelCase_ :Dict = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(SCREAMING_SNAKE_CASE_ , ctypes.byref(SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase_ :List[Any] = False ctypes.windll.kernelaa.SetConsoleCursorInfo(SCREAMING_SNAKE_CASE_ , ctypes.byref(SCREAMING_SNAKE_CASE_ ) ) elif os.name == "posix": sys.stdout.write("""\033[?25l""" ) sys.stdout.flush() def _snake_case ( ) -> Optional[int]: '''simple docstring''' if os.name == "nt": lowerCAmelCase_ :Dict = CursorInfo() lowerCAmelCase_ :Optional[Any] = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(SCREAMING_SNAKE_CASE_ , ctypes.byref(SCREAMING_SNAKE_CASE_ ) ) lowerCAmelCase_ :Tuple = True ctypes.windll.kernelaa.SetConsoleCursorInfo(SCREAMING_SNAKE_CASE_ , ctypes.byref(SCREAMING_SNAKE_CASE_ ) ) elif os.name == "posix": sys.stdout.write("""\033[?25h""" ) sys.stdout.flush() @contextmanager def _snake_case ( ) -> str: '''simple docstring''' try: hide_cursor() yield finally: show_cursor()
365
"""simple docstring""" # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class _SCREAMING_SNAKE_CASE ( A__ , A__ , A__ , unittest.TestCase ): UpperCAmelCase_ :int = StableDiffusionControlNetImgaImgPipeline UpperCAmelCase_ :str = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"height", "width"} UpperCAmelCase_ :Tuple = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCAmelCase_ :List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({"control_image"} ) UpperCAmelCase_ :Optional[Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def __lowerCAmelCase ( self ) -> List[str]: torch.manual_seed(0 ) lowerCAmelCase_ :Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) lowerCAmelCase_ :List[Any] = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[Any] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" , clip_sample=__A , set_alpha_to_one=__A , ) torch.manual_seed(0 ) lowerCAmelCase_ :List[str] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) lowerCAmelCase_ :Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) lowerCAmelCase_ :List[Any] = CLIPTextModel(__A ) lowerCAmelCase_ :int = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) lowerCAmelCase_ :Union[str, Any] = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def __lowerCAmelCase ( self , __A , __A=0 ) -> List[str]: if str(__A ).startswith("""mps""" ): lowerCAmelCase_ :Tuple = torch.manual_seed(__A ) else: lowerCAmelCase_ :Optional[int] = torch.Generator(device=__A ).manual_seed(__A ) lowerCAmelCase_ :List[Any] = 2 lowerCAmelCase_ :int = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__A , device=torch.device(__A ) , ) lowerCAmelCase_ :Optional[int] = floats_tensor(control_image.shape , rng=random.Random(__A ) ).to(__A ) lowerCAmelCase_ :Tuple = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase_ :List[Any] = Image.fromarray(np.uinta(__A ) ).convert("""RGB""" ).resize((64, 64) ) lowerCAmelCase_ :Union[str, Any] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def __lowerCAmelCase ( self ) -> int: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def __lowerCAmelCase ( self ) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def __lowerCAmelCase ( self ) -> List[str]: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) class _SCREAMING_SNAKE_CASE ( A__ , A__ , unittest.TestCase ): UpperCAmelCase_ :List[str] = StableDiffusionControlNetImgaImgPipeline UpperCAmelCase_ :int = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"height", "width"} UpperCAmelCase_ :str = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCAmelCase_ :int = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def __lowerCAmelCase ( self ) -> Optional[int]: torch.manual_seed(0 ) lowerCAmelCase_ :Dict = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(__A ): if isinstance(__A , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) lowerCAmelCase_ :List[Any] = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__A ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[Any] = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__A ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[Any] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" , clip_sample=__A , set_alpha_to_one=__A , ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) lowerCAmelCase_ :str = CLIPTextModel(__A ) lowerCAmelCase_ :str = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) lowerCAmelCase_ :Optional[Any] = MultiControlNetModel([controlneta, controlneta] ) lowerCAmelCase_ :List[Any] = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def __lowerCAmelCase ( self , __A , __A=0 ) -> str: if str(__A ).startswith("""mps""" ): lowerCAmelCase_ :Optional[Any] = torch.manual_seed(__A ) else: lowerCAmelCase_ :List[Any] = torch.Generator(device=__A ).manual_seed(__A ) lowerCAmelCase_ :Optional[Any] = 2 lowerCAmelCase_ :Optional[int] = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__A , device=torch.device(__A ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__A , device=torch.device(__A ) , ), ] lowerCAmelCase_ :int = floats_tensor(control_image[0].shape , rng=random.Random(__A ) ).to(__A ) lowerCAmelCase_ :Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase_ :List[Any] = Image.fromarray(np.uinta(__A ) ).convert("""RGB""" ).resize((64, 64) ) lowerCAmelCase_ :List[str] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def __lowerCAmelCase ( self ) -> Optional[Any]: lowerCAmelCase_ :List[str] = self.get_dummy_components() lowerCAmelCase_ :Tuple = self.pipeline_class(**__A ) pipe.to(__A ) lowerCAmelCase_ :Union[str, Any] = 1_0.0 lowerCAmelCase_ :Union[str, Any] = 4 lowerCAmelCase_ :Tuple = self.get_dummy_inputs(__A ) lowerCAmelCase_ :List[str] = steps lowerCAmelCase_ :int = scale lowerCAmelCase_ :Union[str, Any] = pipe(**__A )[0] lowerCAmelCase_ :Any = self.get_dummy_inputs(__A ) lowerCAmelCase_ :str = steps lowerCAmelCase_ :str = scale lowerCAmelCase_ :Tuple = pipe(**__A , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] lowerCAmelCase_ :Optional[Any] = self.get_dummy_inputs(__A ) lowerCAmelCase_ :Union[str, Any] = steps lowerCAmelCase_ :Union[str, Any] = scale lowerCAmelCase_ :str = pipe(**__A , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] lowerCAmelCase_ :List[str] = self.get_dummy_inputs(__A ) lowerCAmelCase_ :Optional[int] = steps lowerCAmelCase_ :Tuple = scale lowerCAmelCase_ :str = pipe(**__A , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 def __lowerCAmelCase ( self ) -> Dict: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def __lowerCAmelCase ( self ) -> Tuple: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def __lowerCAmelCase ( self ) -> Optional[int]: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) def __lowerCAmelCase ( self ) -> List[str]: lowerCAmelCase_ :str = self.get_dummy_components() lowerCAmelCase_ :Tuple = self.pipeline_class(**__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(__A ) except NotImplementedError: pass @slow @require_torch_gpu class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self ) -> str: lowerCAmelCase_ :Any = ControlNetModel.from_pretrained("""lllyasviel/sd-controlnet-canny""" ) lowerCAmelCase_ :int = StableDiffusionControlNetImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , safety_checker=__A , controlnet=__A ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__A ) lowerCAmelCase_ :List[str] = torch.Generator(device="""cpu""" ).manual_seed(0 ) lowerCAmelCase_ :List[Any] = """evil space-punk bird""" lowerCAmelCase_ :List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png""" ).resize((512, 512) ) lowerCAmelCase_ :int = load_image( """https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png""" ).resize((512, 512) ) lowerCAmelCase_ :Union[str, Any] = pipe( __A , __A , control_image=__A , generator=__A , output_type="""np""" , num_inference_steps=50 , strength=0.6 , ) lowerCAmelCase_ :Tuple = output.images[0] assert image.shape == (512, 512, 3) lowerCAmelCase_ :Tuple = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy""" ) assert np.abs(expected_image - image ).max() < 9E-2
1
0
"""simple docstring""" import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __lowerCamelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def lowerCamelCase ( self : str ): lowerCAmelCase_ : List[Any] = tempfile.mkdtemp() lowerCAmelCase_ : List[Any] = 8 # DPR tok lowerCAmelCase_ : Optional[int] = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] lowerCAmelCase_ : Dict = os.path.join(self.tmpdirname , "dpr_tokenizer" ) os.makedirs(a_ , exist_ok=a_ ) lowerCAmelCase_ : Dict = os.path.join(a_ , DPR_VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) # BART tok lowerCAmelCase_ : Union[str, Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] lowerCAmelCase_ : Dict = dict(zip(a_ , range(len(a_ ) ) ) ) lowerCAmelCase_ : Optional[Any] = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowerCAmelCase_ : List[Any] = {"unk_token": "<unk>"} lowerCAmelCase_ : Optional[Any] = os.path.join(self.tmpdirname , "bart_tokenizer" ) os.makedirs(a_ , exist_ok=a_ ) lowerCAmelCase_ : Tuple = os.path.join(a_ , BART_VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : Union[str, Any] = os.path.join(a_ , BART_VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(a_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(a_ ) ) def lowerCamelCase ( self : List[Any] ): return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , "dpr_tokenizer" ) ) def lowerCamelCase ( self : Any ): return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , "dpr_tokenizer" ) ) def lowerCamelCase ( self : int ): return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , "bart_tokenizer" ) ) def lowerCamelCase ( self : List[Any] ): shutil.rmtree(self.tmpdirname ) def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : Optional[int] = Dataset.from_dict( { "id": ["0", "1"], "text": ["foo", "bar"], "title": ["Foo", "Bar"], "embeddings": [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index("embeddings" , string_factory="Flat" , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def lowerCamelCase ( self : List[str] ): lowerCAmelCase_ : Any = self.get_dummy_dataset() lowerCAmelCase_ : Union[str, Any] = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch("transformers.models.rag.retrieval_rag.load_dataset" ) as mock_load_dataset: lowerCAmelCase_ : Any = dataset lowerCAmelCase_ : str = RagRetriever( a_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def lowerCamelCase ( self : Union[str, Any] , a_ : bool ): lowerCAmelCase_ : int = self.get_dummy_dataset() lowerCAmelCase_ : Optional[int] = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="custom" , ) if from_disk: lowerCAmelCase_ : Dict = os.path.join(self.tmpdirname , "dataset" ) lowerCAmelCase_ : Optional[int] = os.path.join(self.tmpdirname , "index.faiss" ) dataset.get_index("embeddings" ).save(os.path.join(self.tmpdirname , "index.faiss" ) ) dataset.drop_index("embeddings" ) dataset.save_to_disk(os.path.join(self.tmpdirname , "dataset" ) ) del dataset lowerCAmelCase_ : List[Any] = RagRetriever( a_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: lowerCAmelCase_ : str = RagRetriever( a_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , a_ ) , ) return retriever def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : Optional[Any] = Dataset.from_dict( { "id": ["0", "1"], "text": ["foo", "bar"], "title": ["Foo", "Bar"], "embeddings": [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index("embeddings" , string_factory="Flat" , metric_type=faiss.METRIC_INNER_PRODUCT ) lowerCAmelCase_ : Union[str, Any] = os.path.join(self.tmpdirname , "hf_bert_base.hnswSQ8_correct_phi_128.c_index" ) dataset.save_faiss_index("embeddings" , index_file_name + ".index.dpr" ) pickle.dump(dataset["id"] , open(index_file_name + ".index_meta.dpr" , "wb" ) ) lowerCAmelCase_ : List[Any] = os.path.join(self.tmpdirname , "psgs_w100.tsv.pkl" ) lowerCAmelCase_ : List[Any] = {sample["id"]: [sample["text"], sample["title"]] for sample in dataset} pickle.dump(a_ , open(a_ , "wb" ) ) lowerCAmelCase_ : Dict = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="legacy" , index_path=self.tmpdirname , ) lowerCAmelCase_ : List[str] = RagRetriever( a_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def lowerCamelCase ( self : Tuple ): lowerCAmelCase_ : Optional[Any] = 1 lowerCAmelCase_ : List[Any] = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase_ : Tuple = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = retriever.retrieve(a_ , n_docs=a_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(a_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["embeddings", "id", "text", "title"] ) self.assertEqual(len(doc_dicts[0]["id"] ) , a_ ) self.assertEqual(doc_dicts[0]["id"][0] , "1" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["id"][0] , "0" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self : Dict ): lowerCAmelCase_ : Optional[int] = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch("transformers.models.rag.retrieval_rag.load_dataset" ) as mock_load_dataset: lowerCAmelCase_ : Tuple = self.get_dummy_dataset() retriever.save_pretrained(a_ ) lowerCAmelCase_ : Union[str, Any] = RagRetriever.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) lowerCAmelCase_ : str = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ : Dict = retriever.retrieve(a_ , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self : Optional[Any] ): lowerCAmelCase_ : Any = 1 lowerCAmelCase_ : Union[str, Any] = self.get_dummy_custom_hf_index_retriever(from_disk=a_ ) lowerCAmelCase_ : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : int = retriever.retrieve(a_ , n_docs=a_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(a_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["embeddings", "id", "text", "title"] ) self.assertEqual(len(doc_dicts[0]["id"] ) , a_ ) self.assertEqual(doc_dicts[0]["id"][0] , "1" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["id"][0] , "0" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : Union[str, Any] = self.get_dummy_custom_hf_index_retriever(from_disk=a_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(a_ ) lowerCAmelCase_ : Optional[Any] = RagRetriever.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) lowerCAmelCase_ : Tuple = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ : int = retriever.retrieve(a_ , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self : Tuple ): lowerCAmelCase_ : List[str] = 1 lowerCAmelCase_ : str = self.get_dummy_custom_hf_index_retriever(from_disk=a_ ) lowerCAmelCase_ : Optional[int] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = retriever.retrieve(a_ , n_docs=a_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(a_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["embeddings", "id", "text", "title"] ) self.assertEqual(len(doc_dicts[0]["id"] ) , a_ ) self.assertEqual(doc_dicts[0]["id"][0] , "1" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["id"][0] , "0" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self : Any ): lowerCAmelCase_ : Optional[int] = self.get_dummy_custom_hf_index_retriever(from_disk=a_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(a_ ) lowerCAmelCase_ : Dict = RagRetriever.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) lowerCAmelCase_ : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ : Union[str, Any] = retriever.retrieve(a_ , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self : Dict ): lowerCAmelCase_ : Union[str, Any] = 1 lowerCAmelCase_ : str = self.get_dummy_legacy_index_retriever() lowerCAmelCase_ : int = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = retriever.retrieve(a_ , n_docs=a_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(a_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["text", "title"] ) self.assertEqual(len(doc_dicts[0]["text"] ) , a_ ) self.assertEqual(doc_dicts[0]["text"][0] , "bar" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["text"][0] , "foo" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self : Any ): lowerCAmelCase_ : Tuple = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(a_ ) lowerCAmelCase_ : List[Any] = RagRetriever.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) lowerCAmelCase_ : Dict = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ : List[Any] = retriever.retrieve(a_ , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self : Optional[Any] ): import torch lowerCAmelCase_ : Union[str, Any] = 1 lowerCAmelCase_ : Tuple = self.get_dummy_canonical_hf_index_retriever() lowerCAmelCase_ : List[Any] = [[5, 7], [10, 11]] lowerCAmelCase_ : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ : Tuple = retriever(a_ , a_ , prefix=retriever.config.generator.prefix , n_docs=a_ ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = ( out["context_input_ids"], out["context_attention_mask"], out["retrieved_doc_embeds"], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(a_ , a_ ) self.assertIsInstance(a_ , a_ ) self.assertIsInstance(a_ , np.ndarray ) lowerCAmelCase_ : int = retriever( a_ , a_ , prefix=retriever.config.generator.prefix , n_docs=a_ , return_tensors="pt" , ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = ( # noqa: F841 out["context_input_ids"], out["context_attention_mask"], out["retrieved_doc_embeds"], out["doc_ids"], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(a_ , torch.Tensor ) self.assertIsInstance(a_ , torch.Tensor ) self.assertIsInstance(a_ , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self : Optional[Any] ): lowerCAmelCase_ : str = self.get_dpr_ctx_encoder_tokenizer() lowerCAmelCase_ : Dict = 1 lowerCAmelCase_ : str = self.get_dummy_custom_hf_index_retriever(from_disk=a_ ) retriever.set_ctx_encoder_tokenizer(a_ ) lowerCAmelCase_ : List[str] = [[5, 7], [10, 11]] lowerCAmelCase_ : List[str] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) lowerCAmelCase_ : List[str] = retriever(a_ , a_ , prefix=retriever.config.generator.prefix , n_docs=a_ ) self.assertEqual( len(a_ ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ("tokenized_doc_ids", "tokenized_doc_attention_mask") ) , a_ ) # check for doc token related keys in dictionary.
241
def a__ ( __UpperCamelCase ): if not head: return True # split the list to two parts SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = head.next, head while fast and fast.next: SCREAMING_SNAKE_CASE_ = fast.next.next SCREAMING_SNAKE_CASE_ = slow.next SCREAMING_SNAKE_CASE_ = slow.next SCREAMING_SNAKE_CASE_ = None # Don't forget here! But forget still works! # reverse the second part SCREAMING_SNAKE_CASE_ = None while second: SCREAMING_SNAKE_CASE_ = second.next SCREAMING_SNAKE_CASE_ = node SCREAMING_SNAKE_CASE_ = second SCREAMING_SNAKE_CASE_ = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False SCREAMING_SNAKE_CASE_ = node.next SCREAMING_SNAKE_CASE_ = head.next return True def a__ ( __UpperCamelCase ): if not head or not head.next: return True # 1. Get the midpoint (slow) SCREAMING_SNAKE_CASE_ = SCREAMING_SNAKE_CASE_ = SCREAMING_SNAKE_CASE_ = head while fast and fast.next: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = fast.next.next, slow.next # 2. Push the second half into the stack SCREAMING_SNAKE_CASE_ = [slow.val] while slow.next: SCREAMING_SNAKE_CASE_ = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False SCREAMING_SNAKE_CASE_ = cur.next return True def a__ ( __UpperCamelCase ): if not head or not head.next: return True SCREAMING_SNAKE_CASE_ = {} SCREAMING_SNAKE_CASE_ = 0 while head: if head.val in d: d[head.val].append(__UpperCamelCase ) else: SCREAMING_SNAKE_CASE_ = [pos] SCREAMING_SNAKE_CASE_ = head.next pos += 1 SCREAMING_SNAKE_CASE_ = pos - 1 SCREAMING_SNAKE_CASE_ = 0 for v in d.values(): if len(__UpperCamelCase ) % 2 != 0: middle += 1 else: SCREAMING_SNAKE_CASE_ = 0 for i in range(0 , len(__UpperCamelCase ) ): if v[i] + v[len(__UpperCamelCase ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
118
0
"""simple docstring""" from __future__ import annotations from typing import Any class A_ ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" pass class A_ : """simple docstring""" def __init__( self :int , lowercase_ :Any ) -> None: UpperCAmelCase = data UpperCAmelCase = None def __iter__( self :int ) -> int: UpperCAmelCase = self UpperCAmelCase = [] while node: if node in visited: raise ContainsLoopError visited.append(lowercase_ ) yield node.data UpperCAmelCase = node.next_node @property def UpperCAmelCase__ ( self :Optional[int] ) -> bool: try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": snake_case_ = Node(1) snake_case_ = Node(2) snake_case_ = Node(3) snake_case_ = Node(4) print(root_node.has_loop) # False snake_case_ = root_node.next_node print(root_node.has_loop) # True snake_case_ = Node(5) snake_case_ = Node(6) snake_case_ = Node(5) snake_case_ = Node(6) print(root_node.has_loop) # False snake_case_ = Node(1) print(root_node.has_loop) # False
359
"""simple docstring""" from collections import deque class A_ : """simple docstring""" def __init__( self :Any , lowercase_ :str , lowercase_ :int , lowercase_ :int ) -> None: UpperCAmelCase = process_name # process name UpperCAmelCase = arrival_time # arrival time of the process # completion time of finished process or last interrupted time UpperCAmelCase = arrival_time UpperCAmelCase = burst_time # remaining burst time UpperCAmelCase = 0 # total time of the process wait in ready queue UpperCAmelCase = 0 # time from arrival time to completion time class A_ : """simple docstring""" def __init__( self :Any , lowercase_ :int , lowercase_ :list[int] , lowercase_ :deque[Process] , lowercase_ :int , ) -> None: # total number of mlfq's queues UpperCAmelCase = number_of_queues # time slice of queues that round robin algorithm applied UpperCAmelCase = time_slices # unfinished process is in this ready_queue UpperCAmelCase = queue # current time UpperCAmelCase = current_time # finished process is in this sequence queue UpperCAmelCase = deque() def UpperCAmelCase__ ( self :Optional[int] ) -> list[str]: UpperCAmelCase = [] for i in range(len(self.finish_queue ) ): sequence.append(self.finish_queue[i].process_name ) return sequence def UpperCAmelCase__ ( self :List[str] , lowercase_ :list[Process] ) -> list[int]: UpperCAmelCase = [] for i in range(len(lowercase_ ) ): waiting_times.append(queue[i].waiting_time ) return waiting_times def UpperCAmelCase__ ( self :List[str] , lowercase_ :list[Process] ) -> list[int]: UpperCAmelCase = [] for i in range(len(lowercase_ ) ): turnaround_times.append(queue[i].turnaround_time ) return turnaround_times def UpperCAmelCase__ ( self :Dict , lowercase_ :list[Process] ) -> list[int]: UpperCAmelCase = [] for i in range(len(lowercase_ ) ): completion_times.append(queue[i].stop_time ) return completion_times def UpperCAmelCase__ ( self :str , lowercase_ :deque[Process] ) -> list[int]: return [q.burst_time for q in queue] def UpperCAmelCase__ ( self :int , lowercase_ :Process ) -> int: process.waiting_time += self.current_time - process.stop_time return process.waiting_time def UpperCAmelCase__ ( self :Optional[int] , lowercase_ :deque[Process] ) -> deque[Process]: UpperCAmelCase = deque() # sequence deque of finished process while len(lowercase_ ) != 0: UpperCAmelCase = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of current process self.update_waiting_time(lowercase_ ) # update current time self.current_time += cp.burst_time # finish the process and set the process's burst-time 0 UpperCAmelCase = 0 # set the process's turnaround time because it is finished UpperCAmelCase = self.current_time - cp.arrival_time # set the completion time UpperCAmelCase = self.current_time # add the process to queue that has finished queue finished.append(lowercase_ ) self.finish_queue.extend(lowercase_ ) # add finished process to finish queue # FCFS will finish all remaining processes return finished def UpperCAmelCase__ ( self :Tuple , lowercase_ :deque[Process] , lowercase_ :int ) -> tuple[deque[Process], deque[Process]]: UpperCAmelCase = deque() # sequence deque of terminated process # just for 1 cycle and unfinished processes will go back to queue for _ in range(len(lowercase_ ) ): UpperCAmelCase = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of unfinished processes self.update_waiting_time(lowercase_ ) # if the burst time of process is bigger than time-slice if cp.burst_time > time_slice: # use CPU for only time-slice self.current_time += time_slice # update remaining burst time cp.burst_time -= time_slice # update end point time UpperCAmelCase = self.current_time # locate the process behind the queue because it is not finished ready_queue.append(lowercase_ ) else: # use CPU for remaining burst time self.current_time += cp.burst_time # set burst time 0 because the process is finished UpperCAmelCase = 0 # set the finish time UpperCAmelCase = self.current_time # update the process' turnaround time because it is finished UpperCAmelCase = self.current_time - cp.arrival_time # add the process to queue that has finished queue finished.append(lowercase_ ) self.finish_queue.extend(lowercase_ ) # add finished process to finish queue # return finished processes queue and remaining processes queue return finished, ready_queue def UpperCAmelCase__ ( self :Optional[Any] ) -> deque[Process]: # all queues except last one have round_robin algorithm for i in range(self.number_of_queues - 1 ): UpperCAmelCase , UpperCAmelCase = self.round_robin( self.ready_queue , self.time_slices[i] ) # the last queue has first_come_first_served algorithm self.first_come_first_served(self.ready_queue ) return self.finish_queue if __name__ == "__main__": import doctest snake_case_ = Process("""P1""", 0, 53) snake_case_ = Process("""P2""", 0, 17) snake_case_ = Process("""P3""", 0, 68) snake_case_ = Process("""P4""", 0, 24) snake_case_ = 3 snake_case_ = [17, 25] snake_case_ = deque([Pa, Pa, Pa, Pa]) if len(time_slices) != number_of_queues - 1: raise SystemExit(0) doctest.testmod(extraglobs={"""queue""": deque([Pa, Pa, Pa, Pa])}) snake_case_ = Process("""P1""", 0, 53) snake_case_ = Process("""P2""", 0, 17) snake_case_ = Process("""P3""", 0, 68) snake_case_ = Process("""P4""", 0, 24) snake_case_ = 3 snake_case_ = [17, 25] snake_case_ = deque([Pa, Pa, Pa, Pa]) snake_case_ = MLFQ(number_of_queues, time_slices, queue, 0) snake_case_ = mlfq.multi_level_feedback_queue() # print total waiting times of processes(P1, P2, P3, P4) print( f'''waiting time:\ \t\t\t{MLFQ.calculate_waiting_time(mlfq, [Pa, Pa, Pa, Pa])}''' ) # print completion times of processes(P1, P2, P3, P4) print( f'''completion time:\ \t\t{MLFQ.calculate_completion_time(mlfq, [Pa, Pa, Pa, Pa])}''' ) # print total turnaround times of processes(P1, P2, P3, P4) print( f'''turnaround time:\ \t\t{MLFQ.calculate_turnaround_time(mlfq, [Pa, Pa, Pa, Pa])}''' ) # print sequence of finished processes print( f'''sequence of finished processes:\ {mlfq.calculate_sequence_of_finish_queue()}''' )
181
0
"""simple docstring""" import string from math import logaa def lowerCAmelCase__ ( _UpperCamelCase : str , _UpperCamelCase : str ) -> int: """simple docstring""" snake_case = document.translate( str.maketrans('' , '' , string.punctuation ) ).replace('\n' , '' ) snake_case = document_without_punctuation.split(' ' ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def lowerCAmelCase__ ( _UpperCamelCase : str , _UpperCamelCase : str ) -> tuple[int, int]: """simple docstring""" snake_case = corpus.lower().translate( str.maketrans('' , '' , string.punctuation ) ) # strip all punctuation and replace it with '' snake_case = corpus_without_punctuation.split('\n' ) snake_case = term.lower() return (len([doc for doc in docs if term in doc] ), len(_UpperCamelCase )) def lowerCAmelCase__ ( _UpperCamelCase : int , _UpperCamelCase : int , _UpperCamelCase : List[Any]=False ) -> float: """simple docstring""" if smoothing: if n == 0: raise ValueError('log10(0) is undefined.' ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError('df must be > 0' ) elif n == 0: raise ValueError('log10(0) is undefined.' ) return round(logaa(n / df ) , 3 ) def lowerCAmelCase__ ( _UpperCamelCase : int , _UpperCamelCase : int ) -> float: """simple docstring""" return round(tf * idf , 3 )
150
"""simple docstring""" import itertools import string from collections.abc import Generator, Iterable def lowerCAmelCase__ ( _UpperCamelCase : Iterable[str] , _UpperCamelCase : int ) -> Generator[tuple[str, ...], None, None]: """simple docstring""" snake_case = iter(_UpperCamelCase ) while True: snake_case = tuple(itertools.islice(_UpperCamelCase , _UpperCamelCase ) ) if not chunk: return yield chunk def lowerCAmelCase__ ( _UpperCamelCase : str ) -> str: """simple docstring""" snake_case = ''.join([c.upper() for c in dirty if c in string.ascii_letters] ) snake_case = '' if len(_UpperCamelCase ) < 2: return dirty for i in range(len(_UpperCamelCase ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(_UpperCamelCase ) & 1: clean += "X" return clean def lowerCAmelCase__ ( _UpperCamelCase : str ) -> list[str]: """simple docstring""" snake_case = 'ABCDEFGHIKLMNOPQRSTUVWXYZ' # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler snake_case = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(_UpperCamelCase ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(_UpperCamelCase ) return table def lowerCAmelCase__ ( _UpperCamelCase : str , _UpperCamelCase : str ) -> str: """simple docstring""" snake_case = generate_table(_UpperCamelCase ) snake_case = prepare_input(_UpperCamelCase ) snake_case = '' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(_UpperCamelCase , 2 ): snake_case ,snake_case = divmod(table.index(_UpperCamelCase ) , 5 ) snake_case ,snake_case = divmod(table.index(_UpperCamelCase ) , 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def lowerCAmelCase__ ( _UpperCamelCase : str , _UpperCamelCase : str ) -> str: """simple docstring""" snake_case = generate_table(_UpperCamelCase ) snake_case = '' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(_UpperCamelCase , 2 ): snake_case ,snake_case = divmod(table.index(_UpperCamelCase ) , 5 ) snake_case ,snake_case = divmod(table.index(_UpperCamelCase ) , 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
150
1
'''simple docstring''' from typing import Any class UpperCAmelCase_ : def __init__( self : Any , UpperCAmelCase__ : Any ) -> Optional[int]: lowerCAmelCase = data lowerCAmelCase = None class UpperCAmelCase_ : def __init__( self : Union[str, Any] ) -> str: lowerCAmelCase = None def __UpperCAmelCase ( self : str ) -> List[Any]: lowerCAmelCase = self.head while temp is not None: print(temp.data , end=' ' ) lowerCAmelCase = temp.next print() def __UpperCAmelCase ( self : List[Any] , UpperCAmelCase__ : Any ) -> Optional[Any]: lowerCAmelCase = Node(UpperCAmelCase__ ) lowerCAmelCase = self.head lowerCAmelCase = new_node def __UpperCAmelCase ( self : int , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Union[str, Any] ) -> List[str]: if node_data_a == node_data_a: return else: lowerCAmelCase = self.head while node_a is not None and node_a.data != node_data_a: lowerCAmelCase = node_a.next lowerCAmelCase = self.head while node_a is not None and node_a.data != node_data_a: lowerCAmelCase = node_a.next if node_a is None or node_a is None: return lowerCAmelCase , lowerCAmelCase = node_a.data, node_a.data if __name__ == "__main__": __snake_case =LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print("""After swapping""") ll.print_list()
352
'''simple docstring''' from __future__ import annotations from scipy.special import comb # type: ignore class UpperCAmelCase_ : def __init__( self : Dict , UpperCAmelCase__ : list[tuple[float, float]] ) -> str: lowerCAmelCase = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. lowerCAmelCase = len(UpperCAmelCase__ ) - 1 def __UpperCAmelCase ( self : str , UpperCAmelCase__ : float ) -> list[float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." lowerCAmelCase = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , UpperCAmelCase__ ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(UpperCAmelCase__ ) , 5 ) == 1 return output_values def __UpperCAmelCase ( self : Union[str, Any] , UpperCAmelCase__ : float ) -> tuple[float, float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." lowerCAmelCase = self.basis_function(UpperCAmelCase__ ) lowerCAmelCase = 0.0 lowerCAmelCase = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : float = 0.01 ) -> Optional[int]: from matplotlib import pyplot as plt # type: ignore lowerCAmelCase = [] # x coordinates of points to plot lowerCAmelCase = [] # y coordinates of points to plot lowerCAmelCase = 0.0 while t <= 1: lowerCAmelCase = self.bezier_curve_function(UpperCAmelCase__ ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size lowerCAmelCase = [i[0] for i in self.list_of_points] lowerCAmelCase = [i[1] for i in self.list_of_points] plt.plot( UpperCAmelCase__ , UpperCAmelCase__ , color='blue' , label='Curve of Degree ' + str(self.degree ) , ) plt.scatter(UpperCAmelCase__ , UpperCAmelCase__ , color='red' , label='Control Points' ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
55
0
"""simple docstring""" import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all BART models at https://huggingface.co/models?filter=bart lowerCAmelCase__ = { '''vocab_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/vocab.json''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/vocab.json''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json''', }, '''merges_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/merges.txt''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/merges.txt''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt''', }, } lowerCAmelCase__ = { '''facebook/bart-base''': 1_024, '''facebook/bart-large''': 1_024, '''facebook/bart-large-mnli''': 1_024, '''facebook/bart-large-cnn''': 1_024, '''facebook/bart-large-xsum''': 1_024, '''yjernite/bart_eli5''': 1_024, } @lru_cache() def a__ ( ): '''simple docstring''' lowerCAmelCase : Optional[Any] = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) lowerCAmelCase : int = bs[:] lowerCAmelCase : List[str] = 0 for b in range(2**8 ): if b not in bs: bs.append(SCREAMING_SNAKE_CASE ) cs.append(2**8 + n ) n += 1 lowerCAmelCase : Optional[int] = [chr(SCREAMING_SNAKE_CASE ) for n in cs] return dict(zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) def a__ ( SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' lowerCAmelCase : Any = set() lowerCAmelCase : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCAmelCase : str = char return pairs class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" a : Any =VOCAB_FILES_NAMES a : List[Any] =PRETRAINED_VOCAB_FILES_MAP a : Optional[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a : Dict =["input_ids", "attention_mask"] def __init__( self , snake_case__ , snake_case__ , snake_case__="replace" , snake_case__="<s>" , snake_case__="</s>" , snake_case__="</s>" , snake_case__="<s>" , snake_case__="<unk>" , snake_case__="<pad>" , snake_case__="<mask>" , snake_case__=False , **snake_case__ , ): """simple docstring""" lowerCAmelCase : List[str] = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else bos_token lowerCAmelCase : Optional[int] = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else eos_token lowerCAmelCase : Tuple = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else sep_token lowerCAmelCase : List[Any] = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else cls_token lowerCAmelCase : Any = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else unk_token lowerCAmelCase : Any = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase : List[str] = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token super().__init__( errors=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , cls_token=snake_case__ , pad_token=snake_case__ , mask_token=snake_case__ , add_prefix_space=snake_case__ , **snake_case__ , ) with open(snake_case__ , encoding="utf-8" ) as vocab_handle: lowerCAmelCase : Union[str, Any] = json.load(snake_case__ ) lowerCAmelCase : Union[str, Any] = {v: k for k, v in self.encoder.items()} lowerCAmelCase : Union[str, Any] = errors # how to handle errors in decoding lowerCAmelCase : int = bytes_to_unicode() lowerCAmelCase : List[Any] = {v: k for k, v in self.byte_encoder.items()} with open(snake_case__ , encoding="utf-8" ) as merges_handle: lowerCAmelCase : Dict = merges_handle.read().split("\n" )[1:-1] lowerCAmelCase : Tuple = [tuple(merge.split() ) for merge in bpe_merges] lowerCAmelCase : Union[str, Any] = dict(zip(snake_case__ , range(len(snake_case__ ) ) ) ) lowerCAmelCase : str = {} lowerCAmelCase : Union[str, Any] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCAmelCase : List[str] = re.compile(r"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property def lowercase__ ( self ): """simple docstring""" return len(self.encoder ) def lowercase__ ( self ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowercase__ ( self , snake_case__ ): """simple docstring""" if token in self.cache: return self.cache[token] lowerCAmelCase : Tuple = tuple(snake_case__ ) lowerCAmelCase : Tuple = get_pairs(snake_case__ ) if not pairs: return token while True: lowerCAmelCase : Optional[Any] = min(snake_case__ , key=lambda snake_case__ : self.bpe_ranks.get(snake_case__ , float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase , lowerCAmelCase : Tuple = bigram lowerCAmelCase : Optional[int] = [] lowerCAmelCase : str = 0 while i < len(snake_case__ ): try: lowerCAmelCase : Dict = word.index(snake_case__ , snake_case__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase : List[str] = j if word[i] == first and i < len(snake_case__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase : Optional[Any] = tuple(snake_case__ ) lowerCAmelCase : Any = new_word if len(snake_case__ ) == 1: break else: lowerCAmelCase : List[str] = get_pairs(snake_case__ ) lowerCAmelCase : List[str] = " ".join(snake_case__ ) lowerCAmelCase : Tuple = word return word def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : List[str] = [] for token in re.findall(self.pat , snake_case__ ): lowerCAmelCase : Optional[Any] = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(snake_case__ ).split(" " ) ) return bpe_tokens def lowercase__ ( self , snake_case__ ): """simple docstring""" return self.encoder.get(snake_case__ , self.encoder.get(self.unk_token ) ) def lowercase__ ( self , snake_case__ ): """simple docstring""" return self.decoder.get(snake_case__ ) def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : List[str] = "".join(snake_case__ ) lowerCAmelCase : Tuple = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def lowercase__ ( self , snake_case__ , snake_case__ = None ): """simple docstring""" if not os.path.isdir(snake_case__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase : Optional[int] = os.path.join( snake_case__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase : Tuple = os.path.join( snake_case__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(snake_case__ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=snake_case__ , ensure_ascii=snake_case__ ) + "\n" ) lowerCAmelCase : int = 0 with open(snake_case__ , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda snake_case__ : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) lowerCAmelCase : int = token_index writer.write(" ".join(snake_case__ ) + "\n" ) index += 1 return vocab_file, merge_file def lowercase__ ( self , snake_case__ , snake_case__ = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase : Union[str, Any] = [self.cls_token_id] lowerCAmelCase : Optional[int] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowercase__ ( self , snake_case__ , snake_case__ = None , snake_case__ = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case__ , token_ids_a=snake_case__ , already_has_special_tokens=snake_case__ ) if token_ids_a is None: return [1] + ([0] * len(snake_case__ )) + [1] return [1] + ([0] * len(snake_case__ )) + [1, 1] + ([0] * len(snake_case__ )) + [1] def lowercase__ ( self , snake_case__ , snake_case__ = None ): """simple docstring""" lowerCAmelCase : Union[str, Any] = [self.sep_token_id] lowerCAmelCase : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase__ ( self , snake_case__ , snake_case__=False , **snake_case__ ): """simple docstring""" lowerCAmelCase : Union[str, Any] = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(snake_case__ ) > 0 and not text[0].isspace()): lowerCAmelCase : List[Any] = " " + text return (text, kwargs)
108
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) A__ : List[Any] =pytest.mark.integration @pytest.mark.parametrize("""path""" , ["""paws""", """csv"""] ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" inspect_dataset(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = path + """.py""" assert script_name in os.listdir(lowerCAmelCase ) assert "__pycache__" not in os.listdir(lowerCAmelCase ) @pytest.mark.filterwarnings("""ignore:inspect_metric is deprecated:FutureWarning""" ) @pytest.mark.filterwarnings("""ignore:metric_module_factory is deprecated:FutureWarning""" ) @pytest.mark.parametrize("""path""" , ["""accuracy"""] ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" inspect_metric(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = path + """.py""" assert script_name in os.listdir(lowerCAmelCase ) assert "__pycache__" not in os.listdir(lowerCAmelCase ) @pytest.mark.parametrize( """path, config_name, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_config_info(lowerCAmelCase , config_name=lowerCAmelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" with pytest.raises(lowerCAmelCase ): get_dataset_config_info(lowerCAmelCase , config_name=lowerCAmelCase ) @pytest.mark.parametrize( """path, expected""" , [ ("""squad""", """plain_text"""), ("""acronym_identification""", """default"""), ("""lhoestq/squad""", """plain_text"""), ("""lhoestq/test""", """default"""), ("""lhoestq/demo1""", """lhoestq--demo1"""), ("""dalle-mini/wit""", """dalle-mini--wit"""), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_config_names(lowerCAmelCase ) assert expected in config_names @pytest.mark.parametrize( """path, expected_configs, expected_splits_in_first_config""" , [ ("""squad""", ["""plain_text"""], ["""train""", """validation"""]), ("""dalle-mini/wit""", ["""dalle-mini--wit"""], ["""train"""]), ("""paws""", ["""labeled_final""", """labeled_swap""", """unlabeled_final"""], ["""train""", """test""", """validation"""]), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_infos(lowerCAmelCase ) assert list(infos.keys() ) == expected_configs _lowerCAmelCase = expected_configs[0] assert expected_config in infos _lowerCAmelCase = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( """path, expected_config, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_infos(lowerCAmelCase ) assert expected_config in infos _lowerCAmelCase = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" with pytest.raises(lowerCAmelCase ): get_dataset_split_names(lowerCAmelCase , config_name=lowerCAmelCase )
70
0
import json import logging import os import socket import git import numpy as np import torch logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - PID: %(process)d - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO, ) lowerCAmelCase_ : str = logging.getLogger(__name__) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = git.Repo(search_parent_directories=a__ ) UpperCAmelCase = { """repo_id""": str(a__ ), """repo_sha""": str(repo.head.object.hexsha ), """repo_branch""": str(repo.active_branch ), } with open(os.path.join(a__ , """git_log.json""" ) , """w""" ) as f: json.dump(a__ , a__ , indent=4 ) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' if params.n_gpu <= 0: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = True UpperCAmelCase = False return assert torch.cuda.is_available() logger.info("""Initializing GPUs""" ) if params.n_gpu > 1: assert params.local_rank != -1 UpperCAmelCase = int(os.environ["""WORLD_SIZE"""] ) UpperCAmelCase = int(os.environ["""N_GPU_NODE"""] ) UpperCAmelCase = int(os.environ["""RANK"""] ) # number of nodes / node ID UpperCAmelCase = params.world_size // params.n_gpu_per_node UpperCAmelCase = params.global_rank // params.n_gpu_per_node UpperCAmelCase = True assert params.n_nodes == int(os.environ["""N_NODES"""] ) assert params.node_id == int(os.environ["""NODE_RANK"""] ) # local job (single GPU) else: assert params.local_rank == -1 UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 1 UpperCAmelCase = False # sanity checks assert params.n_nodes >= 1 assert 0 <= params.node_id < params.n_nodes assert 0 <= params.local_rank <= params.global_rank < params.world_size assert params.world_size == params.n_nodes * params.n_gpu_per_node # define whether this is the master process / if we are in multi-node distributed mode UpperCAmelCase = params.node_id == 0 and params.local_rank == 0 UpperCAmelCase = params.n_nodes > 1 # summary UpperCAmelCase = F'''--- Global rank: {params.global_rank} - ''' logger.info(PREFIX + """Number of nodes: %i""" % params.n_nodes ) logger.info(PREFIX + """Node ID : %i""" % params.node_id ) logger.info(PREFIX + """Local rank : %i""" % params.local_rank ) logger.info(PREFIX + """World size : %i""" % params.world_size ) logger.info(PREFIX + """GPUs per node : %i""" % params.n_gpu_per_node ) logger.info(PREFIX + """Master : %s""" % str(params.is_master ) ) logger.info(PREFIX + """Multi-node : %s""" % str(params.multi_node ) ) logger.info(PREFIX + """Multi-GPU : %s""" % str(params.multi_gpu ) ) logger.info(PREFIX + """Hostname : %s""" % socket.gethostname() ) # set GPU device torch.cuda.set_device(params.local_rank ) # initialize multi-GPU if params.multi_gpu: logger.info("""Initializing PyTorch distributed""" ) torch.distributed.init_process_group( init_method="""env://""" , backend="""nccl""" , ) def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' np.random.seed(args.seed ) torch.manual_seed(args.seed ) if args.n_gpu > 0: torch.cuda.manual_seed_all(args.seed )
371
"""simple docstring""" import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' return (data["data"], data["target"]) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = XGBRegressor(verbosity=0 , random_state=42 ) xgb.fit(lowerCAmelCase , lowerCAmelCase ) # Predict target for test data UpperCAmelCase = xgb.predict(lowerCAmelCase ) UpperCAmelCase = predictions.reshape(len(lowerCAmelCase ) , 1 ) return predictions def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = fetch_california_housing() UpperCAmelCase , UpperCAmelCase = data_handling(lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_test_split( lowerCAmelCase , lowerCAmelCase , test_size=0.25 , random_state=1 ) UpperCAmelCase = xgboost(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # Error printing print(F'''Mean Absolute Error : {mean_absolute_error(lowerCAmelCase , lowerCAmelCase )}''' ) print(F'''Mean Square Error : {mean_squared_error(lowerCAmelCase , lowerCAmelCase )}''' ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
248
0
'''simple docstring''' import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py a : Optional[Any] = "src/transformers" a : int = "docs/source/en/tasks" def lowercase ( __magic_name__ , __magic_name__ , __magic_name__ ): '''simple docstring''' with open(__magic_name__ , "r" , encoding="utf-8" , newline="\n" ) as f: UpperCAmelCase : Any = f.readlines() # Find the start prompt. UpperCAmelCase : Any = 0 while not lines[start_index].startswith(__magic_name__ ): start_index += 1 start_index += 1 UpperCAmelCase : Dict = start_index while not lines[end_index].startswith(__magic_name__ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. a : Any = direct_transformers_import(TRANSFORMERS_PATH) a : Tuple = { "asr.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, "audio_classification.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, "language_modeling.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, "image_classification.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, "masked_language_modeling.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, "multiple_choice.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, "object_detection.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, "question_answering.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, "semantic_segmentation.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, "sequence_classification.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, "summarization.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, "token_classification.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, "translation.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, "video_classification.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, "document_question_answering.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, "monocular_depth_estimation.md": transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). a : Tuple = { "summarization.md": ("nllb",), "translation.md": ("nllb",), } def lowercase ( __magic_name__ ): '''simple docstring''' UpperCAmelCase : List[Any] = TASK_GUIDE_TO_MODELS[task_guide] UpperCAmelCase : int = SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(__magic_name__ , set() ) UpperCAmelCase : Optional[Any] = { code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F"[{name}](../model_doc/{code})" for code, name in model_names.items()] ) + "\n" def lowercase ( __magic_name__ , __magic_name__=False ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Dict = _find_text_in_file( filename=os.path.join(__magic_name__ , __magic_name__ ) , start_prompt="<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->" , end_prompt="<!--End of the generated tip-->" , ) UpperCAmelCase : Tuple = get_model_list_for_task(__magic_name__ ) if current_list != new_list: if overwrite: with open(os.path.join(__magic_name__ , __magic_name__ ) , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F"The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`" " to fix this." ) if __name__ == "__main__": a : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") a : List[Any] = parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
311
'''simple docstring''' from collections.abc import Generator from math import sin def lowercase ( __magic_name__ ): '''simple docstring''' if len(__magic_name__ ) != 32: raise ValueError("Input must be of length 32" ) UpperCAmelCase : Union[str, Any] = b"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def lowercase ( __magic_name__ ): '''simple docstring''' if i < 0: raise ValueError("Input must be non-negative" ) UpperCAmelCase : Dict = format(__magic_name__ , "08x" )[-8:] UpperCAmelCase : List[str] = b"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8" ) return little_endian_hex def lowercase ( __magic_name__ ): '''simple docstring''' UpperCAmelCase : int = b"" for char in message: bit_string += format(__magic_name__ , "08b" ).encode("utf-8" ) UpperCAmelCase : List[Any] = format(len(__magic_name__ ) , "064b" ).encode("utf-8" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(__magic_name__ ) % 512 != 448: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def lowercase ( __magic_name__ ): '''simple docstring''' if len(__magic_name__ ) % 512 != 0: raise ValueError("Input must have length that's a multiple of 512" ) for pos in range(0 , len(__magic_name__ ) , 512 ): UpperCAmelCase : Union[str, Any] = bit_string[pos : pos + 512] UpperCAmelCase : Tuple = [] for i in range(0 , 512 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def lowercase ( __magic_name__ ): '''simple docstring''' if i < 0: raise ValueError("Input must be non-negative" ) UpperCAmelCase : Any = format(__magic_name__ , "032b" ) UpperCAmelCase : int = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(__magic_name__ , 2 ) def lowercase ( __magic_name__ , __magic_name__ ): '''simple docstring''' return (a + b) % 2**32 def lowercase ( __magic_name__ , __magic_name__ ): '''simple docstring''' if i < 0: raise ValueError("Input must be non-negative" ) if shift < 0: raise ValueError("Shift must be non-negative" ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def lowercase ( __magic_name__ ): '''simple docstring''' UpperCAmelCase : Dict = preprocess(__magic_name__ ) UpperCAmelCase : List[Any] = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states UpperCAmelCase : List[str] = 0X67452301 UpperCAmelCase : Tuple = 0XEFCDAB89 UpperCAmelCase : List[Any] = 0X98BADCFE UpperCAmelCase : List[str] = 0X10325476 UpperCAmelCase : Dict = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(__magic_name__ ): UpperCAmelCase : Optional[Any] = aa UpperCAmelCase : List[Any] = ba UpperCAmelCase : Optional[Any] = ca UpperCAmelCase : Any = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f UpperCAmelCase : Tuple = d ^ (b & (c ^ d)) UpperCAmelCase : List[str] = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f UpperCAmelCase : int = c ^ (d & (b ^ c)) UpperCAmelCase : Tuple = (5 * i + 1) % 16 elif i <= 47: UpperCAmelCase : Any = b ^ c ^ d UpperCAmelCase : Union[str, Any] = (3 * i + 5) % 16 else: UpperCAmelCase : Dict = c ^ (b | not_aa(__magic_name__ )) UpperCAmelCase : Dict = (7 * i) % 16 UpperCAmelCase : List[str] = (f + a + added_consts[i] + block_words[g]) % 2**32 UpperCAmelCase : List[Any] = d UpperCAmelCase : Any = c UpperCAmelCase : Dict = b UpperCAmelCase : Union[str, Any] = sum_aa(__magic_name__ , left_rotate_aa(__magic_name__ , shift_amounts[i] ) ) # Add hashed chunk to running total UpperCAmelCase : List[str] = sum_aa(__magic_name__ , __magic_name__ ) UpperCAmelCase : Any = sum_aa(__magic_name__ , __magic_name__ ) UpperCAmelCase : List[Any] = sum_aa(__magic_name__ , __magic_name__ ) UpperCAmelCase : Optional[int] = sum_aa(__magic_name__ , __magic_name__ ) UpperCAmelCase : List[str] = reformat_hex(__magic_name__ ) + reformat_hex(__magic_name__ ) + reformat_hex(__magic_name__ ) + reformat_hex(__magic_name__ ) return digest if __name__ == "__main__": import doctest doctest.testmod()
311
1
def lowerCAmelCase__ ( a__: Optional[Any] ) -> str: '''simple docstring''' _UpperCAmelCase = len(a__ ) for i in range(length - 1 ): _UpperCAmelCase = i for k in range(i + 1 , a__ ): if collection[k] < collection[least]: _UpperCAmelCase = k if least != i: _UpperCAmelCase , _UpperCAmelCase = (collection[i], collection[least]) return collection if __name__ == "__main__": lowerCAmelCase__ :Tuple = input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase__ :Dict = [int(item) for item in user_input.split(''',''')] print(selection_sort(unsorted))
360
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class __a : _a : Dict = BlenderbotConfig _a : Dict = {} _a : Union[str, Any] = 'gelu' def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=99 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=37 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=20 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=0 , ) -> Optional[int]: """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = seq_length _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = vocab_size _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = eos_token_id _UpperCAmelCase = pad_token_id _UpperCAmelCase = bos_token_id def UpperCAmelCase__ ( self ) -> Tuple: """simple docstring""" _UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) _UpperCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) _UpperCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) _UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) _UpperCAmelCase = prepare_blenderbot_inputs_dict(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return config, inputs_dict def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" _UpperCAmelCase = TFBlenderbotModel(config=_SCREAMING_SNAKE_CASE ).get_decoder() _UpperCAmelCase = inputs_dict['input_ids'] _UpperCAmelCase = input_ids[:1, :] _UpperCAmelCase = inputs_dict['attention_mask'][:1, :] _UpperCAmelCase = inputs_dict['head_mask'] _UpperCAmelCase = 1 # first forward pass _UpperCAmelCase = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , head_mask=_SCREAMING_SNAKE_CASE , use_cache=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase , _UpperCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _UpperCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) _UpperCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and _UpperCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) _UpperCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) _UpperCAmelCase = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE )[0] _UpperCAmelCase = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , past_key_values=_SCREAMING_SNAKE_CASE )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice _UpperCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) _UpperCAmelCase = output_from_no_past[:, -3:, random_slice_idx] _UpperCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , rtol=1e-3 ) def lowerCAmelCase__ ( a__: Dict , a__: Dict , a__: Any , a__: Any=None , a__: List[Any]=None , a__: Union[str, Any]=None , a__: Tuple=None , a__: Union[str, Any]=None , ) -> Any: '''simple docstring''' if attention_mask is None: _UpperCAmelCase = tf.cast(tf.math.not_equal(a__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: _UpperCAmelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: _UpperCAmelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _UpperCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _UpperCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __a ( UpperCAmelCase , UpperCAmelCase , unittest.TestCase ): _a : List[Any] = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () _a : List[str] = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () _a : List[str] = ( { 'conversational': TFBlenderbotForConditionalGeneration, 'feature-extraction': TFBlenderbotModel, 'summarization': TFBlenderbotForConditionalGeneration, 'text2text-generation': TFBlenderbotForConditionalGeneration, 'translation': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) _a : Dict = True _a : int = False _a : Union[str, Any] = False def UpperCAmelCase__ ( self ) -> List[Any]: """simple docstring""" _UpperCAmelCase = TFBlenderbotModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self ) -> str: """simple docstring""" self.config_tester.run_common_tests() def UpperCAmelCase__ ( self ) -> List[str]: """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_SCREAMING_SNAKE_CASE ) @require_tokenizers @require_tf class __a ( unittest.TestCase ): _a : int = ['My friends are cool but they eat too many carbs.'] _a : List[Any] = 'facebook/blenderbot-400M-distill' @cached_property def UpperCAmelCase__ ( self ) -> List[Any]: """simple docstring""" return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def UpperCAmelCase__ ( self ) -> Any: """simple docstring""" _UpperCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def UpperCAmelCase__ ( self ) -> str: """simple docstring""" _UpperCAmelCase = self.tokenizer(self.src_text , return_tensors='tf' ) _UpperCAmelCase = self.model.generate( model_inputs.input_ids , ) _UpperCAmelCase = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=_SCREAMING_SNAKE_CASE )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
185
0
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow snake_case : Optional[Any] = logging.getLogger() @unittest.skip('Temporarily disable the doc tests.' ) @require_torch @require_tf @slow class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = True , ): a :Optional[int] = [file for file in os.listdir(_lowerCamelCase ) if os.path.isfile(os.path.join(_lowerCamelCase , _lowerCamelCase ) )] if identifier is not None: a :Optional[Any] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_lowerCamelCase , _lowerCamelCase ): for n_ in n_identifier: a :Optional[Any] = [file for file in files if n_ not in file] else: a :Optional[int] = [file for file in files if n_identifier not in file] a :List[Any] = ignore_files or [] ignore_files.append('''__init__.py''' ) a :Union[str, Any] = [file for file in files if file not in ignore_files] for file in files: # Open all files print('''Testing''' , _lowerCamelCase ) if only_modules: a :Dict = file.split('''.''' )[0] try: a :int = getattr(_lowerCamelCase , _lowerCamelCase ) a :List[Any] = doctest.DocTestSuite(_lowerCamelCase ) a :int = unittest.TextTestRunner().run(_lowerCamelCase ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F'''{module_identifier} is not a module.''' ) else: a :Any = doctest.testfile(str('''..''' / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def SCREAMING_SNAKE_CASE__ ( self ): a :int = Path('''src/transformers''' ) a :Any = '''modeling''' a :Optional[Any] = [ '''modeling_ctrl.py''', '''modeling_tf_ctrl.py''', ] self.analyze_directory(_lowerCamelCase , identifier=_lowerCamelCase , ignore_files=_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :Union[str, Any] = Path('''src/transformers''' ) a :Any = '''tokenization''' self.analyze_directory(_lowerCamelCase , identifier=_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :Any = Path('''src/transformers''' ) a :Union[str, Any] = '''configuration''' self.analyze_directory(_lowerCamelCase , identifier=_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :Dict = Path('''src/transformers''' ) a :Any = ['''configuration''', '''modeling''', '''tokenization'''] self.analyze_directory(_lowerCamelCase , n_identifier=_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :Optional[int] = Path('''docs/source''' ) a :int = ['''favicon.ico'''] self.analyze_directory(_lowerCamelCase , ignore_files=_lowerCamelCase , only_modules=_lowerCamelCase )
94
"""simple docstring""" from sklearn.metrics import recall_score import datasets UpperCAmelCase__ = """ Recall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation: Recall = TP / (TP + FN) Where TP is the true positives and FN is the false negatives. """ UpperCAmelCase__ = """ Args: - **predictions** (`list` of `int`): The predicted labels. - **references** (`list` of `int`): The ground truth labels. - **labels** (`list` of `int`): The set of labels to include when `average` is not set to `binary`, and their order when average is `None`. Labels present in the data can be excluded in this input, for example to calculate a multiclass average ignoring a majority negative class, while labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in y_true and y_pred are used in sorted order. Defaults to None. - **pos_label** (`int`): The class label to use as the 'positive class' when calculating the recall. Defaults to `1`. - **average** (`string`): This parameter is required for multiclass/multilabel targets. If None, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `'binary'`. - `'binary'`: Only report results for the class specified by `pos_label`. This is applicable only if the target labels and predictions are binary. - `'micro'`: Calculate metrics globally by counting the total true positives, false negatives, and false positives. - `'macro'`: Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - `'weighted'`: Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `'macro'` to account for label imbalance. Note that it can result in an F-score that is not between precision and recall. - `'samples'`: Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). - **sample_weight** (`list` of `float`): Sample weights Defaults to `None`. - **zero_division** (): Sets the value to return when there is a zero division. Defaults to . - `'warn'`: If there is a zero division, the return value is `0`, but warnings are also raised. - `0`: If there is a zero division, the return value is `0`. - `1`: If there is a zero division, the return value is `1`. Returns: - **recall** (`float`, or `array` of `float`): Either the general recall score, or the recall scores for individual classes, depending on the values input to `labels` and `average`. Minimum possible value is 0. Maximum possible value is 1. A higher recall means that more of the positive examples have been labeled correctly. Therefore, a higher recall is generally considered better. Examples: Example 1-A simple example with some errors >>> recall_metric = datasets.load_metric('recall') >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1]) >>> print(results) {'recall': 0.6666666666666666} Example 2-The same example as Example 1, but with `pos_label=0` instead of the default `pos_label=1`. >>> recall_metric = datasets.load_metric('recall') >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], pos_label=0) >>> print(results) {'recall': 0.5} Example 3-The same example as Example 1, but with `sample_weight` included. >>> recall_metric = datasets.load_metric('recall') >>> sample_weight = [0.9, 0.2, 0.9, 0.3, 0.8] >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], sample_weight=sample_weight) >>> print(results) {'recall': 0.55} Example 4-A multiclass example, using different averages. >>> recall_metric = datasets.load_metric('recall') >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = recall_metric.compute(predictions=predictions, references=references, average='macro') >>> print(results) {'recall': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average='micro') >>> print(results) {'recall': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average='weighted') >>> print(results) {'recall': 0.3333333333333333} >>> results = recall_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {'recall': array([1., 0., 0.])} """ UpperCAmelCase__ = """ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): def lowerCAmelCase_ ( self : Tuple ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""int32""" ) ), """references""": datasets.Sequence(datasets.Value("""int32""" ) ), } if self.config_name == """multilabel""" else { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=["""https://scikit-learn.org/stable/modules/generated/sklearn.metrics.recall_score.html"""] , ) def lowerCAmelCase_ ( self : List[Any] , __lowerCAmelCase : int , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : int=None , __lowerCAmelCase : Dict=1 , __lowerCAmelCase : List[str]="binary" , __lowerCAmelCase : Any=None , __lowerCAmelCase : int="warn" , ): _UpperCAmelCase = recall_score( __lowerCAmelCase , __lowerCAmelCase , labels=__lowerCAmelCase , pos_label=__lowerCAmelCase , average=__lowerCAmelCase , sample_weight=__lowerCAmelCase , zero_division=__lowerCAmelCase , ) return {"recall": float(__lowerCAmelCase ) if score.size == 1 else score}
289
0
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class __snake_case : def __init__( self : Union[str, Any] , _snake_case : int , _snake_case : List[str]=13 , _snake_case : Any=7 , _snake_case : List[Any]=True , _snake_case : Any=True , _snake_case : Tuple=True , _snake_case : Any=True , _snake_case : int=99 , _snake_case : Dict=32 , _snake_case : Optional[int]=2 , _snake_case : Optional[Any]=4 , _snake_case : str=37 , _snake_case : Dict="gelu" , _snake_case : List[str]=0.1 , _snake_case : Any=0.1 , _snake_case : Optional[int]=512 , _snake_case : str=16 , _snake_case : Union[str, Any]=2 , _snake_case : Optional[Any]=0.0_2 , _snake_case : int=3 , _snake_case : List[str]=4 , _snake_case : Any=None , ): """simple docstring""" UpperCAmelCase_ = parent UpperCAmelCase_ = 13 UpperCAmelCase_ = 7 UpperCAmelCase_ = True UpperCAmelCase_ = True UpperCAmelCase_ = True UpperCAmelCase_ = True UpperCAmelCase_ = 99 UpperCAmelCase_ = 384 UpperCAmelCase_ = 2 UpperCAmelCase_ = 4 UpperCAmelCase_ = 37 UpperCAmelCase_ = '''gelu''' UpperCAmelCase_ = 0.1 UpperCAmelCase_ = 0.1 UpperCAmelCase_ = 512 UpperCAmelCase_ = 16 UpperCAmelCase_ = 2 UpperCAmelCase_ = 0.0_2 UpperCAmelCase_ = 3 UpperCAmelCase_ = 4 UpperCAmelCase_ = 128 UpperCAmelCase_ = 2 UpperCAmelCase_ = 9 UpperCAmelCase_ = 1 UpperCAmelCase_ = None def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) UpperCAmelCase_ = None if self.use_input_mask: UpperCAmelCase_ = random_attention_mask([self.batch_size, self.seq_length]) UpperCAmelCase_ = None if self.use_token_type_ids: UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size) UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_choices) UpperCAmelCase_ = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=_snake_case , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase ( self : str , _snake_case : Any , _snake_case : Dict , _snake_case : str , _snake_case : Tuple , _snake_case : Optional[int] , _snake_case : Union[str, Any] , _snake_case : Tuple): """simple docstring""" UpperCAmelCase_ = TFConvBertModel(config=_snake_case) UpperCAmelCase_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} UpperCAmelCase_ = [input_ids, input_mask] UpperCAmelCase_ = model(_snake_case) UpperCAmelCase_ = model(_snake_case) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def lowerCamelCase ( self : Dict , _snake_case : Optional[int] , _snake_case : Tuple , _snake_case : Dict , _snake_case : Dict , _snake_case : str , _snake_case : Optional[int] , _snake_case : str): """simple docstring""" UpperCAmelCase_ = TFConvBertForMaskedLM(config=_snake_case) UpperCAmelCase_ = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } UpperCAmelCase_ = model(_snake_case) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def lowerCamelCase ( self : Any , _snake_case : int , _snake_case : List[str] , _snake_case : Union[str, Any] , _snake_case : int , _snake_case : List[str] , _snake_case : str , _snake_case : Optional[int]): """simple docstring""" UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = TFConvBertForSequenceClassification(config=_snake_case) UpperCAmelCase_ = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } UpperCAmelCase_ = model(_snake_case) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def lowerCamelCase ( self : Union[str, Any] , _snake_case : List[str] , _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : Union[str, Any] , _snake_case : List[Any] , _snake_case : List[str] , _snake_case : Tuple): """simple docstring""" UpperCAmelCase_ = self.num_choices UpperCAmelCase_ = TFConvBertForMultipleChoice(config=_snake_case) UpperCAmelCase_ = tf.tile(tf.expand_dims(_snake_case , 1) , (1, self.num_choices, 1)) UpperCAmelCase_ = tf.tile(tf.expand_dims(_snake_case , 1) , (1, self.num_choices, 1)) UpperCAmelCase_ = tf.tile(tf.expand_dims(_snake_case , 1) , (1, self.num_choices, 1)) UpperCAmelCase_ = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } UpperCAmelCase_ = model(_snake_case) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def lowerCamelCase ( self : Tuple , _snake_case : Any , _snake_case : str , _snake_case : List[str] , _snake_case : int , _snake_case : List[str] , _snake_case : Optional[int] , _snake_case : Optional[Any]): """simple docstring""" UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = TFConvBertForTokenClassification(config=_snake_case) UpperCAmelCase_ = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } UpperCAmelCase_ = model(_snake_case) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def lowerCamelCase ( self : Tuple , _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : str , _snake_case : Dict , _snake_case : List[Any] , _snake_case : str , _snake_case : Union[str, Any]): """simple docstring""" UpperCAmelCase_ = TFConvBertForQuestionAnswering(config=_snake_case) UpperCAmelCase_ = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } UpperCAmelCase_ = model(_snake_case) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def lowerCamelCase ( self : List[Any]): """simple docstring""" UpperCAmelCase_ = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) = config_and_inputs UpperCAmelCase_ = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __snake_case ( a , a , unittest.TestCase ): UpperCAmelCase__ : List[Any] = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) UpperCAmelCase__ : List[Any] = ( { '''feature-extraction''': TFConvBertModel, '''fill-mask''': TFConvBertForMaskedLM, '''question-answering''': TFConvBertForQuestionAnswering, '''text-classification''': TFConvBertForSequenceClassification, '''token-classification''': TFConvBertForTokenClassification, '''zero-shot''': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) UpperCAmelCase__ : Tuple = False UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : List[str] = False def lowerCamelCase ( self : Optional[Any]): """simple docstring""" UpperCAmelCase_ = TFConvBertModelTester(self) UpperCAmelCase_ = ConfigTester(self , config_class=_snake_case , hidden_size=37) def lowerCamelCase ( self : List[Any]): """simple docstring""" self.config_tester.run_common_tests() def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case) def lowerCamelCase ( self : Any): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_snake_case) def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_snake_case) def lowerCamelCase ( self : int): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_snake_case) def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_snake_case) def lowerCamelCase ( self : Optional[Any]): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_snake_case) @slow def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = True UpperCAmelCase_ = True if hasattr(_snake_case , '''use_cache'''): UpperCAmelCase_ = True UpperCAmelCase_ = getattr(self.model_tester , '''encoder_seq_length''' , self.model_tester.seq_length) UpperCAmelCase_ = getattr(self.model_tester , '''key_length''' , _snake_case) for model_class in self.all_model_classes: UpperCAmelCase_ = self._prepare_for_class(_snake_case , _snake_case) UpperCAmelCase_ = model_class(_snake_case) UpperCAmelCase_ = len(model(_snake_case)) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_snake_case , saved_model=_snake_case) UpperCAmelCase_ = os.path.join(_snake_case , '''saved_model''' , '''1''') UpperCAmelCase_ = tf.keras.models.load_model(_snake_case) UpperCAmelCase_ = model(_snake_case) if self.is_encoder_decoder: UpperCAmelCase_ = outputs['''encoder_hidden_states'''] UpperCAmelCase_ = outputs['''encoder_attentions'''] else: UpperCAmelCase_ = outputs['''hidden_states'''] UpperCAmelCase_ = outputs['''attentions'''] self.assertEqual(len(_snake_case) , _snake_case) UpperCAmelCase_ = getattr( self.model_tester , '''expected_num_hidden_layers''' , self.model_tester.num_hidden_layers + 1) self.assertEqual(len(_snake_case) , _snake_case) self.assertListEqual( list(output_hidden_states[0].shape[-2:]) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(_snake_case) , self.model_tester.num_hidden_layers) self.assertListEqual( list(output_attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def lowerCamelCase ( self : Tuple): """simple docstring""" UpperCAmelCase_ = TFConvBertModel.from_pretrained('''YituTech/conv-bert-base''') self.assertIsNotNone(_snake_case) def lowerCamelCase ( self : Tuple): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = True UpperCAmelCase_ = getattr(self.model_tester , '''decoder_seq_length''' , self.model_tester.seq_length) UpperCAmelCase_ = getattr(self.model_tester , '''encoder_seq_length''' , self.model_tester.seq_length) UpperCAmelCase_ = getattr(self.model_tester , '''key_length''' , _snake_case) UpperCAmelCase_ = getattr(self.model_tester , '''key_length''' , _snake_case) def check_decoder_attentions_output(_snake_case : List[str]): UpperCAmelCase_ = len(_snake_case) self.assertEqual(out_len % 2 , 0) UpperCAmelCase_ = outputs.decoder_attentions self.assertEqual(len(_snake_case) , self.model_tester.num_hidden_layers) self.assertListEqual( list(decoder_attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(_snake_case : Optional[Any]): UpperCAmelCase_ = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(_snake_case) , self.model_tester.num_hidden_layers) self.assertListEqual( list(attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: UpperCAmelCase_ = True UpperCAmelCase_ = False UpperCAmelCase_ = model_class(_snake_case) UpperCAmelCase_ = model(self._prepare_for_class(_snake_case , _snake_case)) UpperCAmelCase_ = len(_snake_case) self.assertEqual(config.output_hidden_states , _snake_case) check_encoder_attentions_output(_snake_case) if self.is_encoder_decoder: UpperCAmelCase_ = model_class(_snake_case) UpperCAmelCase_ = model(self._prepare_for_class(_snake_case , _snake_case)) self.assertEqual(config.output_hidden_states , _snake_case) check_decoder_attentions_output(_snake_case) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCAmelCase_ = True UpperCAmelCase_ = model_class(_snake_case) UpperCAmelCase_ = model(self._prepare_for_class(_snake_case , _snake_case)) self.assertEqual(config.output_hidden_states , _snake_case) check_encoder_attentions_output(_snake_case) # Check attention is always last and order is fine UpperCAmelCase_ = True UpperCAmelCase_ = True UpperCAmelCase_ = model_class(_snake_case) UpperCAmelCase_ = model(self._prepare_for_class(_snake_case , _snake_case)) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(_snake_case)) self.assertEqual(model.config.output_hidden_states , _snake_case) check_encoder_attentions_output(_snake_case) @require_tf class __snake_case ( unittest.TestCase ): @slow def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ = TFConvBertModel.from_pretrained('''YituTech/conv-bert-base''') UpperCAmelCase_ = tf.constant([[0, 1, 2, 3, 4, 5]]) UpperCAmelCase_ = model(_snake_case)[0] UpperCAmelCase_ = [1, 6, 768] self.assertEqual(output.shape , _snake_case) UpperCAmelCase_ = tf.constant( [ [ [-0.0_3_4_7_5_4_9_3, -0.4_6_8_6_0_3_4, -0.3_0_6_3_8_8_3_2], [0.2_2_6_3_7_2_4_8, -0.2_6_9_8_8_6_4_6, -0.7_4_2_3_4_2_4], [0.1_0_3_2_4_8_6_8, -0.4_5_0_1_3_5_0_8, -0.5_8_2_8_0_7_8_4], ] ]) tf.debugging.assert_near(output[:, :3, :3] , _snake_case , atol=1e-4)
363
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig snake_case_ : Union[str, Any] = logging.get_logger(__name__) class __snake_case : def __init__( self : int , _snake_case : List[Any] , _snake_case : Tuple): """simple docstring""" UpperCAmelCase_ = question_encoder UpperCAmelCase_ = generator UpperCAmelCase_ = self.question_encoder def lowerCamelCase ( self : Union[str, Any] , _snake_case : Optional[int]): """simple docstring""" if os.path.isfile(_snake_case): raise ValueError(F"""Provided path ({save_directory}) should be a directory, not a file""") os.makedirs(_snake_case , exist_ok=_snake_case) UpperCAmelCase_ = os.path.join(_snake_case , '''question_encoder_tokenizer''') UpperCAmelCase_ = os.path.join(_snake_case , '''generator_tokenizer''') self.question_encoder.save_pretrained(_snake_case) self.generator.save_pretrained(_snake_case) @classmethod def lowerCamelCase ( cls : Optional[Any] , _snake_case : Optional[Any] , **_snake_case : Optional[int]): """simple docstring""" from ..auto.tokenization_auto import AutoTokenizer UpperCAmelCase_ = kwargs.pop('''config''' , _snake_case) if config is None: UpperCAmelCase_ = RagConfig.from_pretrained(_snake_case) UpperCAmelCase_ = AutoTokenizer.from_pretrained( _snake_case , config=config.question_encoder , subfolder='''question_encoder_tokenizer''') UpperCAmelCase_ = AutoTokenizer.from_pretrained( _snake_case , config=config.generator , subfolder='''generator_tokenizer''') return cls(question_encoder=_snake_case , generator=_snake_case) def __call__( self : List[Any] , *_snake_case : List[str] , **_snake_case : List[Any]): """simple docstring""" return self.current_tokenizer(*_snake_case , **_snake_case) def lowerCamelCase ( self : List[Any] , *_snake_case : str , **_snake_case : Union[str, Any]): """simple docstring""" return self.generator.batch_decode(*_snake_case , **_snake_case) def lowerCamelCase ( self : str , *_snake_case : Optional[int] , **_snake_case : Any): """simple docstring""" return self.generator.decode(*_snake_case , **_snake_case) def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ = self.question_encoder def lowerCamelCase ( self : Tuple): """simple docstring""" UpperCAmelCase_ = self.generator def lowerCamelCase ( self : Optional[Any] , _snake_case : List[str] , _snake_case : Optional[List[str]] = None , _snake_case : Optional[int] = None , _snake_case : Optional[int] = None , _snake_case : str = "longest" , _snake_case : str = None , _snake_case : bool = True , **_snake_case : Optional[int] , ): """simple docstring""" warnings.warn( '''`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the ''' '''regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` ''' '''context manager to prepare your targets. See the documentation of your specific tokenizer for more ''' '''details''' , _snake_case , ) if max_length is None: UpperCAmelCase_ = self.current_tokenizer.model_max_length UpperCAmelCase_ = self( _snake_case , add_special_tokens=_snake_case , return_tensors=_snake_case , max_length=_snake_case , padding=_snake_case , truncation=_snake_case , **_snake_case , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: UpperCAmelCase_ = self.current_tokenizer.model_max_length UpperCAmelCase_ = self( text_target=_snake_case , add_special_tokens=_snake_case , return_tensors=_snake_case , padding=_snake_case , max_length=_snake_case , truncation=_snake_case , **_snake_case , ) UpperCAmelCase_ = labels['''input_ids'''] return model_inputs
7
0
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=_lowerCamelCase ) class A_ ( _lowerCamelCase ): """simple docstring""" __UpperCamelCase = field(default="""automatic-speech-recognition""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) __UpperCamelCase = Features({"""audio""": Audio()} ) __UpperCamelCase = Features({"""transcription""": Value("""string""" )} ) __UpperCamelCase = """audio""" __UpperCamelCase = """transcription""" def UpperCAmelCase__ ( self :Any , lowercase_ :Optional[int] ) -> Optional[Any]: if self.audio_column not in features: raise ValueError(f"""Column {self.audio_column} is not present in features.""" ) if not isinstance(features[self.audio_column] , lowercase_ ): raise ValueError(f"""Column {self.audio_column} is not an Audio type.""" ) UpperCAmelCase = copy.deepcopy(self ) UpperCAmelCase = self.input_schema.copy() UpperCAmelCase = features[self.audio_column] UpperCAmelCase = input_schema return task_template @property def UpperCAmelCase__ ( self :Any ) -> str: return {self.audio_column: "audio", self.transcription_column: "transcription"}
78
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) a : Tuple = { 'configuration_llama': ['LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LlamaConfig'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['LlamaTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = ['LlamaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = [ 'LlamaForCausalLM', 'LlamaModel', 'LlamaPreTrainedModel', 'LlamaForSequenceClassification', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys a : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
56
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __A : Optional[Any] = logging.get_logger(__name__) __A : Tuple = { '''andreasmadsen/efficient_mlm_m0.40''': ( '''https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json''' ), } class lowerCamelCase ( _UpperCAmelCase ): lowercase : Union[str, Any] = 'roberta-prelayernorm' def __init__( self , SCREAMING_SNAKE_CASE_=5_0265 , SCREAMING_SNAKE_CASE_=768 , SCREAMING_SNAKE_CASE_=12 , SCREAMING_SNAKE_CASE_=12 , SCREAMING_SNAKE_CASE_=3072 , SCREAMING_SNAKE_CASE_="gelu" , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=512 , SCREAMING_SNAKE_CASE_=2 , SCREAMING_SNAKE_CASE_=0.02 , SCREAMING_SNAKE_CASE_=1e-12 , SCREAMING_SNAKE_CASE_=1 , SCREAMING_SNAKE_CASE_=0 , SCREAMING_SNAKE_CASE_=2 , SCREAMING_SNAKE_CASE_="absolute" , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=None , **SCREAMING_SNAKE_CASE_ , ): super().__init__(pad_token_id=SCREAMING_SNAKE_CASE_ , bos_token_id=SCREAMING_SNAKE_CASE_ , eos_token_id=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) UpperCamelCase : int = vocab_size UpperCamelCase : List[str] = hidden_size UpperCamelCase : List[Any] = num_hidden_layers UpperCamelCase : List[str] = num_attention_heads UpperCamelCase : Tuple = hidden_act UpperCamelCase : List[str] = intermediate_size UpperCamelCase : List[Any] = hidden_dropout_prob UpperCamelCase : Dict = attention_probs_dropout_prob UpperCamelCase : Union[str, Any] = max_position_embeddings UpperCamelCase : Optional[Any] = type_vocab_size UpperCamelCase : Dict = initializer_range UpperCamelCase : List[str] = layer_norm_eps UpperCamelCase : Any = position_embedding_type UpperCamelCase : List[str] = use_cache UpperCamelCase : Union[str, Any] = classifier_dropout class lowerCamelCase ( _UpperCAmelCase ): @property def a_ ( self ): if self.task == "multiple-choice": UpperCamelCase : Optional[int] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: UpperCamelCase : Any = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
350
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __A : int = { '''configuration_gpt_bigcode''': ['''GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTBigCodeConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple = [ '''GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTBigCodeForSequenceClassification''', '''GPTBigCodeForTokenClassification''', '''GPTBigCodeForCausalLM''', '''GPTBigCodeModel''', '''GPTBigCodePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys __A : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
27
0
'''simple docstring''' from abc import ABC, abstractmethod from typing import List, Optional class _snake_case ( a__ ): def __init__( self): # test for the above condition self.test() def snake_case__ ( self): UpperCAmelCase__ : Dict = 0 UpperCAmelCase__ : int = False while not completed: if counter == 1: self.reset() UpperCAmelCase__ : Dict = self.advance() if not self.does_advance(_lowerCamelCase): raise Exception( """Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.""") UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = self.update(_lowerCamelCase) counter += 1 if counter > 1_0000: raise Exception("""update() does not fulfill the constraint.""") if self.remaining() != 0: raise Exception("""Custom Constraint is not defined correctly.""") @abstractmethod def snake_case__ ( self): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''') @abstractmethod def snake_case__ ( self , _lowerCamelCase): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''') @abstractmethod def snake_case__ ( self , _lowerCamelCase): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''') @abstractmethod def snake_case__ ( self): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''') @abstractmethod def snake_case__ ( self): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''') @abstractmethod def snake_case__ ( self , _lowerCamelCase=False): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''') class _snake_case ( a__ ): def __init__( self , _lowerCamelCase): super(_lowerCamelCase , self).__init__() if not isinstance(_lowerCamelCase , _lowerCamelCase) or len(_lowerCamelCase) == 0: raise ValueError(f'''`token_ids` has to be a non-empty list, but is {token_ids}.''') if any((not isinstance(_lowerCamelCase , _lowerCamelCase) or token_id < 0) for token_id in token_ids): raise ValueError(f'''Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.''') UpperCAmelCase__ : Optional[Any] = token_ids UpperCAmelCase__ : int = len(self.token_ids) UpperCAmelCase__ : Union[str, Any] = -1 # the index of the currently fulfilled step UpperCAmelCase__ : List[str] = False def snake_case__ ( self): if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def snake_case__ ( self , _lowerCamelCase): if not isinstance(_lowerCamelCase , _lowerCamelCase): raise ValueError(f'''`token_id` has to be an `int`, but is {token_id} of type {type(_lowerCamelCase)}''') if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def snake_case__ ( self , _lowerCamelCase): if not isinstance(_lowerCamelCase , _lowerCamelCase): raise ValueError(f'''`token_id` has to be an `int`, but is {token_id} of type {type(_lowerCamelCase)}''') UpperCAmelCase__ : Optional[int] = False UpperCAmelCase__ : Tuple = False UpperCAmelCase__ : int = False if self.does_advance(_lowerCamelCase): self.fulfilled_idx += 1 UpperCAmelCase__ : Union[str, Any] = True if self.fulfilled_idx == (self.seqlen - 1): UpperCAmelCase__ : Any = True UpperCAmelCase__ : Union[str, Any] = completed else: # failed to make progress. UpperCAmelCase__ : int = True self.reset() return stepped, completed, reset def snake_case__ ( self): UpperCAmelCase__ : Optional[int] = False UpperCAmelCase__ : Dict = 0 def snake_case__ ( self): return self.seqlen - (self.fulfilled_idx + 1) def snake_case__ ( self , _lowerCamelCase=False): UpperCAmelCase__ : Union[str, Any] = PhrasalConstraint(self.token_ids) if stateful: UpperCAmelCase__ : Optional[Any] = self.seqlen UpperCAmelCase__ : Optional[Any] = self.fulfilled_idx UpperCAmelCase__ : List[Any] = self.completed return new_constraint class _snake_case : def __init__( self , _lowerCamelCase , _lowerCamelCase=True): UpperCAmelCase__ : str = max([len(_lowerCamelCase) for one in nested_token_ids]) UpperCAmelCase__ : Optional[int] = {} for token_ids in nested_token_ids: UpperCAmelCase__ : Any = root for tidx, token_id in enumerate(_lowerCamelCase): if token_id not in level: UpperCAmelCase__ : List[Any] = {} UpperCAmelCase__ : List[str] = level[token_id] if no_subsets and self.has_subsets(_lowerCamelCase , _lowerCamelCase): raise ValueError( """Each list in `nested_token_ids` can't be a complete subset of another list, but is""" f''' {nested_token_ids}.''') UpperCAmelCase__ : Any = root def snake_case__ ( self , _lowerCamelCase): UpperCAmelCase__ : int = self.trie for current_token in current_seq: UpperCAmelCase__ : Optional[Any] = start[current_token] UpperCAmelCase__ : Dict = list(start.keys()) return next_tokens def snake_case__ ( self , _lowerCamelCase): UpperCAmelCase__ : Any = self.next_tokens(_lowerCamelCase) return len(_lowerCamelCase) == 0 def snake_case__ ( self , _lowerCamelCase): UpperCAmelCase__ : Any = list(root.values()) if len(_lowerCamelCase) == 0: return 1 else: return sum([self.count_leaves(_lowerCamelCase) for nn in next_nodes]) def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase): UpperCAmelCase__ : Union[str, Any] = self.count_leaves(_lowerCamelCase) return len(_lowerCamelCase) != leaf_count class _snake_case ( a__ ): def __init__( self , _lowerCamelCase): super(_lowerCamelCase , self).__init__() if not isinstance(_lowerCamelCase , _lowerCamelCase) or len(_lowerCamelCase) == 0: raise ValueError(f'''`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.''') if any(not isinstance(_lowerCamelCase , _lowerCamelCase) for token_ids in nested_token_ids): raise ValueError(f'''`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.''') if any( any((not isinstance(_lowerCamelCase , _lowerCamelCase) or token_id < 0) for token_id in token_ids) for token_ids in nested_token_ids): raise ValueError( f'''Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.''') UpperCAmelCase__ : int = DisjunctiveTrie(_lowerCamelCase) UpperCAmelCase__ : List[str] = nested_token_ids UpperCAmelCase__ : Dict = self.trie.max_height UpperCAmelCase__ : str = [] UpperCAmelCase__ : List[Any] = False def snake_case__ ( self): UpperCAmelCase__ : int = self.trie.next_tokens(self.current_seq) if len(_lowerCamelCase) == 0: return None else: return token_list def snake_case__ ( self , _lowerCamelCase): if not isinstance(_lowerCamelCase , _lowerCamelCase): raise ValueError(f'''`token_id` is supposed to be type `int`, but is {token_id} of type {type(_lowerCamelCase)}''') UpperCAmelCase__ : Dict = self.trie.next_tokens(self.current_seq) return token_id in next_tokens def snake_case__ ( self , _lowerCamelCase): if not isinstance(_lowerCamelCase , _lowerCamelCase): raise ValueError(f'''`token_id` is supposed to be type `int`, but is {token_id} of type {type(_lowerCamelCase)}''') UpperCAmelCase__ : str = False UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : Dict = False if self.does_advance(_lowerCamelCase): self.current_seq.append(_lowerCamelCase) UpperCAmelCase__ : int = True else: UpperCAmelCase__ : Optional[int] = True self.reset() UpperCAmelCase__ : Any = self.trie.reached_leaf(self.current_seq) UpperCAmelCase__ : int = completed return stepped, completed, reset def snake_case__ ( self): UpperCAmelCase__ : int = False UpperCAmelCase__ : Any = [] def snake_case__ ( self): if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq) def snake_case__ ( self , _lowerCamelCase=False): UpperCAmelCase__ : Union[str, Any] = DisjunctiveConstraint(self.token_ids) if stateful: UpperCAmelCase__ : List[Any] = self.seqlen UpperCAmelCase__ : int = self.current_seq UpperCAmelCase__ : Any = self.completed return new_constraint class _snake_case : def __init__( self , _lowerCamelCase): UpperCAmelCase__ : Union[str, Any] = constraints # max # of steps required to fulfill a given constraint UpperCAmelCase__ : Any = max([c.seqlen for c in constraints]) UpperCAmelCase__ : Tuple = len(_lowerCamelCase) UpperCAmelCase__ : Any = False self.init_state() def snake_case__ ( self): UpperCAmelCase__ : Optional[Any] = [] UpperCAmelCase__ : Optional[int] = None UpperCAmelCase__ : List[Any] = [constraint.copy(stateful=_lowerCamelCase) for constraint in self.constraints] def snake_case__ ( self): UpperCAmelCase__ : Optional[int] = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints) * self.max_seqlen) + add def snake_case__ ( self): UpperCAmelCase__ : Union[str, Any] = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" UpperCAmelCase__ : str = constraint.advance() if isinstance(_lowerCamelCase , _lowerCamelCase): token_list.append(_lowerCamelCase) elif isinstance(_lowerCamelCase , _lowerCamelCase): token_list.extend(_lowerCamelCase) else: UpperCAmelCase__ : Any = self.inprogress_constraint.advance() if isinstance(_lowerCamelCase , _lowerCamelCase): token_list.append(_lowerCamelCase) elif isinstance(_lowerCamelCase , _lowerCamelCase): token_list.extend(_lowerCamelCase) if len(_lowerCamelCase) == 0: return None else: return token_list def snake_case__ ( self , _lowerCamelCase): self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint UpperCAmelCase__ , UpperCAmelCase__ : Any = self.add(_lowerCamelCase) # the entire list of constraints are fulfilled if self.completed: break def snake_case__ ( self , _lowerCamelCase): if not isinstance(_lowerCamelCase , _lowerCamelCase): raise ValueError(f'''`token_id` should be an `int`, but is `{token_id}`.''') UpperCAmelCase__ , UpperCAmelCase__ : List[str] = False, False if self.completed: UpperCAmelCase__ : Tuple = True UpperCAmelCase__ : Optional[Any] = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Dict = self.inprogress_constraint.update(_lowerCamelCase) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=_lowerCamelCase)) UpperCAmelCase__ : str = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint) UpperCAmelCase__ : Tuple = None if len(self.pending_constraints) == 0: # we're done! UpperCAmelCase__ : List[str] = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints): if pending_constraint.does_advance(_lowerCamelCase): UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = pending_constraint.update(_lowerCamelCase) if not stepped: raise Exception( """`constraint.update(token_id)` is not yielding incremental progress, """ """even though `constraint.does_advance(token_id)` is true.""") if complete: self.complete_constraints.append(_lowerCamelCase) UpperCAmelCase__ : Any = None if not complete and stepped: UpperCAmelCase__ : Union[str, Any] = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". UpperCAmelCase__ : Dict = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. UpperCAmelCase__ : Any = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def snake_case__ ( self , _lowerCamelCase=True): UpperCAmelCase__ : List[Any] = ConstraintListState(self.constraints) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: UpperCAmelCase__ : int = [ constraint.copy(stateful=_lowerCamelCase) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: UpperCAmelCase__ : str = self.inprogress_constraint.copy(stateful=_lowerCamelCase) UpperCAmelCase__ : str = [constraint.copy() for constraint in self.pending_constraints] return new_state
163
'''simple docstring''' import unittest from transformers.testing_utils import require_bsa from transformers.utils import is_bsa_available from ...test_feature_extraction_common import FeatureExtractionSavingTestMixin if is_bsa_available(): from transformers import MarkupLMFeatureExtractor class _snake_case ( unittest.TestCase ): def __init__( self , _lowerCamelCase): UpperCAmelCase__ : Any = parent def snake_case__ ( self): return {} def _UpperCamelCase ( ): UpperCAmelCase__ : List[str] = """<HTML> <HEAD> <TITLE>sample document</TITLE> </HEAD> <BODY BGCOLOR=\"FFFFFF\"> <HR> <a href=\"http://google.com\">Goog</a> <H1>This is one header</H1> <H2>This is a another Header</H2> <P>Travel from <P> <B>SFO to JFK</B> <BR> <B><I>on May 2, 2015 at 2:00 pm. For details go to confirm.com </I></B> <HR> <div style=\"color:#0000FF\"> <h3>Traveler <b> name </b> is <p> John Doe </p> </div>""" UpperCAmelCase__ : Tuple = """ <!DOCTYPE html> <html> <body> <h1>My First Heading</h1> <p>My first paragraph.</p> </body> </html> """ return [html_string_a, html_string_a] @require_bsa class _snake_case ( a__ , unittest.TestCase ): lowerCAmelCase :Optional[int] = MarkupLMFeatureExtractor if is_bsa_available() else None def snake_case__ ( self): UpperCAmelCase__ : Union[str, Any] = MarkupLMFeatureExtractionTester(self) @property def snake_case__ ( self): return self.feature_extract_tester.prepare_feat_extract_dict() def snake_case__ ( self): # Initialize feature_extractor UpperCAmelCase__ : List[Any] = self.feature_extraction_class() # Test not batched input UpperCAmelCase__ : Optional[Any] = get_html_strings()[0] UpperCAmelCase__ : Any = feature_extractor(_lowerCamelCase) # fmt: off UpperCAmelCase__ : Dict = [["""sample document""", """Goog""", """This is one header""", """This is a another Header""", """Travel from""", """SFO to JFK""", """on May 2, 2015 at 2:00 pm. For details go to confirm.com""", """Traveler""", """name""", """is""", """John Doe"""]] UpperCAmelCase__ : List[str] = [["""/html/head/title""", """/html/body/a""", """/html/body/h1""", """/html/body/h2""", """/html/body/p""", """/html/body/p/p/b[1]""", """/html/body/p/p/b[2]/i""", """/html/body/p/p/div/h3""", """/html/body/p/p/div/h3/b""", """/html/body/p/p/div/h3""", """/html/body/p/p/div/h3/p"""]] # fmt: on self.assertEqual(encoding.nodes , _lowerCamelCase) self.assertEqual(encoding.xpaths , _lowerCamelCase) # Test batched UpperCAmelCase__ : int = get_html_strings() UpperCAmelCase__ : Optional[Any] = feature_extractor(_lowerCamelCase) # fmt: off UpperCAmelCase__ : List[str] = expected_nodes + [["""My First Heading""", """My first paragraph."""]] UpperCAmelCase__ : str = expected_xpaths + [["""/html/body/h1""", """/html/body/p"""]] self.assertEqual(len(encoding.nodes) , 2) self.assertEqual(len(encoding.xpaths) , 2) self.assertEqual(encoding.nodes , _lowerCamelCase) self.assertEqual(encoding.xpaths , _lowerCamelCase)
163
1
from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : Tuple = logging.get_logger(__name__) a__ : List[Any] = { '''snap-research/efficientformer-l1-300''': ( '''https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json''' ), } class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : Any = "efficientformer" def __init__( self : Any , UpperCAmelCase__ : List[int] = [3, 2, 6, 4] , UpperCAmelCase__ : List[int] = [4_8, 9_6, 2_2_4, 4_4_8] , UpperCAmelCase__ : List[bool] = [True, True, True, True] , UpperCAmelCase__ : int = 4_4_8 , UpperCAmelCase__ : int = 3_2 , UpperCAmelCase__ : int = 4 , UpperCAmelCase__ : int = 7 , UpperCAmelCase__ : int = 5 , UpperCAmelCase__ : int = 8 , UpperCAmelCase__ : int = 4 , UpperCAmelCase__ : float = 0.0 , UpperCAmelCase__ : int = 1_6 , UpperCAmelCase__ : int = 3 , UpperCAmelCase__ : int = 3 , UpperCAmelCase__ : int = 3 , UpperCAmelCase__ : int = 2 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : float = 0.0 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : float = 1E-5 , UpperCAmelCase__ : str = "gelu" , UpperCAmelCase__ : float = 0.02 , UpperCAmelCase__ : float = 1E-12 , UpperCAmelCase__ : int = 2_2_4 , UpperCAmelCase__ : float = 1E-05 , **UpperCAmelCase__ : Tuple , ) -> None: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = hidden_sizes __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = layer_norm_eps __SCREAMING_SNAKE_CASE = patch_size __SCREAMING_SNAKE_CASE = num_channels __SCREAMING_SNAKE_CASE = depths __SCREAMING_SNAKE_CASE = mlp_expansion_ratio __SCREAMING_SNAKE_CASE = downsamples __SCREAMING_SNAKE_CASE = dim __SCREAMING_SNAKE_CASE = key_dim __SCREAMING_SNAKE_CASE = attention_ratio __SCREAMING_SNAKE_CASE = resolution __SCREAMING_SNAKE_CASE = pool_size __SCREAMING_SNAKE_CASE = downsample_patch_size __SCREAMING_SNAKE_CASE = downsample_stride __SCREAMING_SNAKE_CASE = downsample_pad __SCREAMING_SNAKE_CASE = drop_path_rate __SCREAMING_SNAKE_CASE = num_metaad_blocks __SCREAMING_SNAKE_CASE = distillation __SCREAMING_SNAKE_CASE = use_layer_scale __SCREAMING_SNAKE_CASE = layer_scale_init_value __SCREAMING_SNAKE_CASE = image_size __SCREAMING_SNAKE_CASE = batch_norm_eps
366
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' if not nums: return 0 __SCREAMING_SNAKE_CASE = nums[0] __SCREAMING_SNAKE_CASE = 0 for num in nums[1:]: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = ( max_excluding + num, max(lowerCAmelCase_ , lowerCAmelCase_ ), ) return max(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
195
0
from math import pi, sqrt, tan def __snake_case ( __UpperCamelCase : float ): """simple docstring""" if side_length < 0: raise ValueError("surface_area_cube() only accepts non-negative values" ) return 6 * side_length**2 def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if length < 0 or breadth < 0 or height < 0: raise ValueError("surface_area_cuboid() only accepts non-negative values" ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def __snake_case ( __UpperCamelCase : float ): """simple docstring""" if radius < 0: raise ValueError("surface_area_sphere() only accepts non-negative values" ) return 4 * pi * radius**2 def __snake_case ( __UpperCamelCase : float ): """simple docstring""" if radius < 0: raise ValueError("surface_area_hemisphere() only accepts non-negative values" ) return 3 * pi * radius**2 def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if radius < 0 or height < 0: raise ValueError("surface_area_cone() only accepts non-negative values" ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( "surface_area_conical_frustum() only accepts non-negative values" ) A_ = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if radius < 0 or height < 0: raise ValueError("surface_area_cylinder() only accepts non-negative values" ) return 2 * pi * radius * (height + radius) def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if torus_radius < 0 or tube_radius < 0: raise ValueError("surface_area_torus() only accepts non-negative values" ) if torus_radius < tube_radius: raise ValueError( "surface_area_torus() does not support spindle or self intersecting tori" ) return 4 * pow(__UpperCamelCase ,2 ) * torus_radius * tube_radius def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if length < 0 or width < 0: raise ValueError("area_rectangle() only accepts non-negative values" ) return length * width def __snake_case ( __UpperCamelCase : float ): """simple docstring""" if side_length < 0: raise ValueError("area_square() only accepts non-negative values" ) return side_length**2 def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if base < 0 or height < 0: raise ValueError("area_triangle() only accepts non-negative values" ) return (base * height) / 2 def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError("area_triangle_three_sides() only accepts non-negative values" ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError("Given three sides do not form a triangle" ) A_ = (sidea + sidea + sidea) / 2 A_ = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if base < 0 or height < 0: raise ValueError("area_parallelogram() only accepts non-negative values" ) return base * height def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if basea < 0 or basea < 0 or height < 0: raise ValueError("area_trapezium() only accepts non-negative values" ) return 1 / 2 * (basea + basea) * height def __snake_case ( __UpperCamelCase : float ): """simple docstring""" if radius < 0: raise ValueError("area_circle() only accepts non-negative values" ) return pi * radius**2 def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if radius_x < 0 or radius_y < 0: raise ValueError("area_ellipse() only accepts non-negative values" ) return pi * radius_x * radius_y def __snake_case ( __UpperCamelCase : float ,__UpperCamelCase : float ): """simple docstring""" if diagonal_a < 0 or diagonal_a < 0: raise ValueError("area_rhombus() only accepts non-negative values" ) return 1 / 2 * diagonal_a * diagonal_a def __snake_case ( __UpperCamelCase : int ,__UpperCamelCase : float ): """simple docstring""" if not isinstance(__UpperCamelCase ,__UpperCamelCase ) or sides < 3: raise ValueError( "area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides" ) elif length < 0: raise ValueError( "area_reg_polygon() only accepts non-negative values as \ length of a side" ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('[DEMO] Areas of various geometric shapes: \n') print(F"Rectangle: {area_rectangle(10, 20) = }") print(F"Square: {area_square(10) = }") print(F"Triangle: {area_triangle(10, 10) = }") print(F"Triangle: {area_triangle_three_sides(5, 12, 13) = }") print(F"Parallelogram: {area_parallelogram(10, 20) = }") print(F"Rhombus: {area_rhombus(10, 20) = }") print(F"Trapezium: {area_trapezium(10, 20, 30) = }") print(F"Circle: {area_circle(20) = }") print(F"Ellipse: {area_ellipse(10, 20) = }") print('\nSurface Areas of various geometric shapes: \n') print(F"Cube: {surface_area_cube(20) = }") print(F"Cuboid: {surface_area_cuboid(10, 20, 30) = }") print(F"Sphere: {surface_area_sphere(20) = }") print(F"Hemisphere: {surface_area_hemisphere(20) = }") print(F"Cone: {surface_area_cone(10, 20) = }") print(F"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }") print(F"Cylinder: {surface_area_cylinder(10, 20) = }") print(F"Torus: {surface_area_torus(20, 10) = }") print(F"Equilateral Triangle: {area_reg_polygon(3, 10) = }") print(F"Square: {area_reg_polygon(4, 10) = }") print(F"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
312
from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def __snake_case ( __UpperCamelCase : NDArray[floataa] ,__UpperCamelCase : NDArray[floataa] ,__UpperCamelCase : list[int] ,__UpperCamelCase : int ,): """simple docstring""" A_ , A_ = coefficient_matrix.shape A_ , A_ = constant_matrix.shape if rowsa != colsa: A_ = f'''Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}''' raise ValueError(__UpperCamelCase ) if colsa != 1: A_ = f'''Constant matrix must be nx1 but received {rowsa}x{colsa}''' raise ValueError(__UpperCamelCase ) if rowsa != rowsa: A_ = ( "Coefficient and constant matrices dimensions must be nxn and nx1 but " f'''received {rowsa}x{colsa} and {rowsa}x{colsa}''' ) raise ValueError(__UpperCamelCase ) if len(__UpperCamelCase ) != rowsa: A_ = ( "Number of initial values must be equal to number of rows in coefficient " f'''matrix but received {len(__UpperCamelCase )} and {rowsa}''' ) raise ValueError(__UpperCamelCase ) if iterations <= 0: raise ValueError("Iterations must be at least 1" ) A_ = np.concatenate( (coefficient_matrix, constant_matrix) ,axis=1 ) A_ , A_ = table.shape strictly_diagonally_dominant(__UpperCamelCase ) # Iterates the whole matrix for given number of times for _ in range(__UpperCamelCase ): A_ = [] for row in range(__UpperCamelCase ): A_ = 0 for col in range(__UpperCamelCase ): if col == row: A_ = table[row][col] elif col == cols - 1: A_ = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] A_ = (temp + val) / denom new_val.append(__UpperCamelCase ) A_ = new_val return [float(__UpperCamelCase ) for i in new_val] def __snake_case ( __UpperCamelCase : NDArray[floataa] ): """simple docstring""" A_ , A_ = table.shape A_ = True for i in range(0 ,__UpperCamelCase ): A_ = 0 for j in range(0 ,cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError("Coefficient matrix is not strictly diagonally dominant" ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
312
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class UpperCAmelCase_ ( metaclass=_lowercase): snake_case__ = ['''keras_nlp'''] def __init__( self : Any , *__UpperCamelCase : Union[str, Any] , **__UpperCamelCase : int ) -> str: requires_backends(self , ['''keras_nlp'''] )
54
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class UpperCAmelCase_ ( unittest.TestCase): snake_case__ = MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING snake_case__ = TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def _UpperCamelCase ( self : Optional[Any] , __UpperCamelCase : Optional[Any] , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : Optional[int] ) -> int: _UpperCamelCase = TextaTextGenerationPipeline(model=__UpperCamelCase , tokenizer=__UpperCamelCase ) return generator, ["Something to write", "Something else"] def _UpperCamelCase ( self : Tuple , __UpperCamelCase : str , __UpperCamelCase : List[Any] ) -> Union[str, Any]: _UpperCamelCase = generator('''Something there''' ) self.assertEqual(__UpperCamelCase , [{'''generated_text''': ANY(__UpperCamelCase )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]['''generated_text'''].startswith('''Something there''' ) ) _UpperCamelCase = generator(['''This is great !''', '''Something else'''] , num_return_sequences=2 , do_sample=__UpperCamelCase ) self.assertEqual( __UpperCamelCase , [ [{'''generated_text''': ANY(__UpperCamelCase )}, {'''generated_text''': ANY(__UpperCamelCase )}], [{'''generated_text''': ANY(__UpperCamelCase )}, {'''generated_text''': ANY(__UpperCamelCase )}], ] , ) _UpperCamelCase = generator( ['''This is great !''', '''Something else'''] , num_return_sequences=2 , batch_size=2 , do_sample=__UpperCamelCase ) self.assertEqual( __UpperCamelCase , [ [{'''generated_text''': ANY(__UpperCamelCase )}, {'''generated_text''': ANY(__UpperCamelCase )}], [{'''generated_text''': ANY(__UpperCamelCase )}, {'''generated_text''': ANY(__UpperCamelCase )}], ] , ) with self.assertRaises(__UpperCamelCase ): generator(4 ) @require_torch def _UpperCamelCase ( self : List[str] ) -> List[str]: _UpperCamelCase = pipeline('''text2text-generation''' , model='''patrickvonplaten/t5-tiny-random''' , framework='''pt''' ) # do_sample=False necessary for reproducibility _UpperCamelCase = generator('''Something there''' , do_sample=__UpperCamelCase ) self.assertEqual(__UpperCamelCase , [{'''generated_text''': ''''''}] ) _UpperCamelCase = 3 _UpperCamelCase = generator( '''Something there''' , num_return_sequences=__UpperCamelCase , num_beams=__UpperCamelCase , ) _UpperCamelCase = [ {'''generated_text''': '''Beide Beide Beide Beide Beide Beide Beide Beide Beide'''}, {'''generated_text''': '''Beide Beide Beide Beide Beide Beide Beide Beide'''}, {'''generated_text''': ''''''}, ] self.assertEqual(__UpperCamelCase , __UpperCamelCase ) _UpperCamelCase = generator('''This is a test''' , do_sample=__UpperCamelCase , num_return_sequences=2 , return_tensors=__UpperCamelCase ) self.assertEqual( __UpperCamelCase , [ {'''generated_token_ids''': ANY(torch.Tensor )}, {'''generated_token_ids''': ANY(torch.Tensor )}, ] , ) _UpperCamelCase = generator.model.config.eos_token_id _UpperCamelCase = '''<pad>''' _UpperCamelCase = generator( ['''This is a test''', '''This is a second test'''] , do_sample=__UpperCamelCase , num_return_sequences=2 , batch_size=2 , return_tensors=__UpperCamelCase , ) self.assertEqual( __UpperCamelCase , [ [ {'''generated_token_ids''': ANY(torch.Tensor )}, {'''generated_token_ids''': ANY(torch.Tensor )}, ], [ {'''generated_token_ids''': ANY(torch.Tensor )}, {'''generated_token_ids''': ANY(torch.Tensor )}, ], ] , ) @require_tf def _UpperCamelCase ( self : Optional[Any] ) -> List[Any]: _UpperCamelCase = pipeline('''text2text-generation''' , model='''patrickvonplaten/t5-tiny-random''' , framework='''tf''' ) # do_sample=False necessary for reproducibility _UpperCamelCase = generator('''Something there''' , do_sample=__UpperCamelCase ) self.assertEqual(__UpperCamelCase , [{'''generated_text''': ''''''}] )
54
1
"""simple docstring""" def lowercase ( A_ , A_ , A_ , A_ )-> List[Any]: '''simple docstring''' a : List[Any] = [False] * len(A_ ) a : int = [] queue.append(A_ ) a : int = True while queue: a : List[str] = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(A_ ) a : Dict = True a : Optional[int] = u return visited[t] def lowercase ( A_ , A_ , A_ )-> str: '''simple docstring''' a : int = [-1] * (len(A_ )) a : List[Any] = 0 while bfs(A_ , A_ , A_ , A_ ): a : Tuple = float("Inf" ) a : List[str] = sink while s != source: # Find the minimum value in select path a : List[Any] = min(A_ , graph[parent[s]][s] ) a : str = parent[s] max_flow += path_flow a : Optional[Any] = sink while v != source: a : List[Any] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow a : str = parent[v] return max_flow __lowercase = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] __lowercase , __lowercase = 0, 5 print(ford_fulkerson(graph, source, sink))
40
'''simple docstring''' import string from math import logaa def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' UpperCAmelCase__ = document.translate( str.maketrans("""""" , """""" , string.punctuation ) ).replace("""\n""" , """""" ) UpperCAmelCase__ = document_without_punctuation.split(""" """ ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : str ): '''simple docstring''' UpperCAmelCase__ = corpus.lower().translate( str.maketrans("""""" , """""" , string.punctuation ) ) # strip all punctuation and replace it with '' UpperCAmelCase__ = corpus_without_punctuation.split("""\n""" ) UpperCAmelCase__ = term.lower() return (len([doc for doc in docs if term in doc] ), len(SCREAMING_SNAKE_CASE__ )) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Tuple=False ): '''simple docstring''' if smoothing: if n == 0: raise ValueError("""log10(0) is undefined.""" ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError("""df must be > 0""" ) elif n == 0: raise ValueError("""log10(0) is undefined.""" ) return round(logaa(n / df ) , 3 ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' return round(tf * idf , 3 )
346
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices snake_case_ = logging.get_logger(__name__) snake_case_ = { 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , _UpperCAmelCase ): A_ : Dict = 'bit' A_ : Dict = ['preactivation', 'bottleneck'] A_ : Optional[Any] = ['SAME', 'VALID'] def __init__(self : Optional[Any] , a__ : List[str]=3 , a__ : Dict=64 , a__ : Optional[Any]=[256, 512, 1024, 2048] , a__ : Dict=[3, 4, 6, 3] , a__ : List[str]="preactivation" , a__ : str="relu" , a__ : Tuple=None , a__ : int=32 , a__ : str=0.0 , a__ : Optional[Any]=False , a__ : List[str]=32 , a__ : Dict=1 , a__ : Optional[Any]=None , a__ : str=None , **a__ : Optional[int] , ): """simple docstring""" super().__init__(**a__ ) if layer_type not in self.layer_types: raise ValueError(f"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) if global_padding is not None: if global_padding.upper() in self.supported_padding: __snake_case = global_padding.upper() else: raise ValueError(f"""Padding strategy {global_padding} not supported""" ) __snake_case = num_channels __snake_case = embedding_size __snake_case = hidden_sizes __snake_case = depths __snake_case = layer_type __snake_case = hidden_act __snake_case = global_padding __snake_case = num_groups __snake_case = drop_path_rate __snake_case = embedding_dynamic_padding __snake_case = output_stride __snake_case = width_factor __snake_case = ['''stem'''] + [f"""stage{idx}""" for idx in range(1 , len(a__ ) + 1 )] __snake_case , __snake_case = get_aligned_output_features_output_indices( out_features=a__ , out_indices=a__ , stage_names=self.stage_names )
238
from __future__ import annotations import collections import pprint from pathlib import Path def lowerCamelCase__ ( snake_case_ : str ) -> str: return "".join(sorted(snake_case_ ) ) def lowerCamelCase__ ( snake_case_ : str ) -> list[str]: return word_by_signature[signature(snake_case_ )] snake_case_ = Path(__file__).parent.joinpath('words.txt').read_text(encoding='utf-8') snake_case_ = sorted({word.strip().lower() for word in data.splitlines()}) snake_case_ = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": snake_case_ = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('anagrams.txt', 'w') as file: file.write('all_anagrams = \n ') file.write(pprint.pformat(all_anagrams))
238
1
from manim import * class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): def a (self : List[str] ): """simple docstring""" __snake_case = Rectangle(height=0.5 , width=0.5 ) __snake_case = Rectangle(height=0.2_5 , width=0.2_5 ) __snake_case = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) __snake_case = [mem.copy() for i in range(6 )] __snake_case = [mem.copy() for i in range(6 )] __snake_case = VGroup(*a__ ).arrange(a__ , buff=0 ) __snake_case = VGroup(*a__ ).arrange(a__ , buff=0 ) __snake_case = VGroup(a__ , a__ ).arrange(a__ , buff=0 ) __snake_case = Text('''CPU''' , font_size=24 ) __snake_case = Group(a__ , a__ ).arrange(a__ , buff=0.5 , aligned_edge=a__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(a__ ) __snake_case = [mem.copy() for i in range(4 )] __snake_case = VGroup(*a__ ).arrange(a__ , buff=0 ) __snake_case = Text('''GPU''' , font_size=24 ) __snake_case = Group(a__ , a__ ).arrange(a__ , buff=0.5 , aligned_edge=a__ ) gpu.move_to([-1, -1, 0] ) self.add(a__ ) __snake_case = [mem.copy() for i in range(6 )] __snake_case = VGroup(*a__ ).arrange(a__ , buff=0 ) __snake_case = Text('''Model''' , font_size=24 ) __snake_case = Group(a__ , a__ ).arrange(a__ , buff=0.5 , aligned_edge=a__ ) model.move_to([3, -1.0, 0] ) self.add(a__ ) __snake_case = [] __snake_case = [] __snake_case = [] for i, rect in enumerate(a__ ): rect.set_stroke(a__ ) __snake_case = Rectangle(height=0.4_6 / 4 , width=0.4_6 / 3 ).set_stroke(width=0.0 ).set_fill(a__ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=a__ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=a__ , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=a__ , buff=0.0 ) self.add(a__ ) model_cpu_arr.append(a__ ) self.add(*a__ , *a__ , *a__ ) __snake_case = [mem.copy() for i in range(6 )] __snake_case = VGroup(*a__ ).arrange(a__ , buff=0 ) __snake_case = Text('''Loaded Checkpoint''' , font_size=24 ) __snake_case = Group(a__ , a__ ).arrange(a__ , buff=0.5 , aligned_edge=a__ ) checkpoint.move_to([3, 0.5, 0] ) self.add(a__ ) __snake_case = [] __snake_case = [] for i, rect in enumerate(a__ ): __snake_case = fill.copy().set_fill(a__ , opacity=0.7 ) target.move_to(a__ ) ckpt_arr.append(a__ ) __snake_case = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(a__ ) self.add(*a__ , *a__ ) __snake_case = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) __snake_case = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(a__ , a__ ) __snake_case = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(a__ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(a__ ) __snake_case = MarkupText( f"""Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) __snake_case = [meta_mem.copy() for i in range(6 )] __snake_case = [meta_mem.copy() for i in range(6 )] __snake_case = VGroup(*a__ ).arrange(a__ , buff=0 ) __snake_case = VGroup(*a__ ).arrange(a__ , buff=0 ) __snake_case = VGroup(a__ , a__ ).arrange(a__ , buff=0 ) __snake_case = Text('''Disk''' , font_size=24 ) __snake_case = Group(a__ , a__ ).arrange(a__ , buff=0.5 , aligned_edge=a__ ) disk.move_to([-4.0, -1.2_5, 0] ) self.play(Write(a__ , run_time=3 ) , Write(a__ , run_time=1 ) , Create(a__ , run_time=1 ) ) __snake_case = [] for i, rect in enumerate(a__ ): __snake_case = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(a__ , run_time=1.5 ) ) self.play(*a__ ) self.play(FadeOut(a__ ) ) __snake_case = MarkupText(f"""Then, the checkpoint is removed from memory\nthrough garbage collection.""" , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(a__ , run_time=3 ) ) self.play( FadeOut(a__ , a__ , *a__ , *a__ ) , ) self.wait()
24
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A : str = { 'configuration_blenderbot': [ 'BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlenderbotConfig', 'BlenderbotOnnxConfig', ], 'tokenization_blenderbot': ['BlenderbotTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Optional[Any] = ['BlenderbotTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Optional[Any] = [ 'BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BlenderbotForCausalLM', 'BlenderbotForConditionalGeneration', 'BlenderbotModel', 'BlenderbotPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Dict = [ 'TFBlenderbotForConditionalGeneration', 'TFBlenderbotModel', 'TFBlenderbotPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Dict = [ 'FlaxBlenderbotForConditionalGeneration', 'FlaxBlenderbotModel', 'FlaxBlenderbotPreTrainedModel', ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys A : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
6
0
import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizerOutput from transformers.testing_utils import require_phonemizer from ...test_tokenization_common import TokenizerTesterMixin @require_phonemizer class lowerCamelCase (__lowerCamelCase , unittest.TestCase ): """simple docstring""" UpperCAmelCase_ = WavaVecaPhonemeCTCTokenizer UpperCAmelCase_ = False def A_ ( self : List[str] ) -> Dict: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ : Optional[Any] = ( "<s> <pad> </s> <unk> n s t ə l a i k d m ɛ ɾ e ɪ p o ɐ z ð f j v b ɹ ʁ ʊ iː r w ʌ u ɡ æ aɪ ʃ h ɔ ɑː " "ŋ ɚ eɪ β uː y ɑ̃ oʊ ᵻ eː θ aʊ ts oː ɔ̃ ɣ ɜ ɑ dʒ əl x ɜː ç ʒ tʃ ɔː ɑːɹ ɛ̃ ʎ ɔːɹ ʋ aː ɕ œ ø oːɹ ɲ yː " "ʔ iə i5 s. tɕ ?? nʲ ɛː œ̃ ɭ ɔø ʑ tʲ ɨ ɛɹ ts. rʲ ɪɹ ɭʲ i.5 ɔɪ q sʲ u5 ʊɹ iɜ a5 iɛ5 øː ʕ ja əɜ th ɑ5 " "oɪ dʲ ə5 tɕh ts.h mʲ ɯ dʑ vʲ e̞ tʃʲ ei5 o5 onɡ5 ɑu5 iɑ5 ai5 aɪɚ kh ə1 ʐ i2 ʉ ħ t[ aɪə ʲ ju ə2 u2 oɜ " "pː iɛɜ ou5 y5 uɜ tː uo5 d[ uoɜ tsh ɑɜ ɵ i̪5 uei5 ɟ aɜ ɑɨ i.ɜ eʊ o2 ɐ̃ ä pʲ kʲ n̩ ɒ ph ɑu2 uɨ əɪ ɫ ɬ " "yɜ bʲ ɑ2 s̪ aiɜ χ ɐ̃ʊ̃ 1 ə4 yæɜ a2 ɨː t̪ iouɜ ũ onɡɜ aɨ iɛ2 ɔɨ ɑuɜ o̞ ei2 iou2 c kː y2 ɖ oe dˤ yɛɜ " "əʊ S ɡʲ onɡ2 u\" eiɜ ʈ ɯᵝ iou5 dZ r̝̊ i.2 tS s^ ʝ yə5 iɑɜ uə5 pf ɨu iɑ2 ou2 ər2 fʲ ai2 r̝ uəɜ ɳ əɨ " "ua5 uɪ ɽ bː yu5 uo2 yɛ5 l̩ ɻ ərɜ ʂ i̪2 ouɜ uaɜ a. a.ː yæ5 dː r̩ ee ɪu ər5 i̪ ɜ æi u: i.ː t^ o1 ɪ^ " "ai ueiɜ æː ɛɪ eə i. ɴ ie ua2 ɑ1 o4 tʃː o: ɑ: u1 N i̪1 au yæ2 u. qː yəɜ y: kʰ tʃʰ iʊ sx õ uo tʰ " "uai5 bʰ u.ː uə2 ʊə d^ s̪ː yiɜ dʰ r. oe: i1 ɟː yu2 nʲʲ i̪4 uei2 tsʲ ɸ ĩ ɑ4 t̪ː eɑ u4 e: tsː ʈʰ ɡʰ " "ɯɯ dʒʲ ʂʲ X ɵː uaiɜ tɕʲ ã t^ː ẽː yɛ2 cː i.1 ɛʊ dˤdˤ dʒː i4 ɡː yi ɕʲ ɟʰ pʰ dʑʲ yuɜ ua1 ua4 æiː ɐɐ " "ui iou1 ʊː a1 iou4 cʰ iɛ1 yə2 ɖʰ ẽ ʒʲ ää ər4 iːː ɪː iɑ1 ər1 œː øi ɪuː cʰcʰ əː1 iː1 ũ kʰː o̞o̞ xʲ " "ou1 iɛ4 e̞e̞ y1 dzː dʲʲ dʰː ɯᵝɯᵝ lː uo1 i.4 i: yɛ5ʲ a4" ).split(" " ) SCREAMING_SNAKE_CASE__ : Any = dict(zip(_UpperCAmelCase, range(len(_UpperCAmelCase ) ) ) ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = {"pad_token": "<pad>", "unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>"} SCREAMING_SNAKE_CASE__ : Dict = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file, "w", encoding="utf-8" ) as fp: fp.write(json.dumps(_UpperCAmelCase ) + "\n" ) def A_ ( self : str, _UpperCAmelCase : Union[str, Any], _UpperCAmelCase : Optional[Any]=False, _UpperCAmelCase : str=2_0, _UpperCAmelCase : Tuple=5 ) -> Tuple[str, list]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = [(i, tokenizer.decode([i], clean_up_tokenization_spaces=_UpperCAmelCase )) for i in range(len(_UpperCAmelCase ) )] SCREAMING_SNAKE_CASE__ : List[Any] = list(filter(lambda _UpperCAmelCase : [t[0]] == tokenizer.encode(t[1], do_phonemize=_UpperCAmelCase ), _UpperCAmelCase ) ) if max_length is not None and len(_UpperCAmelCase ) > max_length: SCREAMING_SNAKE_CASE__ : Union[str, Any] = toks[:max_length] if min_length is not None and len(_UpperCAmelCase ) < min_length and len(_UpperCAmelCase ) > 0: while len(_UpperCAmelCase ) < min_length: SCREAMING_SNAKE_CASE__ : Dict = toks + toks # toks_str = [t[1] for t in toks] SCREAMING_SNAKE_CASE__ : Any = [t[0] for t in toks] # Ensure consistency SCREAMING_SNAKE_CASE__ : List[Any] = tokenizer.decode(_UpperCAmelCase, clean_up_tokenization_spaces=_UpperCAmelCase ) if " " not in output_txt and len(_UpperCAmelCase ) > 1: SCREAMING_SNAKE_CASE__ : Any = ( tokenizer.decode([toks_ids[0]], clean_up_tokenization_spaces=_UpperCAmelCase ) + " " + tokenizer.decode(toks_ids[1:], clean_up_tokenization_spaces=_UpperCAmelCase ) ) if with_prefix_space: SCREAMING_SNAKE_CASE__ : List[Any] = " " + output_txt SCREAMING_SNAKE_CASE__ : Optional[Any] = tokenizer.encode(_UpperCAmelCase, add_special_tokens=_UpperCAmelCase ) return output_txt, output_ids def A_ ( self : Any, **_UpperCAmelCase : Tuple ) -> int: """simple docstring""" kwargs.update(self.special_tokens_map ) return WavaVecaPhonemeCTCTokenizer.from_pretrained(self.tmpdirname, **_UpperCAmelCase ) def A_ ( self : int ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) # check adding a single token tokenizer.add_tokens("xxx" ) SCREAMING_SNAKE_CASE__ : Tuple = tokenizer("m xxx ɪ", do_phonemize=_UpperCAmelCase ).input_ids self.assertEqual(_UpperCAmelCase, [1_3, 3_9_2, 1_7] ) # xxx should be last token tokenizer.add_tokens(["aaa", "bbb", "ccc"] ) SCREAMING_SNAKE_CASE__ : str = tokenizer("m aaa ɪ ccc", do_phonemize=_UpperCAmelCase ).input_ids self.assertEqual(_UpperCAmelCase, [1_3, 3_9_3, 1_7, 3_9_5] ) # aaa and ccc should be after xxx and 2 after aaa SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer("maɪ c", do_phonemize=_UpperCAmelCase ).input_ids self.assertEqual(_UpperCAmelCase, [3, 2_0_0] ) # mai should be <unk> (=3) def A_ ( self : List[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) SCREAMING_SNAKE_CASE__ : str = "Hello how are you" SCREAMING_SNAKE_CASE__ : Any = tokenizer.phonemize(_UpperCAmelCase, phonemizer_lang="en-us" ) self.assertEqual(_UpperCAmelCase, "h ə l oʊ h aʊ ɑːɹ j uː" ) def A_ ( self : Any ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = "Hello how are you" SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer.phonemize(_UpperCAmelCase, phonemizer_lang="en-us" ) self.assertEqual(tokenizer(_UpperCAmelCase ).input_ids, tokenizer(_UpperCAmelCase, do_phonemize=_UpperCAmelCase ).input_ids ) def A_ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) SCREAMING_SNAKE_CASE__ : str = "Hello how are you" SCREAMING_SNAKE_CASE__ : int = tokenizer.phonemize(_UpperCAmelCase, phonemizer_lang="en-us" ) SCREAMING_SNAKE_CASE__ : str = tokenizer.decode(tokenizer(_UpperCAmelCase ).input_ids ) self.assertEqual(_UpperCAmelCase, _UpperCAmelCase ) def A_ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) SCREAMING_SNAKE_CASE__ : str = [ [1_1, 5, 1_5, tokenizer.pad_token_id, 1_5, 8, 9_8], [2_4, 2_2, 5, 2_4, 2_2, 5, 7_7], ] SCREAMING_SNAKE_CASE__ : Tuple = tokenizer.decode(sample_ids[0] ) SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer.batch_decode(_UpperCAmelCase ) self.assertEqual(_UpperCAmelCase, batch_tokens[0] ) self.assertEqual(_UpperCAmelCase, ["k s ɾ ɾ l ɭʲ", "j ð s j ð s oːɹ"] ) def A_ ( self : int ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft", word_delimiter_token="|" ) tokenizer.add_tokens("|" ) SCREAMING_SNAKE_CASE__ : Optional[Any] = "Hello how are you" SCREAMING_SNAKE_CASE__ : Tuple = tokenizer.phonemize(_UpperCAmelCase, phonemizer_lang="en-us" ) self.assertEqual(_UpperCAmelCase, "h ə l oʊ | h aʊ | ɑːɹ | j uː |" ) def A_ ( self : Dict ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft", word_delimiter_token="|" ) tokenizer.add_tokens("|" ) SCREAMING_SNAKE_CASE__ : List[Any] = "Hello how are you" SCREAMING_SNAKE_CASE__ : List[Any] = tokenizer.phonemize(_UpperCAmelCase, phonemizer_lang="en-us" ) self.assertEqual(tokenizer(_UpperCAmelCase ).input_ids, tokenizer(_UpperCAmelCase, do_phonemize=_UpperCAmelCase ).input_ids ) def A_ ( self : Union[str, Any] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft", word_delimiter_token="|" ) tokenizer.add_tokens("|" ) # fmt: off SCREAMING_SNAKE_CASE__ : Tuple = [ [1_1, 5, 1_5, tokenizer.pad_token_id, tokenizer.word_delimiter_token_id, 1_5, 8, tokenizer.word_delimiter_token_id, 9_8], [tokenizer.word_delimiter_token_id, 2_4, 2_2, tokenizer.word_delimiter_token_id, 5, 2_4, 2_2, 5, 7_7], ] # fmt: on # decode with word_del_token filter SCREAMING_SNAKE_CASE__ : int = tokenizer.decode(sample_ids[0] ) SCREAMING_SNAKE_CASE__ : int = tokenizer.batch_decode(_UpperCAmelCase ) self.assertEqual(_UpperCAmelCase, batch_tokens[0] ) self.assertEqual(_UpperCAmelCase, ["k s ɾ ɾ l ɭʲ", "j ð s j ð s oːɹ"] ) # decode with no word_del_token filter SCREAMING_SNAKE_CASE__ : Tuple = tokenizer.decode(sample_ids[0], filter_word_delimiter_token=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : str = tokenizer.batch_decode(_UpperCAmelCase, filter_word_delimiter_token=_UpperCAmelCase ) self.assertEqual(_UpperCAmelCase, batch_tokens[0] ) self.assertEqual(_UpperCAmelCase, ["k s ɾ | ɾ l | ɭʲ", "| j ð | s j ð s oːɹ"] ) def A_ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft", word_delimiter_token="|" ) tokenizer.add_tokens("|" ) SCREAMING_SNAKE_CASE__ : Optional[Any] = "Hello how are you" SCREAMING_SNAKE_CASE__ : int = tokenizer.phonemize(_UpperCAmelCase, phonemizer_lang="en-us" ) SCREAMING_SNAKE_CASE__ : List[str] = tokenizer.decode(tokenizer(_UpperCAmelCase ).input_ids, filter_word_delimiter_token=_UpperCAmelCase ) self.assertEqual(_UpperCAmelCase, _UpperCAmelCase ) def A_ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft", word_delimiter_token="|" ) tokenizer.add_tokens("|" ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = "Hello how are you" SCREAMING_SNAKE_CASE__ : Tuple = tokenizer.phonemize(_UpperCAmelCase, phonemizer_lang="en-us" ) SCREAMING_SNAKE_CASE__ : str = tokenizer.decode(tokenizer(_UpperCAmelCase ).input_ids, filter_word_delimiter_token=_UpperCAmelCase ) self.assertEqual(" ".join([p.strip() for p in phonemes.split(" |" )] ).strip(), _UpperCAmelCase ) def A_ ( self : Any ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft", word_delimiter_token=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Optional[int] = "Hello how are you" SCREAMING_SNAKE_CASE__ : List[str] = tokenizer(_UpperCAmelCase, phonemizer_lang="en-us" ).input_ids SCREAMING_SNAKE_CASE__ : str = tokenizer(_UpperCAmelCase, phonemizer_lang="fr-fr" ).input_ids self.assertNotEqual(_UpperCAmelCase, _UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Any = tokenizer.decode(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer.decode(_UpperCAmelCase ) self.assertEqual(_UpperCAmelCase, "h ə l oʊ h aʊ ɑːɹ j uː" ) self.assertEqual(_UpperCAmelCase, "ɛ l o h aʊ a ʁ j u" ) def A_ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) SCREAMING_SNAKE_CASE__ : Optional[int] = "Hello how Are you" SCREAMING_SNAKE_CASE__ : Any = "hello how are you" SCREAMING_SNAKE_CASE__ : Dict = tokenizer(_UpperCAmelCase ).input_ids SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer(_UpperCAmelCase ).input_ids self.assertEqual(_UpperCAmelCase, _UpperCAmelCase ) def A_ ( self : Optional[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) tokenizer.add_tokens(["!", "?"] ) tokenizer.add_special_tokens({"cls_token": "$$$"} ) # fmt: off SCREAMING_SNAKE_CASE__ : Union[str, Any] = [ [1_1, 5, 1_5, tokenizer.pad_token_id, 1_5, 8, 9_8, 3_9_2, 3_9_2, 3_9_3, 3_9_2, 3_9_2, 3_9_3, 3_9_4, 3_9_4], [2_4, 2_2, 5, 2_4, 2_2, 5, 7_7, tokenizer.pad_token_id, 3_9_4, 3_9_4], ] # fmt: on SCREAMING_SNAKE_CASE__ : Dict = tokenizer.batch_decode(_UpperCAmelCase ) self.assertEqual(_UpperCAmelCase, ["k s ɾ ɾ l ɭʲ!?!? $$$", "j ð s j ð s oːɹ $$$"] ) @staticmethod def A_ ( _UpperCAmelCase : Any, _UpperCAmelCase : List[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = [d[key] for d in offsets] return retrieved_list def A_ ( self : Optional[int] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = self.get_tokenizer(word_delimiter_token="|" ) tokenizer.add_tokens("|" ) # fmt: off # ksssɾɾ|ɾɾ<pad>ɾɾ|<pad>ɾlll|ɭʲ -> k s ɾ ɾ | ɾ l | ɭʲ" SCREAMING_SNAKE_CASE__ : Any = [1_1, 5, 5, 5, 1_5, 1_5, tokenizer.pad_token_id, 1_5, 1_5, tokenizer.word_delimiter_token_id, tokenizer.pad_token_id, 1_5, 8, 8, 8, tokenizer.word_delimiter_token_id, 9_8] # fmt: on SCREAMING_SNAKE_CASE__ : Any = tokenizer.decode(_UpperCAmelCase, output_char_offsets=_UpperCAmelCase, filter_word_delimiter_token=_UpperCAmelCase ) # check Wav2Vec2CTCTokenizerOutput keys for char self.assertEqual(len(outputs.keys() ), 2 ) self.assertTrue("text" in outputs ) self.assertTrue("char_offsets" in outputs ) self.assertTrue(isinstance(_UpperCAmelCase, _UpperCAmelCase ) ) # check that order of chars is correct and identical for both outputs self.assertEqual(" ".join(self.get_from_offsets(outputs["char_offsets"], "char" ) ), outputs.text ) self.assertListEqual( self.get_from_offsets(outputs["char_offsets"], "char" ), ["k", "s", "ɾ", "ɾ", "|", "ɾ", "l", "|", "ɭʲ"] ) # check that offsets are actually correct for char # 0-1 is 11, 1-4 is 5, 4-6 is first 15, 6-7 is <pad> (thus not shown), 7-9 is second 15, 9-10 is word_delimiter_token, # 10-11 is <pad> (thus not shown), 11-12 is third 15, 12-15 is 8, 15-16 is word_delimiter_token, 16-17 is 98 self.assertListEqual( self.get_from_offsets(outputs["char_offsets"], "start_offset" ), [0, 1, 4, 7, 9, 1_1, 1_2, 1_5, 1_6] ) self.assertListEqual( self.get_from_offsets(outputs["char_offsets"], "end_offset" ), [1, 4, 6, 9, 1_0, 1_2, 1_5, 1_6, 1_7] ) def A_ ( self : Tuple ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = self.get_tokenizer(word_delimiter_token="|" ) def check_list_tuples_equal(_UpperCAmelCase : Any, _UpperCAmelCase : Union[str, Any] ): self.assertTrue(isinstance(_UpperCAmelCase, _UpperCAmelCase ) ) self.assertTrue(isinstance(outputs_list[0], _UpperCAmelCase ) ) # transform list to ModelOutput SCREAMING_SNAKE_CASE__ : Tuple = WavaVecaPhonemeCTCTokenizerOutput( {k: [d[k] for d in outputs_list] for k in outputs_list[0]} ) self.assertListEqual(outputs_batch["text"], outputs_batch_a["text"] ) def recursive_check(_UpperCAmelCase : Optional[Any], _UpperCAmelCase : Union[str, Any] ): if isinstance(_UpperCAmelCase, _UpperCAmelCase ): [recursive_check(_UpperCAmelCase, _UpperCAmelCase ) for la, la in zip(_UpperCAmelCase, _UpperCAmelCase )] self.assertEqual(_UpperCAmelCase, _UpperCAmelCase ) if "char_offsets" in outputs_batch: recursive_check(outputs_batch["char_offsets"], outputs_batch_a["char_offsets"] ) # fmt: off SCREAMING_SNAKE_CASE__ : Dict = [ [1_1, 5, 1_5, tokenizer.pad_token_id, 1_5, 4, 8, 9_8, 3_2, 3_2, 3_2, 3_2, 4, 3_3, tokenizer.word_delimiter_token_id, 3_2, 3_2, 3_3, 3_4, 3_4], [2_4, 2_2, 5, tokenizer.word_delimiter_token_id, tokenizer.word_delimiter_token_id, 2_4, 2_2, 2_2, 2_2, 4, 5, 7_7, tokenizer.pad_token_id, 2_2, 2_2, 4, 3_4, 3_4, 3_4, 3_4], ] # fmt: on # We assume that `decode` works as expected. All we will check now is # the output type is correct and the output is identical to `decode` # char SCREAMING_SNAKE_CASE__ : List[Any] = tokenizer.batch_decode(_UpperCAmelCase, output_char_offsets=_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Optional[Any] = [tokenizer.decode(_UpperCAmelCase, output_char_offsets=_UpperCAmelCase ) for ids in sample_ids] check_list_tuples_equal(_UpperCAmelCase, _UpperCAmelCase ) @unittest.skip("Wav2Vec2PhonemeTokenizer always lower cases letters to correctly map to phonemes" ) def A_ ( self : Any ) -> int: """simple docstring""" pass @unittest.skip("Wav2Vec2PhonemeTokenizer always puts spaces between phonemes" ) def A_ ( self : Optional[int] ) -> Any: """simple docstring""" pass @unittest.skip("encodes to text to ids, but decodes ids to phonemes -> not possible to have internal consistency" ) def A_ ( self : str ) -> List[Any]: """simple docstring""" pass @unittest.skip("Wav2Vec2PhonemeModel has no max model length => no testing" ) def A_ ( self : Any ) -> List[str]: """simple docstring""" pass def A_ ( self : Tuple ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = self.get_tokenizers(do_lower_case=_UpperCAmelCase ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ : Optional[Any] = tokenizer.vocab_size SCREAMING_SNAKE_CASE__ : Union[str, Any] = len(_UpperCAmelCase ) self.assertNotEqual(_UpperCAmelCase, 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) SCREAMING_SNAKE_CASE__ : List[str] = ["aaaaa bbbbbb", "cccccccccdddddddd"] SCREAMING_SNAKE_CASE__ : int = tokenizer.add_tokens(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : List[Any] = tokenizer.vocab_size SCREAMING_SNAKE_CASE__ : List[str] = len(_UpperCAmelCase ) self.assertNotEqual(_UpperCAmelCase, 0 ) self.assertEqual(_UpperCAmelCase, _UpperCAmelCase ) self.assertEqual(_UpperCAmelCase, len(_UpperCAmelCase ) ) self.assertEqual(_UpperCAmelCase, all_size + len(_UpperCAmelCase ) ) SCREAMING_SNAKE_CASE__ : Dict = tokenizer.encode("aaaaa bbbbbb low cccccccccdddddddd l", add_special_tokens=_UpperCAmelCase ) self.assertGreaterEqual(len(_UpperCAmelCase ), 4 ) self.assertGreater(tokens[0], tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3], tokenizer.vocab_size - 1 ) SCREAMING_SNAKE_CASE__ : Optional[int] = {"eos_token": ">>>>|||<||<<|<<", "pad_token": "<<<<<|||>|>>>>|>"} SCREAMING_SNAKE_CASE__ : Any = tokenizer.add_special_tokens(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : str = tokenizer.vocab_size SCREAMING_SNAKE_CASE__ : Optional[Any] = len(_UpperCAmelCase ) self.assertNotEqual(_UpperCAmelCase, 0 ) self.assertEqual(_UpperCAmelCase, _UpperCAmelCase ) self.assertEqual(_UpperCAmelCase, len(_UpperCAmelCase ) ) self.assertEqual(_UpperCAmelCase, all_size_a + len(_UpperCAmelCase ) ) SCREAMING_SNAKE_CASE__ : List[str] = tokenizer.encode( ">>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l", add_special_tokens=_UpperCAmelCase ) self.assertGreaterEqual(len(_UpperCAmelCase ), 6 ) self.assertGreater(tokens[0], tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0], tokens[1] ) self.assertGreater(tokens[-3], tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3], tokens[-4] ) self.assertEqual(tokens[0], tokenizer.eos_token_id ) self.assertEqual(tokens[-3], tokenizer.pad_token_id ) @unittest.skip("The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode." ) def A_ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip("The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode." ) def A_ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" pass def A_ ( self : str ) -> Optional[Any]: """simple docstring""" # The default common tokenizer tests assumes that the output of `convert_tokens_to_string` is a string which # is not the case for Wav2Vec2PhonemeCTCTokenizer. SCREAMING_SNAKE_CASE__ : Optional[Any] = self.get_tokenizers(fast=_UpperCAmelCase, do_lower_case=_UpperCAmelCase ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ : Optional[int] = ["ð", "ɪ", "s", "ɪ", "z", "ɐ", "t", "ɛ", "k", "s", "t"] SCREAMING_SNAKE_CASE__ : Union[str, Any] = tokenizer.convert_tokens_to_string(_UpperCAmelCase ) self.assertIsInstance(output["text"], _UpperCAmelCase )
191
import requests from bsa import BeautifulSoup def _a ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : dict ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE__ , params=SCREAMING_SNAKE_CASE__ ).content , "html.parser" ) SCREAMING_SNAKE_CASE__ : str = soup.find("div" , attrs={"class": "gs_ri"} ) SCREAMING_SNAKE_CASE__ : int = div.find("div" , attrs={"class": "gs_fl"} ).find_all("a" ) return anchors[2].get_text() if __name__ == "__main__": _lowerCamelCase : Any = { '''title''': ( '''Precisely geometry controlled microsupercapacitors for ultrahigh areal ''' '''capacitance, volumetric capacitance, and energy density''' ), '''journal''': '''Chem. Mater.''', '''volume''': 3_0, '''pages''': '''3979-3990''', '''year''': 2_0_1_8, '''hl''': '''en''', } print(get_citation('''https://scholar.google.com/scholar_lookup''', params=params))
191
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : int = logging.get_logger(__name__) UpperCAmelCase_ : Optional[int] = { """vinvino02/glpn-kitti""": """https://huggingface.co/vinvino02/glpn-kitti/resolve/main/config.json""", # See all GLPN models at https://huggingface.co/models?filter=glpn } class lowerCAmelCase__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = "glpn" def __init__( self : Tuple , lowercase_ : Union[str, Any]=3 , lowercase_ : Tuple=4 , lowercase_ : Dict=[2, 2, 2, 2] , lowercase_ : str=[8, 4, 2, 1] , lowercase_ : Optional[Any]=[32, 64, 160, 256] , lowercase_ : Dict=[7, 3, 3, 3] , lowercase_ : List[Any]=[4, 2, 2, 2] , lowercase_ : List[Any]=[1, 2, 5, 8] , lowercase_ : str=[4, 4, 4, 4] , lowercase_ : Any="gelu" , lowercase_ : Union[str, Any]=0.0 , lowercase_ : Union[str, Any]=0.0 , lowercase_ : Union[str, Any]=0.02 , lowercase_ : Dict=0.1 , lowercase_ : Any=1e-6 , lowercase_ : Dict=64 , lowercase_ : Union[str, Any]=10 , lowercase_ : Dict=-1 , **lowercase_ : Dict , ): '''simple docstring''' super().__init__(**lowercase_) SCREAMING_SNAKE_CASE_ : Optional[int] = num_channels SCREAMING_SNAKE_CASE_ : str = num_encoder_blocks SCREAMING_SNAKE_CASE_ : Union[str, Any] = depths SCREAMING_SNAKE_CASE_ : Dict = sr_ratios SCREAMING_SNAKE_CASE_ : Optional[int] = hidden_sizes SCREAMING_SNAKE_CASE_ : Union[str, Any] = patch_sizes SCREAMING_SNAKE_CASE_ : List[Any] = strides SCREAMING_SNAKE_CASE_ : int = mlp_ratios SCREAMING_SNAKE_CASE_ : List[str] = num_attention_heads SCREAMING_SNAKE_CASE_ : List[str] = hidden_act SCREAMING_SNAKE_CASE_ : List[Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE_ : Optional[Any] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ : List[Any] = initializer_range SCREAMING_SNAKE_CASE_ : Union[str, Any] = drop_path_rate SCREAMING_SNAKE_CASE_ : Union[str, Any] = layer_norm_eps SCREAMING_SNAKE_CASE_ : Any = decoder_hidden_size SCREAMING_SNAKE_CASE_ : Dict = max_depth SCREAMING_SNAKE_CASE_ : Optional[Any] = head_in_index
91
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCAmelCase : List[str] = logging.get_logger(__name__) _lowerCAmelCase : Union[str, Any] = { "Salesforce/instruct-blip-flan-t5": "https://huggingface.co/Salesforce/instruct-blip-flan-t5/resolve/main/config.json", } class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = 'instructblip_vision_model' def __init__( self , __snake_case=1408 , __snake_case=6144 , __snake_case=39 , __snake_case=16 , __snake_case=224 , __snake_case=14 , __snake_case="gelu" , __snake_case=1e-6 , __snake_case=0.0 , __snake_case=1e-10 , __snake_case=True , **__snake_case , ) -> str: '''simple docstring''' super().__init__(**__snake_case ) __a =hidden_size __a =intermediate_size __a =num_hidden_layers __a =num_attention_heads __a =patch_size __a =image_size __a =initializer_range __a =attention_dropout __a =layer_norm_eps __a =hidden_act __a =qkv_bias @classmethod def __magic_name__ ( cls , __snake_case , **__snake_case ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(__snake_case ) __a , __a =cls.get_config_dict(__snake_case , **__snake_case ) # get the vision config dict if we are loading from InstructBlipConfig if config_dict.get('model_type' ) == "instructblip": __a =config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(__snake_case , **__snake_case ) class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = 'instructblip_qformer' def __init__( self , __snake_case=3_0522 , __snake_case=768 , __snake_case=12 , __snake_case=12 , __snake_case=3072 , __snake_case="gelu" , __snake_case=0.1 , __snake_case=0.1 , __snake_case=512 , __snake_case=0.02 , __snake_case=1e-12 , __snake_case=0 , __snake_case="absolute" , __snake_case=2 , __snake_case=1408 , **__snake_case , ) -> List[str]: '''simple docstring''' super().__init__(pad_token_id=__snake_case , **__snake_case ) __a =vocab_size __a =hidden_size __a =num_hidden_layers __a =num_attention_heads __a =hidden_act __a =intermediate_size __a =hidden_dropout_prob __a =attention_probs_dropout_prob __a =max_position_embeddings __a =initializer_range __a =layer_norm_eps __a =position_embedding_type __a =cross_attention_frequency __a =encoder_hidden_size @classmethod def __magic_name__ ( cls , __snake_case , **__snake_case ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(__snake_case ) __a , __a =cls.get_config_dict(__snake_case , **__snake_case ) # get the qformer config dict if we are loading from InstructBlipConfig if config_dict.get('model_type' ) == "instructblip": __a =config_dict['qformer_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(__snake_case , **__snake_case ) class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = 'instructblip' SCREAMING_SNAKE_CASE = True def __init__( self , __snake_case=None , __snake_case=None , __snake_case=None , __snake_case=32 , **__snake_case ) -> str: '''simple docstring''' super().__init__(**__snake_case ) if vision_config is None: __a ={} logger.info('vision_config is None. initializing the InstructBlipVisionConfig with default values.' ) if qformer_config is None: __a ={} logger.info('qformer_config is None. Initializing the InstructBlipQFormerConfig with default values.' ) if text_config is None: __a ={} logger.info('text_config is None. Initializing the text config with default values (`OPTConfig`).' ) __a =InstructBlipVisionConfig(**__snake_case ) __a =InstructBlipQFormerConfig(**__snake_case ) __a =text_config['model_type'] if 'model_type' in text_config else 'opt' __a =CONFIG_MAPPING[text_model_type](**__snake_case ) __a =self.text_config.tie_word_embeddings __a =self.text_config.is_encoder_decoder __a =num_query_tokens __a =self.vision_config.hidden_size __a =self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES __a =1.0 __a =0.02 @classmethod def __magic_name__ ( cls , __snake_case , __snake_case , __snake_case , **__snake_case , ) -> Optional[Any]: '''simple docstring''' return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **__snake_case , ) def __magic_name__ ( self ) -> List[Any]: '''simple docstring''' __a =copy.deepcopy(self.__dict__ ) __a =self.vision_config.to_dict() __a =self.qformer_config.to_dict() __a =self.text_config.to_dict() __a =self.__class__.model_type return output
218
0
'''simple docstring''' import math def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" if ( not isinstance(_lowerCAmelCase , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('power_factor must be a valid float value between -1 and 1.' ) return apparent_power * power_factor def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" if ( not isinstance(_lowerCAmelCase , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('power_factor must be a valid float value between -1 and 1.' ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
48
'''simple docstring''' from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCamelCase ( A ): '''simple docstring''' lowerCAmelCase__ = ["""image_processor""", """tokenizer"""] lowerCAmelCase__ = """BridgeTowerImageProcessor""" lowerCAmelCase__ = ("""RobertaTokenizer""", """RobertaTokenizerFast""") def __init__( self : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any]): '''simple docstring''' super().__init__(_lowerCAmelCase , _lowerCAmelCase) def __call__( self : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , _lowerCAmelCase : bool = True , _lowerCAmelCase : Union[bool, str, PaddingStrategy] = False , _lowerCAmelCase : Union[bool, str, TruncationStrategy] = None , _lowerCAmelCase : Optional[int] = None , _lowerCAmelCase : int = 0 , _lowerCAmelCase : Optional[int] = None , _lowerCAmelCase : Optional[bool] = None , _lowerCAmelCase : Optional[bool] = None , _lowerCAmelCase : bool = False , _lowerCAmelCase : bool = False , _lowerCAmelCase : bool = False , _lowerCAmelCase : bool = False , _lowerCAmelCase : bool = True , _lowerCAmelCase : Optional[Union[str, TensorType]] = None , **_lowerCAmelCase : Optional[Any] , ): '''simple docstring''' __lowercase =self.tokenizer( text=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=_lowerCAmelCase , stride=_lowerCAmelCase , pad_to_multiple_of=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , return_overflowing_tokens=_lowerCAmelCase , return_special_tokens_mask=_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , return_length=_lowerCAmelCase , verbose=_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase , ) # add pixel_values + pixel_mask __lowercase =self.image_processor( _lowerCAmelCase , return_tensors=_lowerCAmelCase , do_normalize=_lowerCAmelCase , do_center_crop=_lowerCAmelCase , **_lowerCAmelCase) encoding.update(_lowerCAmelCase) return encoding def __lowerCamelCase ( self : List[str] , *_lowerCAmelCase : Dict , **_lowerCAmelCase : str): '''simple docstring''' return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase) def __lowerCamelCase ( self : Optional[Any] , *_lowerCAmelCase : Tuple , **_lowerCAmelCase : Union[str, Any]): '''simple docstring''' return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase) @property def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =self.tokenizer.model_input_names __lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names))
48
1
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): A_ : str = (PNDMScheduler,) A_ : Dict = (('num_inference_steps', 50),) def a (self : Dict , **a__ : Tuple ): """simple docstring""" __snake_case = { """num_train_timesteps""": 1000, """beta_start""": 0.0_0_0_1, """beta_end""": 0.0_2, """beta_schedule""": """linear""", } config.update(**a__ ) return config def a (self : List[str] , a__ : str=0 , **a__ : Tuple ): """simple docstring""" __snake_case = dict(self.forward_default_kwargs ) __snake_case = kwargs.pop('''num_inference_steps''' , a__ ) __snake_case = self.dummy_sample __snake_case = 0.1 * sample __snake_case = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: __snake_case = self.get_scheduler_config(**a__ ) __snake_case = scheduler_class(**a__ ) scheduler.set_timesteps(a__ ) # copy over dummy past residuals __snake_case = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(a__ ) __snake_case = scheduler_class.from_pretrained(a__ ) new_scheduler.set_timesteps(a__ ) # copy over dummy past residuals __snake_case = dummy_past_residuals[:] __snake_case = scheduler.step_prk(a__ , a__ , a__ , **a__ ).prev_sample __snake_case = new_scheduler.step_prk(a__ , a__ , a__ , **a__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" __snake_case = scheduler.step_plms(a__ , a__ , a__ , **a__ ).prev_sample __snake_case = new_scheduler.step_plms(a__ , a__ , a__ , **a__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def a (self : List[Any] ): """simple docstring""" pass def a (self : Dict , a__ : int=0 , **a__ : Union[str, Any] ): """simple docstring""" __snake_case = dict(self.forward_default_kwargs ) __snake_case = kwargs.pop('''num_inference_steps''' , a__ ) __snake_case = self.dummy_sample __snake_case = 0.1 * sample __snake_case = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a__ ) scheduler.set_timesteps(a__ ) # copy over dummy past residuals (must be after setting timesteps) __snake_case = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(a__ ) __snake_case = scheduler_class.from_pretrained(a__ ) # copy over dummy past residuals new_scheduler.set_timesteps(a__ ) # copy over dummy past residual (must be after setting timesteps) __snake_case = dummy_past_residuals[:] __snake_case = scheduler.step_prk(a__ , a__ , a__ , **a__ ).prev_sample __snake_case = new_scheduler.step_prk(a__ , a__ , a__ , **a__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" __snake_case = scheduler.step_plms(a__ , a__ , a__ , **a__ ).prev_sample __snake_case = new_scheduler.step_plms(a__ , a__ , a__ , **a__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def a (self : Optional[Any] , **a__ : Tuple ): """simple docstring""" __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config(**a__ ) __snake_case = scheduler_class(**a__ ) __snake_case = 10 __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter scheduler.set_timesteps(a__ ) for i, t in enumerate(scheduler.prk_timesteps ): __snake_case = model(a__ , a__ ) __snake_case = scheduler.step_prk(a__ , a__ , a__ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): __snake_case = model(a__ , a__ ) __snake_case = scheduler.step_plms(a__ , a__ , a__ ).prev_sample return sample def a (self : int ): """simple docstring""" __snake_case = dict(self.forward_default_kwargs ) __snake_case = kwargs.pop('''num_inference_steps''' , a__ ) for scheduler_class in self.scheduler_classes: __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a__ ) __snake_case = self.dummy_sample __snake_case = 0.1 * sample if num_inference_steps is not None and hasattr(a__ , '''set_timesteps''' ): scheduler.set_timesteps(a__ ) elif num_inference_steps is not None and not hasattr(a__ , '''set_timesteps''' ): __snake_case = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) __snake_case = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] __snake_case = dummy_past_residuals[:] __snake_case = scheduler.step_prk(a__ , 0 , a__ , **a__ ).prev_sample __snake_case = scheduler.step_prk(a__ , 1 , a__ , **a__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) __snake_case = scheduler.step_plms(a__ , 0 , a__ , **a__ ).prev_sample __snake_case = scheduler.step_plms(a__ , 1 , a__ , **a__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def a (self : List[str] ): """simple docstring""" for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=a__ ) def a (self : Any ): """simple docstring""" for steps_offset in [0, 1]: self.check_over_configs(steps_offset=a__ ) __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config(steps_offset=1 ) __snake_case = scheduler_class(**a__ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def a (self : str ): """simple docstring""" for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1] , [0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=a__ , beta_end=a__ ) def a (self : Optional[int] ): """simple docstring""" for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a__ ) def a (self : Union[str, Any] ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=a__ ) def a (self : Dict ): """simple docstring""" for t in [1, 5, 10]: self.check_over_forward(time_step=a__ ) def a (self : Dict ): """simple docstring""" for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=a__ ) def a (self : int ): """simple docstring""" __snake_case = 27 for scheduler_class in self.scheduler_classes: __snake_case = self.dummy_sample __snake_case = 0.1 * sample __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a__ ) scheduler.set_timesteps(a__ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): __snake_case = scheduler.step_prk(a__ , a__ , a__ ).prev_sample def a (self : Any ): """simple docstring""" with self.assertRaises(a__ ): __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a__ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def a (self : Dict ): """simple docstring""" __snake_case = self.full_loop() __snake_case = torch.sum(torch.abs(a__ ) ) __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_sum.item() - 198.1318 ) < 1E-2 assert abs(result_mean.item() - 0.2_5_8_0 ) < 1E-3 def a (self : str ): """simple docstring""" __snake_case = self.full_loop(prediction_type='''v_prediction''' ) __snake_case = torch.sum(torch.abs(a__ ) ) __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_sum.item() - 67.3986 ) < 1E-2 assert abs(result_mean.item() - 0.0_8_7_8 ) < 1E-3 def a (self : Optional[int] ): """simple docstring""" __snake_case = self.full_loop(set_alpha_to_one=a__ , beta_start=0.0_1 ) __snake_case = torch.sum(torch.abs(a__ ) ) __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_sum.item() - 230.0399 ) < 1E-2 assert abs(result_mean.item() - 0.2_9_9_5 ) < 1E-3 def a (self : Dict ): """simple docstring""" __snake_case = self.full_loop(set_alpha_to_one=a__ , beta_start=0.0_1 ) __snake_case = torch.sum(torch.abs(a__ ) ) __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_sum.item() - 186.9482 ) < 1E-2 assert abs(result_mean.item() - 0.2_4_3_4 ) < 1E-3
24
'''simple docstring''' import inspect from typing import Callable, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import DiffusionPipeline from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import logging a : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCamelCase_ ( __magic_name__ ): def __init__( self , A , A , A , A , A , A , A , ) -> Optional[Any]: super().__init__() self.register_modules( vae=A , text_encoder=A , tokenizer=A , unet=A , scheduler=A , safety_checker=A , feature_extractor=A , ) def _lowercase( self , A = "auto" ) -> List[Any]: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory UpperCAmelCase : Optional[Any] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(A ) def _lowercase( self ) -> Dict: self.enable_attention_slicing(A ) @torch.no_grad() def __call__( self , A , A = 512 , A = 512 , A = 50 , A = 7.5 , A = None , A = 1 , A = 0.0 , A = None , A = None , A = "pil" , A = True , A = None , A = 1 , A = None , **A , ) -> List[Any]: if isinstance(A , A ): UpperCAmelCase : List[str] = 1 elif isinstance(A , A ): UpperCAmelCase : Dict = len(A ) else: raise ValueError(f'''`prompt` has to be of type `str` or `list` but is {type(A )}''' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` have to be divisible by 8 but are {height} and {width}.''' ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(A , A ) or callback_steps <= 0) ): raise ValueError( f'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' f''' {type(A )}.''' ) # get prompt text embeddings UpperCAmelCase : List[str] = self.tokenizer( A , padding="""max_length""" , max_length=self.tokenizer.model_max_length , return_tensors="""pt""" , ) UpperCAmelCase : List[Any] = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: UpperCAmelCase : int = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( """The following part of your input was truncated because CLIP can only handle sequences up to""" f''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) UpperCAmelCase : Tuple = text_input_ids[:, : self.tokenizer.model_max_length] if text_embeddings is None: UpperCAmelCase : Union[str, Any] = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Union[str, Any] = text_embeddings.shape UpperCAmelCase : List[str] = text_embeddings.repeat(1 , A , 1 ) UpperCAmelCase : List[Any] = text_embeddings.view(bs_embed * num_images_per_prompt , A , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. UpperCAmelCase : Optional[int] = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: UpperCAmelCase : List[str] if negative_prompt is None: UpperCAmelCase : Any = [""""""] elif type(A ) is not type(A ): raise TypeError( f'''`negative_prompt` should be the same type to `prompt`, but got {type(A )} !=''' f''' {type(A )}.''' ) elif isinstance(A , A ): UpperCAmelCase : Optional[int] = [negative_prompt] elif batch_size != len(A ): raise ValueError( f'''`negative_prompt`: {negative_prompt} has batch size {len(A )}, but `prompt`:''' f''' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches''' """ the batch size of `prompt`.""" ) else: UpperCAmelCase : Any = negative_prompt UpperCAmelCase : Dict = text_input_ids.shape[-1] UpperCAmelCase : List[Any] = self.tokenizer( A , padding="""max_length""" , max_length=A , truncation=A , return_tensors="""pt""" , ) UpperCAmelCase : Tuple = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method UpperCAmelCase : int = uncond_embeddings.shape[1] UpperCAmelCase : List[Any] = uncond_embeddings.repeat(A , A , 1 ) UpperCAmelCase : List[str] = uncond_embeddings.view(batch_size * num_images_per_prompt , A , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCAmelCase : List[str] = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. UpperCAmelCase : Tuple = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) UpperCAmelCase : Optional[int] = (batch_size * num_images_per_prompt, self.unet.config.in_channels, 64, 64) UpperCAmelCase : str = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps UpperCAmelCase : Dict = torch.randn( A , generator=A , device="""cpu""" , dtype=A ).to(self.device ) UpperCAmelCase : int = torch.randn(A , generator=A , device="""cpu""" , dtype=A ).to( self.device ) else: UpperCAmelCase : int = torch.randn( A , generator=A , device=self.device , dtype=A ) UpperCAmelCase : int = torch.randn(A , generator=A , device=self.device , dtype=A ) else: if latents_reference.shape != latents_shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) UpperCAmelCase : Optional[Any] = latents_reference.to(self.device ) UpperCAmelCase : Tuple = latents.to(self.device ) # This is the key part of the pipeline where we # try to ensure that the generated images w/ the same seed # but different sizes actually result in similar images UpperCAmelCase : int = (latents_shape[3] - latents_shape_reference[3]) // 2 UpperCAmelCase : List[str] = (latents_shape[2] - latents_shape_reference[2]) // 2 UpperCAmelCase : Union[str, Any] = latents_shape_reference[3] if dx >= 0 else latents_shape_reference[3] + 2 * dx UpperCAmelCase : Union[str, Any] = latents_shape_reference[2] if dy >= 0 else latents_shape_reference[2] + 2 * dy UpperCAmelCase : Optional[int] = 0 if dx < 0 else dx UpperCAmelCase : List[str] = 0 if dy < 0 else dy UpperCAmelCase : Union[str, Any] = max(-dx , 0 ) UpperCAmelCase : List[Any] = max(-dy , 0 ) # import pdb # pdb.set_trace() UpperCAmelCase : str = latents_reference[:, :, dy : dy + h, dx : dx + w] # set timesteps self.scheduler.set_timesteps(A ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand UpperCAmelCase : Union[str, Any] = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler UpperCAmelCase : Optional[int] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] UpperCAmelCase : int = """eta""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) UpperCAmelCase : Optional[Any] = {} if accepts_eta: UpperCAmelCase : List[str] = eta for i, t in enumerate(self.progress_bar(A ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase : Optional[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase : str = self.scheduler.scale_model_input(A , A ) # predict the noise residual UpperCAmelCase : Any = self.unet(A , A , encoder_hidden_states=A ).sample # perform guidance if do_classifier_free_guidance: UpperCAmelCase , UpperCAmelCase : Any = noise_pred.chunk(2 ) UpperCAmelCase : List[Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase : Dict = self.scheduler.step(A , A , A , **A ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(A , A , A ) UpperCAmelCase : Union[str, Any] = 1 / 0.1_8_2_1_5 * latents UpperCAmelCase : Tuple = self.vae.decode(A ).sample UpperCAmelCase : Union[str, Any] = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 UpperCAmelCase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if self.safety_checker is not None: UpperCAmelCase : int = self.feature_extractor(self.numpy_to_pil(A ) , return_tensors="""pt""" ).to( self.device ) UpperCAmelCase , UpperCAmelCase : int = self.safety_checker( images=A , clip_input=safety_checker_input.pixel_values.to(text_embeddings.dtype ) ) else: UpperCAmelCase : Any = None if output_type == "pil": UpperCAmelCase : int = self.numpy_to_pil(A ) if not return_dict: return (image, has_nsfw_concept) return StableDiffusionPipelineOutput(images=A , nsfw_content_detected=A )
265
0
import torch from diffusers import StableDiffusionPipeline UpperCamelCase_ = '''path-to-your-trained-model''' UpperCamelCase_ = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to('''cuda''') UpperCamelCase_ = '''A photo of sks dog in a bucket''' UpperCamelCase_ = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save('''dog-bucket.png''')
362
import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() UpperCamelCase_ = logging.get_logger('''transformers.models.speecht5''') def lowerCamelCase_ ( _a : str , _a : int , _a : Union[str, Any] ): '''simple docstring''' hf_model.apply_weight_norm() UpperCAmelCase_ : Optional[int] = checkpoint["""input_conv.weight_g"""] UpperCAmelCase_ : str = checkpoint["""input_conv.weight_v"""] UpperCAmelCase_ : str = checkpoint["""input_conv.bias"""] for i in range(len(config.upsample_rates ) ): UpperCAmelCase_ : Dict = checkpoint[F'''upsamples.{i}.1.weight_g'''] UpperCAmelCase_ : Any = checkpoint[F'''upsamples.{i}.1.weight_v'''] UpperCAmelCase_ : Union[str, Any] = checkpoint[F'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): UpperCAmelCase_ : Tuple = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_g'''] UpperCAmelCase_ : Dict = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_v'''] UpperCAmelCase_ : Optional[Any] = checkpoint[F'''blocks.{i}.convs1.{j}.1.bias'''] UpperCAmelCase_ : Tuple = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_g'''] UpperCAmelCase_ : Optional[Any] = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_v'''] UpperCAmelCase_ : Tuple = checkpoint[F'''blocks.{i}.convs2.{j}.1.bias'''] UpperCAmelCase_ : Optional[int] = checkpoint["""output_conv.1.weight_g"""] UpperCAmelCase_ : Optional[Any] = checkpoint["""output_conv.1.weight_v"""] UpperCAmelCase_ : Union[str, Any] = checkpoint["""output_conv.1.bias"""] hf_model.remove_weight_norm() @torch.no_grad() def lowerCamelCase_ ( _a : Tuple , _a : int , _a : Any , _a : Tuple=None , _a : Dict=None , ): '''simple docstring''' if config_path is not None: UpperCAmelCase_ : Union[str, Any] = SpeechTaHifiGanConfig.from_pretrained(_a ) else: UpperCAmelCase_ : str = SpeechTaHifiGanConfig() UpperCAmelCase_ : List[str] = SpeechTaHifiGan(_a ) UpperCAmelCase_ : int = torch.load(_a ) load_weights(orig_checkpoint["""model"""]["""generator"""] , _a , _a ) UpperCAmelCase_ : List[Any] = np.load(_a ) UpperCAmelCase_ : Optional[Any] = stats[0].reshape(-1 ) UpperCAmelCase_ : int = stats[1].reshape(-1 ) UpperCAmelCase_ : Any = torch.from_numpy(_a ).float() UpperCAmelCase_ : int = torch.from_numpy(_a ).float() model.save_pretrained(_a ) if repo_id: print("""Pushing to the hub...""" ) model.push_to_hub(_a ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to original checkpoint''') parser.add_argument('''--stats_path''', required=True, default=None, type=str, help='''Path to stats.npy file''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) UpperCamelCase_ = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
59
0
'''simple docstring''' import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def __lowerCamelCase ( A__ ) -> tuple: """simple docstring""" return (data["data"], data["target"]) def __lowerCamelCase ( A__ , A__ ) -> XGBClassifier: """simple docstring""" UpperCamelCase = XGBClassifier() classifier.fit(A__ , A__ ) return classifier def __lowerCamelCase ( ) -> None: """simple docstring""" UpperCamelCase = load_iris() UpperCamelCase , UpperCamelCase = data_handling(A__ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = train_test_split( A__ , A__ , test_size=0.25 ) UpperCamelCase = iris['target_names'] # Create an XGBoost Classifier from the training data UpperCamelCase = xgboost(A__ , A__ ) # Display the confusion matrix of the classifier with both training and test sets ConfusionMatrixDisplay.from_estimator( A__ , A__ , A__ , display_labels=A__ , cmap='Blues' , normalize='true' , ) plt.title('Normalized Confusion Matrix - IRIS Dataset' ) plt.show() if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
28
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def a ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): '''simple docstring''' lowercase__ = ('''dense.weight''', '''attention.self.query''', '''attention.self.key''', '''attention.self.value''') lowercase__ = ( ('''layer.''', '''layer_'''), ('''word_embeddings.weight''', '''word_embeddings'''), ('''position_embeddings.weight''', '''position_embeddings'''), ('''token_type_embeddings.weight''', '''token_type_embeddings'''), ('''.''', '''/'''), ('''LayerNorm/weight''', '''LayerNorm/gamma'''), ('''LayerNorm/bias''', '''LayerNorm/beta'''), ('''weight''', '''kernel'''), ) if not os.path.isdir(lowerCamelCase_ ): os.makedirs(lowerCamelCase_ ) lowercase__ = model.state_dict() def to_tf_var_name(lowerCamelCase_ ): for patt, repl in iter(lowerCamelCase_ ): lowercase__ = name.replace(lowerCamelCase_ , lowerCamelCase_ ) return F"""bert/{name}""" def create_tf_var(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): lowercase__ = tf.dtypes.as_dtype(tensor.dtype ) lowercase__ = tf.get_variable(dtype=lowerCamelCase_ , shape=tensor.shape , name=lowerCamelCase_ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(lowerCamelCase_ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: lowercase__ = to_tf_var_name(lowerCamelCase_ ) lowercase__ = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): lowercase__ = torch_tensor.T lowercase__ = create_tf_var(tensor=lowerCamelCase_ , name=lowerCamelCase_ , session=lowerCamelCase_ ) tf.keras.backend.set_value(lowerCamelCase_ , lowerCamelCase_ ) lowercase__ = session.run(lowerCamelCase_ ) print(F"""Successfully created {tf_name}: {np.allclose(lowerCamelCase_ , lowerCamelCase_ )}""" ) lowercase__ = tf.train.Saver(tf.trainable_variables() ) saver.save(lowerCamelCase_ , os.path.join(lowerCamelCase_ , model_name.replace('''-''' , '''_''' ) + '''.ckpt''' ) ) def a ( lowerCamelCase_=None ): '''simple docstring''' lowercase__ = argparse.ArgumentParser() parser.add_argument('''--model_name''' , type=lowerCamelCase_ , required=lowerCamelCase_ , help='''model name e.g. bert-base-uncased''' ) parser.add_argument( '''--cache_dir''' , type=lowerCamelCase_ , default=lowerCamelCase_ , required=lowerCamelCase_ , help='''Directory containing pytorch model''' ) parser.add_argument('''--pytorch_model_path''' , type=lowerCamelCase_ , required=lowerCamelCase_ , help='''/path/to/<pytorch-model-name>.bin''' ) parser.add_argument('''--tf_cache_dir''' , type=lowerCamelCase_ , required=lowerCamelCase_ , help='''Directory in which to save tensorflow model''' ) lowercase__ = parser.parse_args(lowerCamelCase_ ) lowercase__ = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=lowerCamelCase_ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
207
0
import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class lowercase__ : def __init__( self : Union[str, Any] , UpperCamelCase__ : Dict , UpperCamelCase__ : Any=13 , UpperCamelCase__ : Optional[int]=30 , UpperCamelCase__ : Optional[int]=2 , UpperCamelCase__ : Any=3 , UpperCamelCase__ : List[str]=True , UpperCamelCase__ : List[Any]=True , UpperCamelCase__ : Tuple=32 , UpperCamelCase__ : Any=5 , UpperCamelCase__ : int=4 , UpperCamelCase__ : Union[str, Any]=37 , UpperCamelCase__ : Tuple="gelu" , UpperCamelCase__ : List[str]=0.1 , UpperCamelCase__ : Dict=0.1 , UpperCamelCase__ : str=10 , UpperCamelCase__ : Tuple=0.02 , UpperCamelCase__ : Dict=3 , UpperCamelCase__ : Any=None , UpperCamelCase__ : Optional[Any]=2 , ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = parent SCREAMING_SNAKE_CASE : str = batch_size SCREAMING_SNAKE_CASE : Any = image_size SCREAMING_SNAKE_CASE : List[Any] = patch_size SCREAMING_SNAKE_CASE : List[str] = num_channels SCREAMING_SNAKE_CASE : int = is_training SCREAMING_SNAKE_CASE : Dict = use_labels SCREAMING_SNAKE_CASE : Dict = hidden_size SCREAMING_SNAKE_CASE : Tuple = num_hidden_layers SCREAMING_SNAKE_CASE : List[Any] = num_attention_heads SCREAMING_SNAKE_CASE : Dict = intermediate_size SCREAMING_SNAKE_CASE : int = hidden_act SCREAMING_SNAKE_CASE : Tuple = hidden_dropout_prob SCREAMING_SNAKE_CASE : Tuple = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : Any = type_sequence_label_size SCREAMING_SNAKE_CASE : Union[str, Any] = initializer_range SCREAMING_SNAKE_CASE : Optional[int] = scope SCREAMING_SNAKE_CASE : Optional[int] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) SCREAMING_SNAKE_CASE : Dict = (image_size // patch_size) ** 2 SCREAMING_SNAKE_CASE : Tuple = num_patches + 2 def __A ( self : int ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE : Optional[Any] = None if self.use_labels: SCREAMING_SNAKE_CASE : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE : List[Any] = self.get_config() return config, pixel_values, labels def __A ( self : Optional[Any] ): '''simple docstring''' return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase_ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __A ( self : int , UpperCamelCase__ : Any , UpperCamelCase__ : int , UpperCamelCase__ : Any ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = DeiTModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() SCREAMING_SNAKE_CASE : Tuple = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __A ( self : Tuple , UpperCamelCase__ : Tuple , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = DeiTForMaskedImageModeling(config=lowercase_ ) model.to(lowercase_ ) model.eval() SCREAMING_SNAKE_CASE : Optional[Any] = model(lowercase_ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images SCREAMING_SNAKE_CASE : Union[str, Any] = 1 SCREAMING_SNAKE_CASE : Dict = DeiTForMaskedImageModeling(lowercase_ ) model.to(lowercase_ ) model.eval() SCREAMING_SNAKE_CASE : int = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE : int = model(lowercase_ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __A ( self : Optional[int] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Union[str, Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = self.type_sequence_label_size SCREAMING_SNAKE_CASE : Any = DeiTForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() SCREAMING_SNAKE_CASE : Optional[int] = model(lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images SCREAMING_SNAKE_CASE : List[Any] = 1 SCREAMING_SNAKE_CASE : Optional[Any] = DeiTForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() SCREAMING_SNAKE_CASE : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE : int = model(lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __A ( self : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = self.prepare_config_and_inputs() ( SCREAMING_SNAKE_CASE ) : Any = config_and_inputs SCREAMING_SNAKE_CASE : Dict = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase__ ( a_ , a_ , unittest.TestCase): UpperCamelCase_ = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) UpperCamelCase_ = ( { """feature-extraction""": DeiTModel, """image-classification""": (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False def __A ( self : Optional[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = DeiTModelTester(self ) SCREAMING_SNAKE_CASE : str = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ , hidden_size=37 ) def __A ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''DeiT does not use inputs_embeds''' ) def __A ( self : Union[str, Any] ): '''simple docstring''' pass def __A ( self : List[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : List[Any] = model_class(lowercase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) SCREAMING_SNAKE_CASE : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase_ , nn.Linear ) ) def __A ( self : str ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : Any = model_class(lowercase_ ) SCREAMING_SNAKE_CASE : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE : Optional[Any] = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowercase_ ) def __A ( self : Tuple ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def __A ( self : Tuple ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowercase_ ) def __A ( self : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) def __A ( self : Tuple , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Dict=False ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = super()._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def __A ( self : Any ): '''simple docstring''' if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE : Dict = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowercase_ ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue SCREAMING_SNAKE_CASE : Optional[Any] = model_class(lowercase_ ) model.to(lowercase_ ) model.train() SCREAMING_SNAKE_CASE : Dict = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) SCREAMING_SNAKE_CASE : Optional[int] = model(**lowercase_ ).loss loss.backward() def __A ( self : List[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : List[Any] = True for model_class in self.all_model_classes: if model_class in get_values(lowercase_ ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue SCREAMING_SNAKE_CASE : List[Any] = model_class(lowercase_ ) model.gradient_checkpointing_enable() model.to(lowercase_ ) model.train() SCREAMING_SNAKE_CASE : Optional[int] = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) SCREAMING_SNAKE_CASE : Optional[int] = model(**lowercase_ ).loss loss.backward() def __A ( self : Tuple ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE : int = [ {'''title''': '''multi_label_classification''', '''num_labels''': 2, '''dtype''': torch.float}, {'''title''': '''single_label_classification''', '''num_labels''': 1, '''dtype''': torch.long}, {'''title''': '''regression''', '''num_labels''': 1, '''dtype''': torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowercase_ ), *get_values(lowercase_ ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f"""Testing {model_class} with {problem_type['title']}""" ): SCREAMING_SNAKE_CASE : str = problem_type['''title'''] SCREAMING_SNAKE_CASE : Tuple = problem_type['''num_labels'''] SCREAMING_SNAKE_CASE : Any = model_class(lowercase_ ) model.to(lowercase_ ) model.train() SCREAMING_SNAKE_CASE : Optional[Any] = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) if problem_type["num_labels"] > 1: SCREAMING_SNAKE_CASE : int = inputs['''labels'''].unsqueeze(1 ).repeat(1 , problem_type['''num_labels'''] ) SCREAMING_SNAKE_CASE : Optional[int] = inputs['''labels'''].to(problem_type['''dtype'''] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowercase_ ) as warning_list: SCREAMING_SNAKE_CASE : Dict = model(**lowercase_ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def __A ( self : List[str] ): '''simple docstring''' for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE : Union[str, Any] = DeiTModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) def A ( ): SCREAMING_SNAKE_CASE : Tuple = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase__ ( unittest.TestCase): @cached_property def __A ( self : Any ): '''simple docstring''' return ( DeiTImageProcessor.from_pretrained('''facebook/deit-base-distilled-patch16-224''' ) if is_vision_available() else None ) @slow def __A ( self : Any ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = DeiTForImageClassificationWithTeacher.from_pretrained('''facebook/deit-base-distilled-patch16-224''' ).to( lowercase_ ) SCREAMING_SNAKE_CASE : Optional[Any] = self.default_image_processor SCREAMING_SNAKE_CASE : Dict = prepare_img() SCREAMING_SNAKE_CASE : Tuple = image_processor(images=lowercase_ , return_tensors='''pt''' ).to(lowercase_ ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE : Tuple = model(**lowercase_ ) # verify the logits SCREAMING_SNAKE_CASE : Optional[Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowercase_ ) SCREAMING_SNAKE_CASE : Dict = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(lowercase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase_ , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def __A ( self : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = DeiTModel.from_pretrained( '''facebook/deit-base-distilled-patch16-224''' , torch_dtype=torch.floataa , device_map='''auto''' ) SCREAMING_SNAKE_CASE : Optional[Any] = self.default_image_processor SCREAMING_SNAKE_CASE : List[Any] = prepare_img() SCREAMING_SNAKE_CASE : int = image_processor(images=lowercase_ , return_tensors='''pt''' ) SCREAMING_SNAKE_CASE : Tuple = inputs.pixel_values.to(lowercase_ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): SCREAMING_SNAKE_CASE : Tuple = model(lowercase_ )
358
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def A ( _lowercase , _lowercase , _lowercase , _lowercase ): if isinstance(_lowercase , _lowercase ): SCREAMING_SNAKE_CASE : Any = np.full((len(_lowercase ), sequence_length, 2) , _lowercase ) else: SCREAMING_SNAKE_CASE : List[Any] = np.full((len(_lowercase ), sequence_length) , _lowercase ) for i, tensor in enumerate(_lowercase ): if padding_side == "right": if isinstance(_lowercase , _lowercase ): SCREAMING_SNAKE_CASE : Tuple = tensor[:sequence_length] else: SCREAMING_SNAKE_CASE : Any = tensor[:sequence_length] else: if isinstance(_lowercase , _lowercase ): SCREAMING_SNAKE_CASE : Any = tensor[:sequence_length] else: SCREAMING_SNAKE_CASE : Union[str, Any] = tensor[:sequence_length] return out_tensor.tolist() def A ( _lowercase ): SCREAMING_SNAKE_CASE : Union[str, Any] = ord(_lowercase ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 123 and cp <= 126): return True SCREAMING_SNAKE_CASE : Optional[Any] = unicodedata.category(_lowercase ) if cat.startswith('''P''' ): return True return False @dataclass class lowercase__ ( UpperCamelCase_): UpperCamelCase_ = 42 UpperCamelCase_ = True UpperCamelCase_ = None UpperCamelCase_ = None UpperCamelCase_ = -100 UpperCamelCase_ = "pt" def __A ( self : Optional[int] , UpperCamelCase__ : List[Any] ): '''simple docstring''' import torch SCREAMING_SNAKE_CASE : str = '''label''' if '''label''' in features[0].keys() else '''labels''' SCREAMING_SNAKE_CASE : str = [feature[label_name] for feature in features] if label_name in features[0].keys() else None SCREAMING_SNAKE_CASE : Dict = self.tokenizer.pad( UpperCamelCase__ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch SCREAMING_SNAKE_CASE : Tuple = torch.tensor(batch['''entity_ids'''] ).shape[1] SCREAMING_SNAKE_CASE : Tuple = self.tokenizer.padding_side if padding_side == "right": SCREAMING_SNAKE_CASE : int = [ list(UpperCamelCase__ ) + [self.label_pad_token_id] * (sequence_length - len(UpperCamelCase__ )) for label in labels ] else: SCREAMING_SNAKE_CASE : str = [ [self.label_pad_token_id] * (sequence_length - len(UpperCamelCase__ )) + list(UpperCamelCase__ ) for label in labels ] SCREAMING_SNAKE_CASE : List[str] = [feature['''ner_tags'''] for feature in features] SCREAMING_SNAKE_CASE : Dict = padding_tensor(UpperCamelCase__ , -1 , UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Any = [feature['''original_entity_spans'''] for feature in features] SCREAMING_SNAKE_CASE : Dict = padding_tensor(UpperCamelCase__ , (-1, -1) , UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = {k: torch.tensor(UpperCamelCase__ , dtype=torch.intaa ) for k, v in batch.items()} return batch
258
0
"""simple docstring""" def lowercase ( lowerCAmelCase__ : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: __a = set() # Replace all the whitespace in our sentence __a = input_str.replace(''' ''' , '''''' ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(lowerCAmelCase__ ) == 26 def lowercase ( lowerCAmelCase__ : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: __a = [False] * 26 for char in input_str: if char.islower(): __a = True elif char.isupper(): __a = True return all(lowerCAmelCase__ ) def lowercase ( lowerCAmelCase__ : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def lowercase ( ) -> None: from timeit import timeit __a = '''from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest''' print(timeit('''is_pangram()''' , setup=lowerCAmelCase__ ) ) print(timeit('''is_pangram_faster()''' , setup=lowerCAmelCase__ ) ) print(timeit('''is_pangram_fastest()''' , setup=lowerCAmelCase__ ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
45
"""simple docstring""" from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class __lowerCAmelCase : '''simple docstring''' __UpperCAmelCase : List[str] __UpperCAmelCase : Optional[str] = None # Automatically constructed __UpperCAmelCase : ClassVar[str] = "dict" __UpperCAmelCase : ClassVar[Any] = None __UpperCAmelCase : str = field(default='Translation' , init=__SCREAMING_SNAKE_CASE , repr=__SCREAMING_SNAKE_CASE ) def __call__( self ): return pa.struct({lang: pa.string() for lang in sorted(self.languages )} ) def __UpperCAmelCase ( self ): from .features import Value return {k: Value('''string''' ) for k in sorted(self.languages )} @dataclass class __lowerCAmelCase : '''simple docstring''' __UpperCAmelCase : Optional[List] = None __UpperCAmelCase : Optional[int] = None __UpperCAmelCase : Optional[str] = None # Automatically constructed __UpperCAmelCase : ClassVar[str] = "dict" __UpperCAmelCase : ClassVar[Any] = None __UpperCAmelCase : str = field(default='TranslationVariableLanguages' , init=__SCREAMING_SNAKE_CASE , repr=__SCREAMING_SNAKE_CASE ) def __UpperCAmelCase ( self ): __a = sorted(set(self.languages ) ) if self.languages else None __a = len(self.languages ) if self.languages else None def __call__( self ): return pa.struct({'''language''': pa.list_(pa.string() ), '''translation''': pa.list_(pa.string() )} ) def __UpperCAmelCase ( self , _a ): __a = set(self.languages ) if self.languages and set(_a ) - lang_set: raise ValueError( f'''Some languages in example ({', '.join(sorted(set(_a ) - lang_set ) )}) are not in valid set ({', '.join(_a )}).''' ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. __a = [] for lang, text in translation_dict.items(): if isinstance(_a , _a ): translation_tuples.append((lang, text) ) else: translation_tuples.extend([(lang, el) for el in text] ) # Ensure translations are in ascending order by language code. __a , __a = zip(*sorted(_a ) ) return {"language": languages, "translation": translations} def __UpperCAmelCase ( self ): from .features import Sequence, Value return { "language": Sequence(Value('''string''' ) ), "translation": Sequence(Value('''string''' ) ), }
45
1
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class lowercase : def __init__( self : Optional[int] , _UpperCamelCase : List[str] , _UpperCamelCase : str=100 , _UpperCamelCase : List[str]=13 , _UpperCamelCase : Dict=30 , _UpperCamelCase : Optional[Any]=2 , _UpperCamelCase : int=3 , _UpperCamelCase : List[str]=True , _UpperCamelCase : List[Any]=True , _UpperCamelCase : List[str]=32 , _UpperCamelCase : Optional[Any]=4 , _UpperCamelCase : List[str]=4 , _UpperCamelCase : Optional[Any]=37 , _UpperCamelCase : int="gelu" , _UpperCamelCase : Dict=0.1 , _UpperCamelCase : Optional[Any]=0.1 , _UpperCamelCase : Any=10 , _UpperCamelCase : Dict=0.0_2 , _UpperCamelCase : str=3 , _UpperCamelCase : Optional[Any]=None , _UpperCamelCase : Tuple=[0, 1, 2, 3] , ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = 100 SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = image_size SCREAMING_SNAKE_CASE = patch_size SCREAMING_SNAKE_CASE = num_channels SCREAMING_SNAKE_CASE = is_training SCREAMING_SNAKE_CASE = use_labels SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = intermediate_size SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = hidden_dropout_prob SCREAMING_SNAKE_CASE = attention_probs_dropout_prob SCREAMING_SNAKE_CASE = type_sequence_label_size SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = scope SCREAMING_SNAKE_CASE = out_indices SCREAMING_SNAKE_CASE = num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) SCREAMING_SNAKE_CASE = (image_size // patch_size) ** 2 SCREAMING_SNAKE_CASE = num_patches + 1 def __snake_case( self : Optional[Any] ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None if self.use_labels: SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case( self : List[Any] ) -> Dict: '''simple docstring''' return BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_a , initializer_range=self.initializer_range , out_indices=self.out_indices , ) def __snake_case( self : Union[str, Any] , _UpperCamelCase : Optional[int] , _UpperCamelCase : List[Any] , _UpperCamelCase : Optional[Any] , _UpperCamelCase : List[str] ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = BeitModel(config=_a ) model.to(_a ) model.eval() SCREAMING_SNAKE_CASE = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case( self : List[Any] , _UpperCamelCase : int , _UpperCamelCase : Optional[int] , _UpperCamelCase : int , _UpperCamelCase : List[Any] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = BeitForMaskedImageModeling(config=_a ) model.to(_a ) model.eval() SCREAMING_SNAKE_CASE = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def __snake_case( self : Tuple , _UpperCamelCase : List[Any] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : List[Any] , _UpperCamelCase : str ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = self.type_sequence_label_size SCREAMING_SNAKE_CASE = BeitForImageClassification(_a ) model.to(_a ) model.eval() SCREAMING_SNAKE_CASE = model(_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images SCREAMING_SNAKE_CASE = 1 SCREAMING_SNAKE_CASE = BeitForImageClassification(_a ) model.to(_a ) model.eval() SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE = model(_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __snake_case( self : List[str] , _UpperCamelCase : List[Any] , _UpperCamelCase : List[Any] , _UpperCamelCase : Dict , _UpperCamelCase : Any ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.num_labels SCREAMING_SNAKE_CASE = BeitForSemanticSegmentation(_a ) model.to(_a ) model.eval() SCREAMING_SNAKE_CASE = model(_a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) SCREAMING_SNAKE_CASE = model(_a , labels=_a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def __snake_case( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE = config_and_inputs SCREAMING_SNAKE_CASE = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowercase ( a , a , unittest.TestCase ): lowercase__ : Optional[Any] = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) lowercase__ : Optional[Any] = ( { """feature-extraction""": BeitModel, """image-classification""": BeitForImageClassification, """image-segmentation""": BeitForSemanticSegmentation, } if is_torch_available() else {} ) lowercase__ : str = False lowercase__ : Union[str, Any] = False lowercase__ : Dict = False def __snake_case( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = BeitModelTester(self ) SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def __snake_case( self : Optional[int] ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="BEiT does not use inputs_embeds" ) def __snake_case( self : List[Any] ) -> Optional[Any]: '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason="BEiT has some layers using `add_module` which doesn't work well with `nn.DataParallel`" ) def __snake_case( self : Any ) -> Tuple: '''simple docstring''' pass def __snake_case( self : List[str] ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = model_class(_a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) SCREAMING_SNAKE_CASE = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_a , nn.Linear ) ) def __snake_case( self : str ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = model_class(_a ) SCREAMING_SNAKE_CASE = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE = ["pixel_values"] self.assertListEqual(arg_names[:1] , _a ) def __snake_case( self : List[Any] ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def __snake_case( self : Optional[int] ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_a ) def __snake_case( self : List[Any] ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) def __snake_case( self : str ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_a ) def __snake_case( self : Dict ) -> Optional[Any]: '''simple docstring''' if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(_a ), BeitForMaskedImageModeling]: continue SCREAMING_SNAKE_CASE = model_class(_a ) model.to(_a ) model.train() SCREAMING_SNAKE_CASE = self._prepare_for_class(_a , _a , return_labels=_a ) SCREAMING_SNAKE_CASE = model(**_a ).loss loss.backward() def __snake_case( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(_a ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue SCREAMING_SNAKE_CASE = model_class(_a ) model.gradient_checkpointing_enable() model.to(_a ) model.train() SCREAMING_SNAKE_CASE = self._prepare_for_class(_a , _a , return_labels=_a ) SCREAMING_SNAKE_CASE = model(**_a ).loss loss.backward() def __snake_case( self : List[str] ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE = _config_zero_init(_a ) for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = model_class(config=_a ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"Parameter {name} of model {model_class} seems not properly initialized" , ) @slow def __snake_case( self : str ) -> Any: '''simple docstring''' for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE = BeitModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowerCamelCase (): SCREAMING_SNAKE_CASE = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowercase ( unittest.TestCase ): @cached_property def __snake_case( self : Union[str, Any] ) -> List[Any]: '''simple docstring''' return BeitImageProcessor.from_pretrained("microsoft/beit-base-patch16-224" ) if is_vision_available() else None @slow def __snake_case( self : Optional[int] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = BeitForMaskedImageModeling.from_pretrained("microsoft/beit-base-patch16-224-pt22k" ).to(_a ) SCREAMING_SNAKE_CASE = self.default_image_processor SCREAMING_SNAKE_CASE = prepare_img() SCREAMING_SNAKE_CASE = image_processor(images=_a , return_tensors="pt" ).pixel_values.to(_a ) # prepare bool_masked_pos SCREAMING_SNAKE_CASE = torch.ones((1, 196) , dtype=torch.bool ).to(_a ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE = model(pixel_values=_a , bool_masked_pos=_a ) SCREAMING_SNAKE_CASE = outputs.logits # verify the logits SCREAMING_SNAKE_CASE = torch.Size((1, 196, 8_192) ) self.assertEqual(logits.shape , _a ) SCREAMING_SNAKE_CASE = torch.tensor( [[-3.2_4_3_7, 0.5_0_7_2, -1_3.9_1_7_4], [-3.2_4_5_6, 0.4_9_4_8, -1_3.9_4_0_1], [-3.2_0_3_3, 0.5_1_2_1, -1_3.8_5_5_0]] ).to(_a ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] , _a , atol=1e-2 ) ) @slow def __snake_case( self : Dict ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = BeitForImageClassification.from_pretrained("microsoft/beit-base-patch16-224" ).to(_a ) SCREAMING_SNAKE_CASE = self.default_image_processor SCREAMING_SNAKE_CASE = prepare_img() SCREAMING_SNAKE_CASE = image_processor(images=_a , return_tensors="pt" ).to(_a ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE = model(**_a ) SCREAMING_SNAKE_CASE = outputs.logits # verify the logits SCREAMING_SNAKE_CASE = torch.Size((1, 1_000) ) self.assertEqual(logits.shape , _a ) SCREAMING_SNAKE_CASE = torch.tensor([-1.2_3_8_5, -1.0_9_8_7, -1.0_1_0_8] ).to(_a ) self.assertTrue(torch.allclose(logits[0, :3] , _a , atol=1e-4 ) ) SCREAMING_SNAKE_CASE = 281 self.assertEqual(logits.argmax(-1 ).item() , _a ) @slow def __snake_case( self : int ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = BeitForImageClassification.from_pretrained("microsoft/beit-large-patch16-224-pt22k-ft22k" ).to( _a ) SCREAMING_SNAKE_CASE = self.default_image_processor SCREAMING_SNAKE_CASE = prepare_img() SCREAMING_SNAKE_CASE = image_processor(images=_a , return_tensors="pt" ).to(_a ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE = model(**_a ) SCREAMING_SNAKE_CASE = outputs.logits # verify the logits SCREAMING_SNAKE_CASE = torch.Size((1, 21_841) ) self.assertEqual(logits.shape , _a ) SCREAMING_SNAKE_CASE = torch.tensor([1.6_8_8_1, -0.2_7_8_7, 0.5_9_0_1] ).to(_a ) self.assertTrue(torch.allclose(logits[0, :3] , _a , atol=1e-4 ) ) SCREAMING_SNAKE_CASE = 2_396 self.assertEqual(logits.argmax(-1 ).item() , _a ) @slow def __snake_case( self : Union[str, Any] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = BeitForSemanticSegmentation.from_pretrained("microsoft/beit-base-finetuned-ade-640-640" ) SCREAMING_SNAKE_CASE = model.to(_a ) SCREAMING_SNAKE_CASE = BeitImageProcessor(do_resize=_a , size=640 , do_center_crop=_a ) SCREAMING_SNAKE_CASE = load_dataset("hf-internal-testing/fixtures_ade20k" , split="test" ) SCREAMING_SNAKE_CASE = Image.open(ds[0]["file"] ) SCREAMING_SNAKE_CASE = image_processor(images=_a , return_tensors="pt" ).to(_a ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE = model(**_a ) SCREAMING_SNAKE_CASE = outputs.logits # verify the logits SCREAMING_SNAKE_CASE = torch.Size((1, 150, 160, 160) ) self.assertEqual(logits.shape , _a ) SCREAMING_SNAKE_CASE = version.parse(PIL.__version__ ) < version.parse("9.0.0" ) if is_pillow_less_than_a: SCREAMING_SNAKE_CASE = torch.tensor( [ [[-4.9_2_2_5, -2.3_9_5_4, -3.0_5_2_2], [-2.8_8_2_2, -1.0_0_4_6, -1.7_5_6_1], [-2.9_5_4_9, -1.3_2_2_8, -2.1_3_4_7]], [[-5.8_1_6_8, -3.4_1_2_9, -4.0_7_7_8], [-3.8_6_5_1, -2.2_2_1_4, -3.0_2_7_7], [-3.8_3_5_6, -2.4_6_4_3, -3.3_5_3_5]], [[-0.0_0_7_8, 3.9_9_5_2, 4.0_7_5_4], [2.9_8_5_6, 4.6_9_4_4, 5.0_0_3_5], [3.2_4_1_3, 4.7_8_1_3, 4.9_9_6_9]], ] , device=_a , ) else: SCREAMING_SNAKE_CASE = torch.tensor( [ [[-4.8_9_6_0, -2.3_6_8_8, -3.0_3_5_5], [-2.8_4_7_8, -0.9_8_3_6, -1.7_4_1_8], [-2.9_4_4_9, -1.3_3_3_2, -2.1_4_5_6]], [[-5.8_0_8_1, -3.4_1_2_4, -4.1_0_0_6], [-3.8_5_6_1, -2.2_0_8_1, -3.0_3_2_3], [-3.8_3_6_5, -2.4_6_0_1, -3.3_6_6_9]], [[-0.0_3_0_9, 3.9_8_6_8, 4.0_5_4_0], [2.9_6_4_0, 4.6_8_7_7, 4.9_9_7_6], [3.2_0_8_1, 4.7_6_9_0, 4.9_9_4_2]], ] , device=_a , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , _a , atol=1e-4 ) ) @slow def __snake_case( self : str ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = BeitForSemanticSegmentation.from_pretrained("microsoft/beit-base-finetuned-ade-640-640" ) SCREAMING_SNAKE_CASE = model.to(_a ) SCREAMING_SNAKE_CASE = BeitImageProcessor(do_resize=_a , size=640 , do_center_crop=_a ) SCREAMING_SNAKE_CASE = load_dataset("hf-internal-testing/fixtures_ade20k" , split="test" ) SCREAMING_SNAKE_CASE = Image.open(ds[0]["file"] ) SCREAMING_SNAKE_CASE = image_processor(images=_a , return_tensors="pt" ).to(_a ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE = model(**_a ) SCREAMING_SNAKE_CASE = outputs.logits.detach().cpu() SCREAMING_SNAKE_CASE = image_processor.post_process_semantic_segmentation(outputs=_a , target_sizes=[(500, 300)] ) SCREAMING_SNAKE_CASE = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , _a ) SCREAMING_SNAKE_CASE = image_processor.post_process_semantic_segmentation(outputs=_a ) SCREAMING_SNAKE_CASE = torch.Size((160, 160) ) self.assertEqual(segmentation[0].shape , _a )
369
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase : int = logging.get_logger(__name__) _lowerCamelCase : Union[str, Any] = { '''microsoft/unispeech-sat-base-100h-libri-ft''': ( '''https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json''' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class lowercase ( a ): lowercase__ : Tuple = """unispeech-sat""" def __init__( self : str , _UpperCamelCase : Tuple=32 , _UpperCamelCase : Union[str, Any]=768 , _UpperCamelCase : Tuple=12 , _UpperCamelCase : List[str]=12 , _UpperCamelCase : Tuple=3_072 , _UpperCamelCase : List[str]="gelu" , _UpperCamelCase : Tuple=0.1 , _UpperCamelCase : Any=0.1 , _UpperCamelCase : Union[str, Any]=0.1 , _UpperCamelCase : Dict=0.0 , _UpperCamelCase : Tuple=0.0 , _UpperCamelCase : Tuple=0.1 , _UpperCamelCase : Optional[Any]=0.1 , _UpperCamelCase : Tuple=0.0_2 , _UpperCamelCase : Optional[int]=1e-5 , _UpperCamelCase : Union[str, Any]="group" , _UpperCamelCase : Optional[int]="gelu" , _UpperCamelCase : Tuple=(512, 512, 512, 512, 512, 512, 512) , _UpperCamelCase : List[str]=(5, 2, 2, 2, 2, 2, 2) , _UpperCamelCase : Optional[int]=(10, 3, 3, 3, 3, 2, 2) , _UpperCamelCase : Optional[int]=False , _UpperCamelCase : Dict=128 , _UpperCamelCase : Optional[int]=16 , _UpperCamelCase : Tuple=False , _UpperCamelCase : Union[str, Any]=True , _UpperCamelCase : Optional[Any]=0.0_5 , _UpperCamelCase : Union[str, Any]=10 , _UpperCamelCase : Union[str, Any]=2 , _UpperCamelCase : str=0.0 , _UpperCamelCase : List[Any]=10 , _UpperCamelCase : Optional[int]=0 , _UpperCamelCase : Any=320 , _UpperCamelCase : List[Any]=2 , _UpperCamelCase : str=0.1 , _UpperCamelCase : str=100 , _UpperCamelCase : int=256 , _UpperCamelCase : Optional[Any]=256 , _UpperCamelCase : List[Any]=0.1 , _UpperCamelCase : str="mean" , _UpperCamelCase : int=False , _UpperCamelCase : Optional[Any]=False , _UpperCamelCase : Any=256 , _UpperCamelCase : str=(512, 512, 512, 512, 1_500) , _UpperCamelCase : List[Any]=(5, 3, 3, 1, 1) , _UpperCamelCase : Union[str, Any]=(1, 2, 3, 1, 1) , _UpperCamelCase : Any=512 , _UpperCamelCase : str=0 , _UpperCamelCase : int=1 , _UpperCamelCase : Any=2 , _UpperCamelCase : Optional[Any]=504 , **_UpperCamelCase : str , ) -> int: '''simple docstring''' super().__init__(**_UpperCamelCase , pad_token_id=_UpperCamelCase , bos_token_id=_UpperCamelCase , eos_token_id=_UpperCamelCase ) SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = feat_extract_norm SCREAMING_SNAKE_CASE = feat_extract_activation SCREAMING_SNAKE_CASE = list(_UpperCamelCase ) SCREAMING_SNAKE_CASE = list(_UpperCamelCase ) SCREAMING_SNAKE_CASE = list(_UpperCamelCase ) SCREAMING_SNAKE_CASE = conv_bias SCREAMING_SNAKE_CASE = num_conv_pos_embeddings SCREAMING_SNAKE_CASE = num_conv_pos_embedding_groups SCREAMING_SNAKE_CASE = len(self.conv_dim ) SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = intermediate_size SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = hidden_dropout SCREAMING_SNAKE_CASE = attention_dropout SCREAMING_SNAKE_CASE = activation_dropout SCREAMING_SNAKE_CASE = feat_proj_dropout SCREAMING_SNAKE_CASE = final_dropout SCREAMING_SNAKE_CASE = layerdrop SCREAMING_SNAKE_CASE = layer_norm_eps SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = num_clusters SCREAMING_SNAKE_CASE = do_stable_layer_norm SCREAMING_SNAKE_CASE = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==" " `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =" F" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`," F" `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 SCREAMING_SNAKE_CASE = apply_spec_augment SCREAMING_SNAKE_CASE = mask_time_prob SCREAMING_SNAKE_CASE = mask_time_length SCREAMING_SNAKE_CASE = mask_time_min_masks SCREAMING_SNAKE_CASE = mask_feature_prob SCREAMING_SNAKE_CASE = mask_feature_length SCREAMING_SNAKE_CASE = mask_feature_min_masks # parameters for pretraining with codevector quantized representations SCREAMING_SNAKE_CASE = num_codevectors_per_group SCREAMING_SNAKE_CASE = num_codevector_groups SCREAMING_SNAKE_CASE = contrastive_logits_temperature SCREAMING_SNAKE_CASE = feat_quantizer_dropout SCREAMING_SNAKE_CASE = num_negatives SCREAMING_SNAKE_CASE = codevector_dim SCREAMING_SNAKE_CASE = proj_codevector_dim SCREAMING_SNAKE_CASE = diversity_loss_weight # ctc loss SCREAMING_SNAKE_CASE = ctc_loss_reduction SCREAMING_SNAKE_CASE = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. SCREAMING_SNAKE_CASE = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. SCREAMING_SNAKE_CASE = list(_UpperCamelCase ) SCREAMING_SNAKE_CASE = list(_UpperCamelCase ) SCREAMING_SNAKE_CASE = list(_UpperCamelCase ) SCREAMING_SNAKE_CASE = xvector_output_dim @property def __snake_case( self : Tuple ) -> str: '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
206
0
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def __UpperCAmelCase ( a_ , a_ , a_=None): # set parameter of one layer assert torch_layer.weight.shape == weight.shape, f'''{torch_layer} layer.weight does not match''' snake_case_ = nn.Parameter(_lowercase) if bias is not None: assert torch_layer.bias.shape == bias.shape, f'''{torch_layer} layer.bias does not match''' snake_case_ = nn.Parameter(_lowercase) def __UpperCAmelCase ( a_ , a_ , a_): # set torch weights for 1-to-1 comparison snake_case_ = np.asarray(weights[0]) snake_case_ = np.asarray(weights[1]) snake_case_ = np.asarray(weights[2]) set_param( torch_layer.self_attention.query_key , torch.tensor(_lowercase).transpose(1 , 2).contiguous().view(-1 , _lowercase) , ) set_param( torch_layer.self_attention.value , torch.tensor(_lowercase).transpose(1 , 2).contiguous().view(-1 , _lowercase) , ) set_param( torch_layer.output.dense , torch.tensor(_lowercase).view(-1 , _lowercase).contiguous().transpose(0 , 1) , ) def __UpperCAmelCase ( a_ , a_ , a_): # set torch weights for 1-to-1 comparison snake_case_ = np.asarray(weights[0]) snake_case_ = np.asarray(weights[1]) snake_case_ = np.asarray(weights[2]) snake_case_ = np.asarray(weights[3]) set_param( torch_layer.self_attention.query , torch.tensor(_lowercase).transpose(1 , 2).contiguous().view(-1 , _lowercase) , ) set_param( torch_layer.self_attention.key , torch.tensor(_lowercase).transpose(1 , 2).contiguous().view(-1 , _lowercase) , ) set_param( torch_layer.self_attention.value , torch.tensor(_lowercase).transpose(1 , 2).contiguous().view(-1 , _lowercase) , ) set_param( torch_layer.output.dense , torch.tensor(_lowercase).view(-1 , _lowercase).contiguous().transpose(0 , 1) , ) def __UpperCAmelCase ( a_ , a_ , a_): # layernorm 1 snake_case_ = weights[0][0][0] snake_case_ = np.asarray(layer_norm_a[0]) snake_case_ = np.asarray(layer_norm_a[1]) set_param( torch_block.attention.layer_norm , torch.tensor(_lowercase) , torch.tensor(_lowercase) , ) # lsh weights + output snake_case_ = weights[0][1] if len(_lowercase) < 4: set_layer_weights_in_torch_lsh(_lowercase , torch_block.attention , _lowercase) else: set_layer_weights_in_torch_local(_lowercase , torch_block.attention , _lowercase) # intermediate weighs snake_case_ = weights[2][0][1][2] # Chunked Feed Forward if len(_lowercase) == 4: snake_case_ = intermediate_weights[2] # layernorm 2 snake_case_ = np.asarray(intermediate_weights[0][0]) snake_case_ = np.asarray(intermediate_weights[0][1]) set_param( torch_block.feed_forward.layer_norm , torch.tensor(_lowercase) , torch.tensor(_lowercase) , ) # intermediate dense snake_case_ = np.asarray(intermediate_weights[1][0]) snake_case_ = np.asarray(intermediate_weights[1][1]) set_param( torch_block.feed_forward.dense.dense , torch.tensor(_lowercase).transpose(0 , 1).contiguous() , torch.tensor(_lowercase) , ) # intermediate out snake_case_ = np.asarray(intermediate_weights[4][0]) snake_case_ = np.asarray(intermediate_weights[4][1]) set_param( torch_block.feed_forward.output.dense , torch.tensor(_lowercase).transpose(0 , 1).contiguous() , torch.tensor(_lowercase) , ) def __UpperCAmelCase ( a_ , a_ , a_): # reformer model snake_case_ = torch_model.reformer # word embeds snake_case_ = np.asarray(weights[1]) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(_lowercase) , ) if isinstance(weights[3] , _lowercase): snake_case_ = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights)): snake_case_ = np.asarray(weights[3][emb_idx][0]) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f'''{position_embeddings[emb_idx]} emb does not match''' snake_case_ = nn.Parameter(torch.tensor(_lowercase)) snake_case_ = weights[5] assert len(torch_model_reformer.encoder.layers) * 4 == len( _lowercase), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers): snake_case_ = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(_lowercase , _lowercase , _lowercase) # output layer norm snake_case_ = np.asarray(weights[7][0]) snake_case_ = np.asarray(weights[7][1]) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(_lowercase) , torch.tensor(_lowercase) , ) # output embeddings snake_case_ = np.asarray(weights[9][0]) snake_case_ = np.asarray(weights[9][1]) set_param( torch_model.lm_head.decoder , torch.tensor(_lowercase).transpose(0 , 1).contiguous() , torch.tensor(_lowercase) , ) def __UpperCAmelCase ( a_ , a_ , a_): # Initialise PyTorch model snake_case_ = ReformerConfig.from_json_file(_lowercase) print(f'''Building PyTorch model from configuration: {config}''') snake_case_ = ReformerModelWithLMHead(_lowercase) with open(_lowercase , 'rb') as f: snake_case_ = pickle.load(_lowercase)["""weights"""] set_model_weights_in_torch(_lowercase , _lowercase , config.hidden_size) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''') torch.save(model.state_dict() , _lowercase) if __name__ == "__main__": lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--trax_model_pkl_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained Reformer model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) lowercase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
178
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def __lowerCamelCase ( _lowercase ) -> Tuple: return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def __lowerCamelCase ( _lowercase ) -> List[Any]: UpperCAmelCase : Any = create_tensor(_lowercase ) UpperCAmelCase : Union[str, Any] = gather(_lowercase ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def __lowerCamelCase ( _lowercase ) -> Optional[int]: UpperCAmelCase : Any = [state.process_index] UpperCAmelCase : Union[str, Any] = gather_object(_lowercase ) assert len(_lowercase ) == state.num_processes, F'''{gathered_obj}, {len(_lowercase )} != {state.num_processes}''' assert gathered_obj == list(range(state.num_processes ) ), F'''{gathered_obj} != {list(range(state.num_processes ) )}''' def __lowerCamelCase ( _lowercase ) -> List[Any]: UpperCAmelCase : Optional[int] = create_tensor(_lowercase ) UpperCAmelCase : List[str] = broadcast(_lowercase ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def __lowerCamelCase ( _lowercase ) -> Tuple: # We need to pad the tensor with one more element if we are the main process # to ensure that we can pad if state.is_main_process: UpperCAmelCase : Optional[Any] = torch.arange(state.num_processes + 1 ).to(state.device ) else: UpperCAmelCase : Tuple = torch.arange(state.num_processes ).to(state.device ) UpperCAmelCase : Optional[Any] = pad_across_processes(_lowercase ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def __lowerCamelCase ( _lowercase ) -> Dict: # For now runs on only two processes if state.num_processes != 2: return UpperCAmelCase : Optional[Any] = create_tensor(_lowercase ) UpperCAmelCase : Optional[Any] = reduce(_lowercase , """sum""" ) UpperCAmelCase : Optional[Any] = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(_lowercase , _lowercase ), F'''{reduced_tensor} != {truth_tensor}''' def __lowerCamelCase ( _lowercase ) -> Optional[Any]: # For now runs on only two processes if state.num_processes != 2: return UpperCAmelCase : Tuple = create_tensor(_lowercase ) UpperCAmelCase : Optional[int] = reduce(_lowercase , """mean""" ) UpperCAmelCase : str = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(_lowercase , _lowercase ), F'''{reduced_tensor} != {truth_tensor}''' def __lowerCamelCase ( _lowercase ) -> Optional[int]: # For xla_spawn (TPUs) main() def __lowerCamelCase ( ) -> int: UpperCAmelCase : List[Any] = PartialState() state.print(F'''State: {state}''' ) state.print("""testing gather""" ) test_gather(_lowercase ) state.print("""testing gather_object""" ) test_gather_object(_lowercase ) state.print("""testing broadcast""" ) test_broadcast(_lowercase ) state.print("""testing pad_across_processes""" ) test_pad_across_processes(_lowercase ) state.print("""testing reduce_sum""" ) test_reduce_sum(_lowercase ) state.print("""testing reduce_mean""" ) test_reduce_mean(_lowercase ) if __name__ == "__main__": main()
265
0
"""simple docstring""" def SCREAMING_SNAKE_CASE__ ( snake_case_ ) -> list: """simple docstring""" for i in range(len(snake_case_ ) - 1, 0, -1 ): a = False for j in range(snake_case_, 0, -1 ): if unsorted[j] < unsorted[j - 1]: a , a = unsorted[j - 1], unsorted[j] a = True for j in range(snake_case_ ): if unsorted[j] > unsorted[j + 1]: a , a = unsorted[j + 1], unsorted[j] a = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ : Optional[Any] = input("""Enter numbers separated by a comma:\n""").strip() UpperCamelCase__ : str = [int(item) for item in user_input.split(""",""")] print(F"{cocktail_shaker_sort(unsorted) = }")
355
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ : str = logging.get_logger(__name__) UpperCamelCase__ : Optional[int] = { """studio-ousia/luke-base""": """https://huggingface.co/studio-ousia/luke-base/resolve/main/config.json""", """studio-ousia/luke-large""": """https://huggingface.co/studio-ousia/luke-large/resolve/main/config.json""", } class lowerCamelCase_ ( a_ ): SCREAMING_SNAKE_CASE_ = 'luke' def __init__( self : Dict ,__lowerCamelCase : Optional[Any]=5_02_67 ,__lowerCamelCase : str=50_00_00 ,__lowerCamelCase : Any=7_68 ,__lowerCamelCase : int=2_56 ,__lowerCamelCase : Optional[int]=12 ,__lowerCamelCase : Tuple=12 ,__lowerCamelCase : Any=30_72 ,__lowerCamelCase : Any="gelu" ,__lowerCamelCase : Any=0.1 ,__lowerCamelCase : Tuple=0.1 ,__lowerCamelCase : Tuple=5_12 ,__lowerCamelCase : int=2 ,__lowerCamelCase : Optional[int]=0.02 ,__lowerCamelCase : List[Any]=1e-12 ,__lowerCamelCase : Dict=True ,__lowerCamelCase : Tuple=None ,__lowerCamelCase : Any=1 ,__lowerCamelCase : Dict=0 ,__lowerCamelCase : Any=2 ,**__lowerCamelCase : str ,): '''simple docstring''' super().__init__(pad_token_id=__lowerCamelCase ,bos_token_id=__lowerCamelCase ,eos_token_id=__lowerCamelCase ,**__lowerCamelCase ) a = vocab_size a = entity_vocab_size a = hidden_size a = entity_emb_size a = num_hidden_layers a = num_attention_heads a = hidden_act a = intermediate_size a = hidden_dropout_prob a = attention_probs_dropout_prob a = max_position_embeddings a = type_vocab_size a = initializer_range a = layer_norm_eps a = use_entity_aware_attention a = classifier_dropout
330
0
"""simple docstring""" import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class _lowerCAmelCase ( lowercase ): """simple docstring""" def __init__( self : Optional[int], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[str]=1_3, UpperCAmelCase__ : Union[str, Any]=7, UpperCAmelCase__ : Optional[int]=True, UpperCAmelCase__ : List[str]=True, UpperCAmelCase__ : Optional[Any]=False, UpperCAmelCase__ : Union[str, Any]=True, UpperCAmelCase__ : Any=9_9, UpperCAmelCase__ : List[Any]=3_2, UpperCAmelCase__ : int=5, UpperCAmelCase__ : Dict=4, UpperCAmelCase__ : Dict=3_7, UpperCAmelCase__ : List[str]="gelu", UpperCAmelCase__ : Optional[int]=0.1, UpperCAmelCase__ : Optional[Any]=0.1, UpperCAmelCase__ : Union[str, Any]=5_1_2, UpperCAmelCase__ : int=1_6, UpperCAmelCase__ : int=2, UpperCAmelCase__ : Optional[Any]=0.02, UpperCAmelCase__ : Optional[int]=3, UpperCAmelCase__ : Tuple=4, UpperCAmelCase__ : List[str]=None, ): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def _lowercase ( self : Any ): __lowercase = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size], self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) __lowercase = ids_tensor([self.batch_size], self.num_choices ) __lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowercase ( self : str ): return DistilBertConfig( vocab_size=self.vocab_size, dim=self.hidden_size, n_layers=self.num_hidden_layers, n_heads=self.num_attention_heads, hidden_dim=self.intermediate_size, hidden_act=self.hidden_act, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, initializer_range=self.initializer_range, ) def _lowercase ( self : Dict, UpperCAmelCase__ : List[str], UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : str, UpperCAmelCase__ : Any, UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : str ): __lowercase = DistilBertModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowercase = model(UpperCAmelCase__, UpperCAmelCase__ ) __lowercase = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[int], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : List[str], UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[Any] ): __lowercase = DistilBertForMaskedLM(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowercase = model(UpperCAmelCase__, attention_mask=UpperCAmelCase__, labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : Optional[int], UpperCAmelCase__ : List[str], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : Optional[Any] ): __lowercase = DistilBertForQuestionAnswering(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowercase = model( UpperCAmelCase__, attention_mask=UpperCAmelCase__, start_positions=UpperCAmelCase__, end_positions=UpperCAmelCase__ ) self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length) ) def _lowercase ( self : Optional[Any], UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Tuple ): __lowercase = self.num_labels __lowercase = DistilBertForSequenceClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowercase = model(UpperCAmelCase__, attention_mask=UpperCAmelCase__, labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def _lowercase ( self : Any, UpperCAmelCase__ : Any, UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : int, UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : int ): __lowercase = self.num_labels __lowercase = DistilBertForTokenClassification(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowercase = model(UpperCAmelCase__, attention_mask=UpperCAmelCase__, labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self : Optional[int], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[str], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Any ): __lowercase = self.num_choices __lowercase = DistilBertForMultipleChoice(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __lowercase = input_ids.unsqueeze(1 ).expand(-1, self.num_choices, -1 ).contiguous() __lowercase = input_mask.unsqueeze(1 ).expand(-1, self.num_choices, -1 ).contiguous() __lowercase = model( UpperCAmelCase__, attention_mask=UpperCAmelCase__, labels=UpperCAmelCase__, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices) ) def _lowercase ( self : Optional[Any] ): __lowercase = self.prepare_config_and_inputs() ((__lowercase) ,(__lowercase) ,(__lowercase) ,(__lowercase) ,(__lowercase) ,(__lowercase)) = config_and_inputs __lowercase = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _lowerCAmelCase ( lowercase ,lowercase ,unittest.TestCase ): """simple docstring""" __UpperCAmelCase : int = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __UpperCAmelCase : List[Any] = ( { "feature-extraction": DistilBertModel, "fill-mask": DistilBertForMaskedLM, "question-answering": DistilBertForQuestionAnswering, "text-classification": DistilBertForSequenceClassification, "token-classification": DistilBertForTokenClassification, "zero-shot": DistilBertForSequenceClassification, } if is_torch_available() else {} ) __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : Optional[Any] = True __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : str = True def _lowercase ( self : Tuple ): __lowercase = DistilBertModelTester(self ) __lowercase = ConfigTester(self, config_class=UpperCAmelCase__, dim=3_7 ) def _lowercase ( self : Optional[int] ): self.config_tester.run_common_tests() def _lowercase ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*UpperCAmelCase__ ) def _lowercase ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*UpperCAmelCase__ ) def _lowercase ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*UpperCAmelCase__ ) def _lowercase ( self : str ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*UpperCAmelCase__ ) def _lowercase ( self : str ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*UpperCAmelCase__ ) @slow def _lowercase ( self : List[str] ): for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DistilBertModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) @slow @require_torch_gpu def _lowercase ( self : Any ): __lowercase ,__lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase = True __lowercase = model_class(config=UpperCAmelCase__ ) __lowercase = self._prepare_for_class(UpperCAmelCase__, UpperCAmelCase__ ) __lowercase = torch.jit.trace( UpperCAmelCase__, (inputs_dict["input_ids"].to("cpu" ), inputs_dict["attention_mask"].to("cpu" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(UpperCAmelCase__, os.path.join(UpperCAmelCase__, "traced_model.pt" ) ) __lowercase = torch.jit.load(os.path.join(UpperCAmelCase__, "traced_model.pt" ), map_location=UpperCAmelCase__ ) loaded(inputs_dict["input_ids"].to(UpperCAmelCase__ ), inputs_dict["attention_mask"].to(UpperCAmelCase__ ) ) @require_torch class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def _lowercase ( self : List[str] ): __lowercase = DistilBertModel.from_pretrained("distilbert-base-uncased" ) __lowercase = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) __lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase = model(UpperCAmelCase__, attention_mask=UpperCAmelCase__ )[0] __lowercase = torch.Size((1, 1_1, 7_6_8) ) self.assertEqual(output.shape, UpperCAmelCase__ ) __lowercase = torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4], UpperCAmelCase__, atol=1E-4 ) )
17
import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __lowerCAmelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): @register_to_config def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = False , ) -> Tuple: '''simple docstring''' super().__init__() _lowercase =nn.Embedding(lowerCAmelCase , lowerCAmelCase ) _lowercase =nn.Embedding(lowerCAmelCase , lowerCAmelCase ) _lowercase =False _lowercase =nn.Dropout(p=lowerCAmelCase ) _lowercase =TaConfig( vocab_size=lowerCAmelCase , d_model=lowerCAmelCase , num_heads=lowerCAmelCase , d_kv=lowerCAmelCase , d_ff=lowerCAmelCase , dropout_rate=lowerCAmelCase , feed_forward_proj=lowerCAmelCase , is_decoder=lowerCAmelCase , is_encoder_decoder=lowerCAmelCase , ) _lowercase =nn.ModuleList() for lyr_num in range(lowerCAmelCase ): _lowercase =TaBlock(lowerCAmelCase ) self.encoders.append(lowerCAmelCase ) _lowercase =TaLayerNorm(lowerCAmelCase ) _lowercase =nn.Dropout(p=lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ) -> Dict: '''simple docstring''' _lowercase =self.token_embedder(lowerCAmelCase ) _lowercase =encoder_input_tokens.shape[1] _lowercase =torch.arange(lowerCAmelCase , device=encoder_input_tokens.device ) x += self.position_encoding(lowerCAmelCase ) _lowercase =self.dropout_pre(lowerCAmelCase ) # inverted the attention mask _lowercase =encoder_input_tokens.size() _lowercase =self.get_extended_attention_mask(lowerCAmelCase , lowerCAmelCase ) for lyr in self.encoders: _lowercase =lyr(lowerCAmelCase , lowerCAmelCase )[0] _lowercase =self.layer_norm(lowerCAmelCase ) return self.dropout_post(lowerCAmelCase ), encoder_inputs_mask
205
0
"""simple docstring""" import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) __UpperCamelCase = pytest.mark.integration @pytest.mark.parametrize('path' , ['paws', 'csv'] ) def lowercase (SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : str ) -> Any: inspect_dataset(_lowerCAmelCase , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = path + ".py" assert script_name in os.listdir(_lowerCAmelCase ) assert "__pycache__" not in os.listdir(_lowerCAmelCase ) @pytest.mark.filterwarnings('ignore:inspect_metric is deprecated:FutureWarning' ) @pytest.mark.filterwarnings('ignore:metric_module_factory is deprecated:FutureWarning' ) @pytest.mark.parametrize('path' , ['accuracy'] ) def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : str ) -> Optional[int]: inspect_metric(_lowerCAmelCase , _lowerCAmelCase ) SCREAMING_SNAKE_CASE = path + ".py" assert script_name in os.listdir(_lowerCAmelCase ) assert "__pycache__" not in os.listdir(_lowerCAmelCase ) @pytest.mark.parametrize( 'path, config_name, expected_splits' , [ ('squad', 'plain_text', ['train', 'validation']), ('dalle-mini/wit', 'dalle-mini--wit', ['train']), ('paws', 'labeled_final', ['train', 'test', 'validation']), ] , ) def lowercase (SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Union[str, Any] ) -> Optional[int]: SCREAMING_SNAKE_CASE = get_dataset_config_info(_lowerCAmelCase , config_name=_lowerCAmelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( 'path, config_name, expected_exception' , [ ('paws', None, ValueError), ] , ) def lowercase (SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Any ) -> int: with pytest.raises(_lowerCAmelCase ): get_dataset_config_info(_lowerCAmelCase , config_name=_lowerCAmelCase ) @pytest.mark.parametrize( 'path, expected' , [ ('squad', 'plain_text'), ('acronym_identification', 'default'), ('lhoestq/squad', 'plain_text'), ('lhoestq/test', 'default'), ('lhoestq/demo1', 'lhoestq--demo1'), ('dalle-mini/wit', 'dalle-mini--wit'), ] , ) def lowercase (SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Any: SCREAMING_SNAKE_CASE = get_dataset_config_names(_lowerCAmelCase ) assert expected in config_names @pytest.mark.parametrize( 'path, expected_configs, expected_splits_in_first_config' , [ ('squad', ['plain_text'], ['train', 'validation']), ('dalle-mini/wit', ['dalle-mini--wit'], ['train']), ('paws', ['labeled_final', 'labeled_swap', 'unlabeled_final'], ['train', 'test', 'validation']), ] , ) def lowercase (SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : List[Any] ) -> List[str]: SCREAMING_SNAKE_CASE = get_dataset_infos(_lowerCAmelCase ) assert list(infos.keys() ) == expected_configs SCREAMING_SNAKE_CASE = expected_configs[0] assert expected_config in infos SCREAMING_SNAKE_CASE = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( 'path, expected_config, expected_splits' , [ ('squad', 'plain_text', ['train', 'validation']), ('dalle-mini/wit', 'dalle-mini--wit', ['train']), ('paws', 'labeled_final', ['train', 'test', 'validation']), ] , ) def lowercase (SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : List[Any] ) -> List[Any]: SCREAMING_SNAKE_CASE = get_dataset_infos(_lowerCAmelCase ) assert expected_config in infos SCREAMING_SNAKE_CASE = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( 'path, config_name, expected_exception' , [ ('paws', None, ValueError), ] , ) def lowercase (SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : int ) -> int: with pytest.raises(_lowerCAmelCase ): get_dataset_split_names(_lowerCAmelCase , config_name=_lowerCAmelCase )
353
"""simple docstring""" import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowercase (SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : str ) -> Any: # Load configuration defined in the metadata file with open(SCREAMING_SNAKE_CASE_ ) as metadata_file: SCREAMING_SNAKE_CASE = json.load(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = LukeConfig(use_entity_aware_attention=SCREAMING_SNAKE_CASE_ , **metadata['model_config'] ) # Load in the weights from the checkpoint_path SCREAMING_SNAKE_CASE = torch.load(SCREAMING_SNAKE_CASE_ , map_location='cpu' )['module'] # Load the entity vocab file SCREAMING_SNAKE_CASE = load_original_entity_vocab(SCREAMING_SNAKE_CASE_ ) # add an entry for [MASK2] SCREAMING_SNAKE_CASE = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 SCREAMING_SNAKE_CASE = XLMRobertaTokenizer.from_pretrained(metadata['model_config']['bert_model_name'] ) # Add special tokens to the token vocabulary for downstream tasks SCREAMING_SNAKE_CASE = AddedToken('<ent>' , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = AddedToken('<ent2>' , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) tokenizer.add_special_tokens({'additional_special_tokens': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F'Saving tokenizer to {pytorch_dump_folder_path}' ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE_ ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , 'tokenizer_config.json' ) , 'r' ) as f: SCREAMING_SNAKE_CASE = json.load(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = 'MLukeTokenizer' with open(os.path.join(SCREAMING_SNAKE_CASE_ , 'tokenizer_config.json' ) , 'w' ) as f: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , MLukeTokenizer.vocab_files_names['entity_vocab_file'] ) , 'w' ) as f: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) # Initialize the embeddings of the special tokens SCREAMING_SNAKE_CASE = tokenizer.convert_tokens_to_ids(['@'] )[0] SCREAMING_SNAKE_CASE = tokenizer.convert_tokens_to_ids(['#'] )[0] SCREAMING_SNAKE_CASE = state_dict['embeddings.word_embeddings.weight'] SCREAMING_SNAKE_CASE = word_emb[ent_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE = word_emb[enta_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: SCREAMING_SNAKE_CASE = state_dict[bias_name] SCREAMING_SNAKE_CASE = decoder_bias[ent_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE = decoder_bias[enta_init_index].unsqueeze(0 ) SCREAMING_SNAKE_CASE = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: SCREAMING_SNAKE_CASE = F'encoder.layer.{layer_index}.attention.self.' SCREAMING_SNAKE_CASE = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks SCREAMING_SNAKE_CASE = state_dict['entity_embeddings.entity_embeddings.weight'] SCREAMING_SNAKE_CASE = entity_emb[entity_vocab['[MASK]']].unsqueeze(0 ) SCREAMING_SNAKE_CASE = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' SCREAMING_SNAKE_CASE = state_dict['entity_predictions.bias'] SCREAMING_SNAKE_CASE = entity_prediction_bias[entity_vocab['[MASK]']].unsqueeze(0 ) SCREAMING_SNAKE_CASE = torch.cat([entity_prediction_bias, entity_mask_bias] ) SCREAMING_SNAKE_CASE = LukeForMaskedLM(config=SCREAMING_SNAKE_CASE_ ).eval() state_dict.pop('entity_predictions.decoder.weight' ) state_dict.pop('lm_head.decoder.weight' ) state_dict.pop('lm_head.decoder.bias' ) SCREAMING_SNAKE_CASE = OrderedDict() for key, value in state_dict.items(): if not (key.startswith('lm_head' ) or key.startswith('entity_predictions' )): SCREAMING_SNAKE_CASE = state_dict[key] else: SCREAMING_SNAKE_CASE = state_dict[key] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = model.load_state_dict(SCREAMING_SNAKE_CASE_ , strict=SCREAMING_SNAKE_CASE_ ) if set(SCREAMING_SNAKE_CASE_ ) != {"luke.embeddings.position_ids"}: raise ValueError(F'Unexpected unexpected_keys: {unexpected_keys}' ) if set(SCREAMING_SNAKE_CASE_ ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(F'Unexpected missing_keys: {missing_keys}' ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs SCREAMING_SNAKE_CASE = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ , task='entity_classification' ) SCREAMING_SNAKE_CASE = 'ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan).' SCREAMING_SNAKE_CASE = (0, 9) SCREAMING_SNAKE_CASE = tokenizer(SCREAMING_SNAKE_CASE_ , entity_spans=[span] , return_tensors='pt' ) SCREAMING_SNAKE_CASE = model(**SCREAMING_SNAKE_CASE_ ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base SCREAMING_SNAKE_CASE = torch.Size((1, 33, 7_68) ) SCREAMING_SNAKE_CASE = torch.tensor([[0.08_92, 0.05_96, -0.28_19], [0.01_34, 0.11_99, 0.05_73], [-0.01_69, 0.09_27, 0.06_44]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F'Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=1E-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base SCREAMING_SNAKE_CASE = torch.Size((1, 1, 7_68) ) SCREAMING_SNAKE_CASE = torch.tensor([[-0.14_82, 0.06_09, 0.03_22]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( F'Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is' F' {expected_shape}' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=1E-4 ): raise ValueError # Verify masked word/entity prediction SCREAMING_SNAKE_CASE = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = 'Tokyo is the capital of <mask>.' SCREAMING_SNAKE_CASE = (24, 30) SCREAMING_SNAKE_CASE = tokenizer(SCREAMING_SNAKE_CASE_ , entity_spans=[span] , return_tensors='pt' ) SCREAMING_SNAKE_CASE = model(**SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = encoding['input_ids'][0].tolist() SCREAMING_SNAKE_CASE = input_ids.index(tokenizer.convert_tokens_to_ids('<mask>' ) ) SCREAMING_SNAKE_CASE = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = outputs.entity_logits[0][0].argmax().item() SCREAMING_SNAKE_CASE = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith('en:' )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print('Saving PyTorch model to {}'.format(SCREAMING_SNAKE_CASE_ ) ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) def lowercase (SCREAMING_SNAKE_CASE_ : int ) -> int: SCREAMING_SNAKE_CASE = ['[MASK]', '[PAD]', '[UNK]'] SCREAMING_SNAKE_CASE = [json.loads(SCREAMING_SNAKE_CASE_ ) for line in open(SCREAMING_SNAKE_CASE_ )] SCREAMING_SNAKE_CASE = {} for entry in data: SCREAMING_SNAKE_CASE = entry['id'] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: SCREAMING_SNAKE_CASE = entity_id break SCREAMING_SNAKE_CASE = F'{language}:{entity_name}' SCREAMING_SNAKE_CASE = entity_id return new_mapping if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) __UpperCamelCase = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
38
0
from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def _A ( SCREAMING_SNAKE_CASE : np.ndarray , SCREAMING_SNAKE_CASE : np.ndarray , SCREAMING_SNAKE_CASE : np.ndarray , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ): """simple docstring""" a__ : Any =cva.getAffineTransform(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return cva.warpAffine(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , (rows, cols) ) if __name__ == "__main__": # read original image UpperCAmelCase : Optional[Any] = cva.imread( str(Path(__file__).resolve().parent.parent / """image_data""" / """lena.jpg""") ) # turn image in gray scale value UpperCAmelCase : Tuple = cva.cvtColor(image, cva.COLOR_BGR2GRAY) # get image shape UpperCAmelCase , UpperCAmelCase : Optional[int] = gray_img.shape # set different points to rotate image UpperCAmelCase : Optional[Any] = np.array([[50, 50], [200, 50], [50, 200]], np.floataa) UpperCAmelCase : Optional[Any] = np.array([[10, 100], [200, 50], [100, 250]], np.floataa) UpperCAmelCase : Dict = np.array([[50, 50], [150, 50], [120, 200]], np.floataa) UpperCAmelCase : Any = np.array([[10, 100], [80, 50], [180, 250]], np.floataa) # add all rotated images in a list UpperCAmelCase : Optional[Any] = [ gray_img, get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), ] # plot different image rotations UpperCAmelCase : int = plt.figure(1) UpperCAmelCase : Any = ["""Original""", """Rotation 1""", """Rotation 2""", """Rotation 3"""] for i, image in enumerate(images): plt.subplot(2, 2, i + 1), plt.imshow(image, """gray""") plt.title(titles[i]) plt.axis("""off""") plt.subplots_adjust(left=0.0, bottom=0.0_5, right=1.0, top=0.9_5) plt.show()
95
"""simple docstring""" import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin SCREAMING_SNAKE_CASE_ = 1E-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class lowerCAmelCase_ : '''simple docstring''' def __init__( self , snake_case_ , snake_case_=16 , snake_case_=13 , snake_case_=7 , snake_case_=14 , snake_case_=10 , snake_case_=19 , snake_case_=5 , snake_case_=4 , snake_case_=True , snake_case_=16 , snake_case_=2 , snake_case_=4 , snake_case_=4 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=[1, 2, 3, 4, 5] , snake_case_=25 , snake_case_=5 , ) -> Tuple: __lowerCAmelCase = d_model __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = prediction_length __lowerCAmelCase = context_length __lowerCAmelCase = cardinality __lowerCAmelCase = num_time_features __lowerCAmelCase = lags_sequence __lowerCAmelCase = embedding_dimension __lowerCAmelCase = is_training __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = context_length __lowerCAmelCase = prediction_length + label_length __lowerCAmelCase = label_length __lowerCAmelCase = moving_average __lowerCAmelCase = autocorrelation_factor def A__ ( self ) -> List[Any]: return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def A__ ( self , snake_case_ ) -> Any: __lowerCAmelCase = config.context_length + max(config.lags_sequence ) __lowerCAmelCase = ids_tensor([self.batch_size, 1] , config.cardinality[0] ) __lowerCAmelCase = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) __lowerCAmelCase = floats_tensor([self.batch_size, _past_length] ) __lowerCAmelCase = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs __lowerCAmelCase = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) __lowerCAmelCase = floats_tensor([self.batch_size, config.prediction_length] ) __lowerCAmelCase = { """past_values""": past_values, """static_categorical_features""": static_categorical_features, """past_time_features""": past_time_features, """past_observed_mask""": past_observed_mask, """future_time_features""": future_time_features, """future_values""": future_values, } return inputs_dict def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.get_config() __lowerCAmelCase = self.prepare_autoformer_inputs_dict(snake_case_ ) return config, inputs_dict def A__ ( self ) -> int: __lowerCAmelCase , __lowerCAmelCase = self.prepare_config_and_inputs() return config, inputs_dict def A__ ( self , snake_case_ , snake_case_ ) -> int: __lowerCAmelCase = AutoformerModel(config=snake_case_ ).to(snake_case_ ).eval() __lowerCAmelCase = model(**snake_case_ ) __lowerCAmelCase = outputs.encoder_last_hidden_state __lowerCAmelCase = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = model.get_encoder() encoder.save_pretrained(snake_case_ ) __lowerCAmelCase = AutoformerEncoder.from_pretrained(snake_case_ ).to(snake_case_ ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = model.create_network_inputs(**snake_case_ ) __lowerCAmelCase , __lowerCAmelCase = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) __lowerCAmelCase = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) __lowerCAmelCase = encoder(inputs_embeds=snake_case_ )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) __lowerCAmelCase = ( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) __lowerCAmelCase = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) __lowerCAmelCase = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) __lowerCAmelCase = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = model.get_decoder() decoder.save_pretrained(snake_case_ ) __lowerCAmelCase = AutoformerDecoder.from_pretrained(snake_case_ ).to(snake_case_ ) __lowerCAmelCase = decoder( trend=snake_case_ , inputs_embeds=snake_case_ , encoder_hidden_states=snake_case_ , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class lowerCAmelCase_ ( A__ , A__ , unittest.TestCase ): '''simple docstring''' _snake_case = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () _snake_case = (AutoformerForPrediction,) if is_torch_available() else () _snake_case = {'''feature-extraction''': AutoformerModel} if is_torch_available() else {} _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = False def A__ ( self ) -> Optional[int]: __lowerCAmelCase = AutoformerModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ ) def A__ ( self ) -> Optional[int]: self.config_tester.run_common_tests() def A__ ( self ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(snake_case_ ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case_ ) __lowerCAmelCase , __lowerCAmelCase = model_class.from_pretrained(snake_case_ , output_loading_info=snake_case_ ) self.assertEqual(info["""missing_keys"""] , [] ) def A__ ( self ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*snake_case_ ) @unittest.skip(reason="""Model has no tokens embeddings""" ) def A__ ( self ) -> Any: pass def A__ ( self ) -> str: __lowerCAmelCase = inspect.signature(getattr(snake_case_ , """forward""" ) ) # The main input is the name of the argument after `self` __lowerCAmelCase = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , snake_case_ ) def A__ ( self ) -> Any: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(snake_case_ ) __lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase = [*signature.parameters.keys()] __lowerCAmelCase = [ """past_values""", """past_time_features""", """past_observed_mask""", """static_categorical_features""", """static_real_features""", """future_values""", """future_time_features""", ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append("""future_observed_mask""" ) expected_arg_names.extend( [ """decoder_attention_mask""", """head_mask""", """decoder_head_mask""", """cross_attn_head_mask""", """encoder_outputs""", """past_key_values""", """output_hidden_states""", """output_attentions""", """use_cache""", """return_dict""", ] ) self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) def A__ ( self ) -> Optional[int]: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase = True __lowerCAmelCase = getattr(self.model_tester , """seq_length""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """decoder_seq_length""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """encoder_seq_length""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """d_model""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """num_attention_heads""" , snake_case_ ) __lowerCAmelCase = d_model // num_attention_heads for model_class in self.all_model_classes: __lowerCAmelCase = True __lowerCAmelCase = False __lowerCAmelCase = True __lowerCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) __lowerCAmelCase = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __lowerCAmelCase = True __lowerCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) __lowerCAmelCase = outputs.encoder_attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) __lowerCAmelCase = len(snake_case_ ) __lowerCAmelCase = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(snake_case_ , snake_case_ ) # decoder attentions __lowerCAmelCase = outputs.decoder_attentions self.assertIsInstance(snake_case_ , (list, tuple) ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions __lowerCAmelCase = outputs.cross_attentions self.assertIsInstance(snake_case_ , (list, tuple) ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine __lowerCAmelCase = True __lowerCAmelCase = True __lowerCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(out_len + 2 , len(snake_case_ ) ) __lowerCAmelCase = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def A__ ( self ) -> int: super().test_retain_grad_hidden_states_attentions() def lowercase (_lowerCAmelCase="train-batch.pt" ): __lowerCAmelCase = hf_hub_download(repo_id="""hf-internal-testing/tourism-monthly-batch""" , filename=_lowerCAmelCase , repo_type="""dataset""" ) __lowerCAmelCase = torch.load(_lowerCAmelCase , map_location=_lowerCAmelCase ) return batch @require_torch @slow class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def A__ ( self ) -> int: __lowerCAmelCase = AutoformerModel.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(snake_case_ ) __lowerCAmelCase = prepare_batch() with torch.no_grad(): __lowerCAmelCase = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , future_values=batch["""future_values"""] , future_time_features=batch["""future_time_features"""] , )[0] __lowerCAmelCase = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , snake_case_ ) __lowerCAmelCase = torch.tensor( [[0.3_593, -1.3_398, 0.6_330], [0.2_279, 1.5_396, -0.1_792], [0.0_450, 1.3_225, -0.2_335]] , device=snake_case_ ) self.assertTrue(torch.allclose(output[0, :3, :3] , snake_case_ , atol=snake_case_ ) ) def A__ ( self ) -> List[str]: __lowerCAmelCase = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(snake_case_ ) __lowerCAmelCase = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): __lowerCAmelCase = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , ).encoder_last_hidden_state __lowerCAmelCase = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , snake_case_ ) __lowerCAmelCase = torch.tensor( [[-0.0_734, -0.9_036, 0.8_358], [4.7_186, 2.4_113, 1.9_581], [1.7_953, 2.3_558, 1.2_970]] , device=snake_case_ ) self.assertTrue(torch.allclose(output[0, :3, :3] , snake_case_ , atol=snake_case_ ) ) def A__ ( self ) -> Any: __lowerCAmelCase = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(snake_case_ ) __lowerCAmelCase = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): __lowerCAmelCase = model.generate( static_categorical_features=batch["""static_categorical_features"""] , past_time_features=batch["""past_time_features"""] , past_values=batch["""past_values"""] , future_time_features=batch["""future_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , ) __lowerCAmelCase = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , snake_case_ ) __lowerCAmelCase = torch.tensor([3_130.6_763, 4_056.5_293, 7_053.0_786] , device=snake_case_ ) __lowerCAmelCase = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , snake_case_ , rtol=1e-1 ) )
301
0
'''simple docstring''' import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCamelCase : Tuple = logging.get_logger(__name__) lowerCamelCase : Dict = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} lowerCamelCase : List[Any] = { "tokenizer_file": { "EleutherAI/gpt-neox-20b": "https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json", }, } lowerCamelCase : List[Any] = { "gpt-neox-20b": 2_0_4_8, } class A__ ( A__ ): A__ = VOCAB_FILES_NAMES A__ = PRETRAINED_VOCAB_FILES_MAP A__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ = ['input_ids', 'attention_mask'] def __init__( self : Any , _a : List[str]=None , _a : Any=None , _a : Tuple=None , _a : Union[str, Any]="<|endoftext|>" , _a : str="<|endoftext|>" , _a : List[str]="<|endoftext|>" , _a : List[Any]=False , **_a : Optional[Any] , ) -> int: '''simple docstring''' super().__init__( _a , _a , tokenizer_file=_a , unk_token=_a , bos_token=_a , eos_token=_a , add_prefix_space=_a , **_a , ) _SCREAMING_SNAKE_CASE =json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , _a ) != add_prefix_space: _SCREAMING_SNAKE_CASE =getattr(_a , pre_tok_state.pop('type' ) ) _SCREAMING_SNAKE_CASE =add_prefix_space _SCREAMING_SNAKE_CASE =pre_tok_class(**_a ) _SCREAMING_SNAKE_CASE =add_prefix_space def A ( self : Union[str, Any] , _a : str , _a : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' _SCREAMING_SNAKE_CASE =self._tokenizer.model.save(_a , name=_a ) return tuple(_a ) def A ( self : Union[str, Any] , _a : "Conversation" ) -> List[int]: '''simple docstring''' _SCREAMING_SNAKE_CASE =[] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(_a , add_special_tokens=_a ) + [self.eos_token_id] ) if len(_a ) > self.model_max_length: _SCREAMING_SNAKE_CASE =input_ids[-self.model_max_length :] return input_ids
114
'''simple docstring''' from ..utils import DummyObject, requires_backends class A__ ( metaclass=A__ ): A__ = ['note_seq'] def __init__( self : List[str] , *_a : Any , **_a : Dict ) -> Optional[Any]: '''simple docstring''' requires_backends(self , ['note_seq'] ) @classmethod def A ( cls : Any , *_a : str , **_a : List[Any] ) -> List[str]: '''simple docstring''' requires_backends(cls , ['note_seq'] ) @classmethod def A ( cls : int , *_a : Optional[Any] , **_a : Optional[int] ) -> Tuple: '''simple docstring''' requires_backends(cls , ['note_seq'] )
114
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available _lowercase : List[str] ={"tokenization_herbert": ["HerbertTokenizer"]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Optional[int] =["HerbertTokenizerFast"] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys _lowercase : Optional[Any] =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
170
"""simple docstring""" import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging a = logging.get_logger(__name__) class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : Any = ['''input_values''', '''attention_mask'''] def __init__( self : Dict , _UpperCAmelCase : int = 1 , _UpperCAmelCase : int = 16_000 , _UpperCAmelCase : float = 0.0 , _UpperCAmelCase : bool = False , _UpperCAmelCase : int = 80 , _UpperCAmelCase : int = 16 , _UpperCAmelCase : int = 64 , _UpperCAmelCase : str = "hann_window" , _UpperCAmelCase : float = 1.0 , _UpperCAmelCase : float = 80 , _UpperCAmelCase : float = 7_600 , _UpperCAmelCase : float = 1E-1_0 , _UpperCAmelCase : int = 2 , _UpperCAmelCase : bool = True , **_UpperCAmelCase : List[Any] , ): super().__init__(feature_size=_UpperCAmelCase , sampling_rate=_UpperCAmelCase , padding_value=_UpperCAmelCase , **_UpperCAmelCase ) _A = do_normalize _A = return_attention_mask _A = num_mel_bins _A = hop_length _A = win_length _A = win_function _A = frame_signal_scale _A = fmin _A = fmax _A = mel_floor _A = reduction_factor _A = win_length * sampling_rate // 1_000 _A = hop_length * sampling_rate // 1_000 _A = optimal_fft_length(self.sample_size ) _A = (self.n_fft // 2) + 1 _A = window_function(window_length=self.sample_size , name=self.win_function , periodic=_UpperCAmelCase ) _A = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm='slaney' , mel_scale='slaney' , ) if frame_signal_scale != 1.0: warnings.warn( 'The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers' , _UpperCAmelCase , ) if reduction_factor != 2.0: warnings.warn( 'The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers' , _UpperCAmelCase , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def lowerCAmelCase_ ( _UpperCAmelCase : List[np.ndarray] , _UpperCAmelCase : List[np.ndarray] , _UpperCAmelCase : float = 0.0 ): if attention_mask is not None: _A = np.array(_UpperCAmelCase , np.intaa ) _A = [] for vector, length in zip(_UpperCAmelCase , attention_mask.sum(-1 ) ): _A = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: _A = padding_value normed_input_values.append(_UpperCAmelCase ) else: _A = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : np.ndarray , ): _A = spectrogram( _UpperCAmelCase , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel='log10' , ) return log_mel_spec.T def __call__( self : int , _UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , _UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , _UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , _UpperCAmelCase : Optional[int] = None , _UpperCAmelCase : bool = False , _UpperCAmelCase : Optional[int] = None , _UpperCAmelCase : Optional[bool] = None , _UpperCAmelCase : Optional[Union[str, TensorType]] = None , _UpperCAmelCase : Optional[int] = None , **_UpperCAmelCase : Optional[int] , ): if audio is None and audio_target is None: raise ValueError('You must provide either `audio` or `audio_target` values.' ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided audio input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the ``sampling_rate`` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) if audio is not None: _A = self._process_audio( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase , ) else: _A = None if audio_target is not None: _A = self._process_audio( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase , ) if inputs is None: return inputs_target else: _A = inputs_target['input_values'] _A = inputs_target.get('attention_mask' ) if decoder_attention_mask is not None: _A = decoder_attention_mask return inputs def lowerCAmelCase_ ( self : Optional[int] , _UpperCAmelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , _UpperCAmelCase : bool = False , _UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , _UpperCAmelCase : Optional[int] = None , _UpperCAmelCase : bool = False , _UpperCAmelCase : Optional[int] = None , _UpperCAmelCase : Optional[bool] = None , _UpperCAmelCase : Optional[Union[str, TensorType]] = None , **_UpperCAmelCase : List[Any] , ): _A = isinstance(_UpperCAmelCase , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) _A = is_batched_numpy or ( isinstance(_UpperCAmelCase , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: _A = [np.asarray(_UpperCAmelCase , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(_UpperCAmelCase , np.ndarray ): _A = np.asarray(_UpperCAmelCase , dtype=np.floataa ) elif isinstance(_UpperCAmelCase , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): _A = speech.astype(np.floataa ) # always return batch if not is_batched: _A = [speech] # needed to make pad() work on spectrogram inputs _A = self.feature_size # convert into correct format for padding if is_target: _A = [self._extract_mel_features(_UpperCAmelCase ) for waveform in speech] _A = BatchFeature({'input_values': features} ) _A = self.num_mel_bins else: _A = BatchFeature({'input_values': speech} ) _A = self.pad( _UpperCAmelCase , padding=_UpperCAmelCase , max_length=_UpperCAmelCase , truncation=_UpperCAmelCase , pad_to_multiple_of=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase , **_UpperCAmelCase , ) _A = feature_size_hack # convert input values to correct format _A = padded_inputs['input_values'] if not isinstance(input_values[0] , np.ndarray ): _A = [np.asarray(_UpperCAmelCase , dtype=np.floataa ) for array in input_values] elif ( not isinstance(_UpperCAmelCase , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): _A = [array.astype(np.floataa ) for array in input_values] elif isinstance(_UpperCAmelCase , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): _A = input_values.astype(np.floataa ) # convert attention_mask to correct format _A = padded_inputs.get('attention_mask' ) if attention_mask is not None: _A = [np.asarray(_UpperCAmelCase , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: _A = ( attention_mask if self._get_padding_strategies(_UpperCAmelCase , max_length=_UpperCAmelCase ) is not PaddingStrategy.DO_NOT_PAD else None ) _A = self.zero_mean_unit_var_norm( padded_inputs['input_values'] , attention_mask=_UpperCAmelCase , padding_value=self.padding_value ) if return_tensors is not None: _A = padded_inputs.convert_to_tensors(_UpperCAmelCase ) return padded_inputs def lowerCAmelCase_ ( self : Any ): _A = super().to_dict() # Don't serialize these as they are derived from the other properties. _A = ['window', 'mel_filters', 'sample_size', 'sample_stride', 'n_fft', 'n_freqs'] for name in names: if name in output: del output[name] return output
315
0
import unittest from transformers.testing_utils import require_bsa from transformers.utils import is_bsa_available from ...test_feature_extraction_common import FeatureExtractionSavingTestMixin if is_bsa_available(): from transformers import MarkupLMFeatureExtractor class _lowerCAmelCase( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase ): UpperCamelCase_: int = parent def _a ( self ): return {} def snake_case () -> List[str]: UpperCamelCase_: Any = '<HTML>\n\n <HEAD>\n <TITLE>sample document</TITLE>\n </HEAD>\n\n <BODY BGCOLOR="FFFFFF">\n <HR>\n <a href="http://google.com">Goog</a>\n <H1>This is one header</H1>\n <H2>This is a another Header</H2>\n <P>Travel from\n <P>\n <B>SFO to JFK</B>\n <BR>\n <B><I>on May 2, 2015 at 2:00 pm. For details go to confirm.com </I></B>\n <HR>\n <div style="color:#0000FF">\n <h3>Traveler <b> name </b> is\n <p> John Doe </p>\n </div>' UpperCamelCase_: Dict = '\n <!DOCTYPE html>\n <html>\n <body>\n\n <h1>My First Heading</h1>\n <p>My first paragraph.</p>\n\n </body>\n </html>\n ' return [html_string_a, html_string_a] @require_bsa class _lowerCAmelCase( UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" a : Union[str, Any] =MarkupLMFeatureExtractor if is_bsa_available() else None def _a ( self ): UpperCamelCase_: int = MarkupLMFeatureExtractionTester(self ) @property def _a ( self ): return self.feature_extract_tester.prepare_feat_extract_dict() def _a ( self ): # Initialize feature_extractor UpperCamelCase_: Optional[int] = self.feature_extraction_class() # Test not batched input UpperCamelCase_: Any = get_html_strings()[0] UpperCamelCase_: int = feature_extractor(_lowerCamelCase ) # fmt: off UpperCamelCase_: str = [['sample document', 'Goog', 'This is one header', 'This is a another Header', 'Travel from', 'SFO to JFK', 'on May 2, 2015 at 2:00 pm. For details go to confirm.com', 'Traveler', 'name', 'is', 'John Doe']] UpperCamelCase_: List[Any] = [['/html/head/title', '/html/body/a', '/html/body/h1', '/html/body/h2', '/html/body/p', '/html/body/p/p/b[1]', '/html/body/p/p/b[2]/i', '/html/body/p/p/div/h3', '/html/body/p/p/div/h3/b', '/html/body/p/p/div/h3', '/html/body/p/p/div/h3/p']] # fmt: on self.assertEqual(encoding.nodes , _lowerCamelCase ) self.assertEqual(encoding.xpaths , _lowerCamelCase ) # Test batched UpperCamelCase_: Union[str, Any] = get_html_strings() UpperCamelCase_: Optional[int] = feature_extractor(_lowerCamelCase ) # fmt: off UpperCamelCase_: Tuple = expected_nodes + [['My First Heading', 'My first paragraph.']] UpperCamelCase_: List[Any] = expected_xpaths + [['/html/body/h1', '/html/body/p']] self.assertEqual(len(encoding.nodes ) , 2 ) self.assertEqual(len(encoding.xpaths ) , 2 ) self.assertEqual(encoding.nodes , _lowerCamelCase ) self.assertEqual(encoding.xpaths , _lowerCamelCase )
292
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def snake_case (UpperCAmelCase__ , UpperCAmelCase__ ) -> np.array: UpperCamelCase_: Dict = F'''{sampling_rate}''' UpperCamelCase_: Any = '1' UpperCamelCase_: Any = 'f32le' UpperCamelCase_: Union[str, Any] = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(UpperCAmelCase__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: UpperCamelCase_: Optional[Any] = ffmpeg_process.communicate(UpperCAmelCase__ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error UpperCamelCase_: Union[str, Any] = output_stream[0] UpperCamelCase_: List[str] = np.frombuffer(UpperCAmelCase__ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = "f32le" , ) -> Tuple: UpperCamelCase_: Any = F'''{sampling_rate}''' UpperCamelCase_: Union[str, Any] = '1' if format_for_conversion == "s16le": UpperCamelCase_: Optional[Any] = 2 elif format_for_conversion == "f32le": UpperCamelCase_: Any = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) UpperCamelCase_: int = platform.system() if system == "Linux": UpperCamelCase_: Tuple = 'alsa' UpperCamelCase_: List[str] = 'default' elif system == "Darwin": UpperCamelCase_: int = 'avfoundation' UpperCamelCase_: Union[str, Any] = ':0' elif system == "Windows": UpperCamelCase_: Tuple = 'dshow' UpperCamelCase_: Dict = 'default' UpperCamelCase_: Any = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] UpperCamelCase_: Tuple = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample UpperCamelCase_: Optional[int] = _ffmpeg_stream(UpperCAmelCase__ , UpperCAmelCase__ ) for item in iterator: yield item def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = None , UpperCAmelCase__ = None , UpperCAmelCase__ = "f32le" , ) -> Any: if stream_chunk_s is not None: UpperCamelCase_: List[Any] = stream_chunk_s else: UpperCamelCase_: Dict = chunk_length_s UpperCamelCase_: List[str] = ffmpeg_microphone(UpperCAmelCase__ , UpperCAmelCase__ , format_for_conversion=UpperCAmelCase__ ) if format_for_conversion == "s16le": UpperCamelCase_: Union[str, Any] = np.intaa UpperCamelCase_: List[Any] = 2 elif format_for_conversion == "f32le": UpperCamelCase_: str = np.floataa UpperCamelCase_: Tuple = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: UpperCamelCase_: int = chunk_length_s / 6 UpperCamelCase_: Tuple = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(UpperCAmelCase__ , (int, float) ): UpperCamelCase_: Union[str, Any] = [stride_length_s, stride_length_s] UpperCamelCase_: Any = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample UpperCamelCase_: Dict = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample UpperCamelCase_: Optional[int] = datetime.datetime.now() UpperCamelCase_: Optional[int] = datetime.timedelta(seconds=UpperCAmelCase__ ) for item in chunk_bytes_iter(UpperCAmelCase__ , UpperCAmelCase__ , stride=(stride_left, stride_right) , stream=UpperCAmelCase__ ): # Put everything back in numpy scale UpperCamelCase_: Tuple = np.frombuffer(item['raw'] , dtype=UpperCAmelCase__ ) UpperCamelCase_: Optional[int] = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) UpperCamelCase_: int = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 1_0 * delta: # We're late !! SKIP continue yield item def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = False ) -> int: UpperCamelCase_: str = b'' UpperCamelCase_ ,UpperCamelCase_: Union[str, Any] = stride if stride_left + stride_right >= chunk_len: raise ValueError( F'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) UpperCamelCase_: List[str] = 0 for raw in iterator: acc += raw if stream and len(UpperCAmelCase__ ) < chunk_len: UpperCamelCase_: Optional[Any] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(UpperCAmelCase__ ) >= chunk_len: # We are flushing the accumulator UpperCamelCase_: int = (_stride_left, stride_right) UpperCamelCase_: Optional[Any] = {'raw': acc[:chunk_len], 'stride': stride} if stream: UpperCamelCase_: Any = False yield item UpperCamelCase_: Optional[int] = stride_left UpperCamelCase_: Optional[Any] = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(UpperCAmelCase__ ) > stride_left: UpperCamelCase_: int = {'raw': acc, 'stride': (_stride_left, 0)} if stream: UpperCamelCase_: Optional[Any] = False yield item def snake_case (UpperCAmelCase__ , UpperCAmelCase__ ) -> int: UpperCamelCase_: Any = 2**2_4 # 16Mo try: with subprocess.Popen(UpperCAmelCase__ , stdout=subprocess.PIPE , bufsize=UpperCAmelCase__ ) as ffmpeg_process: while True: UpperCamelCase_: Any = ffmpeg_process.stdout.read(UpperCAmelCase__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
292
1
import logging import math from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence, Tuple, Union import torch from .tensor_utils import tensor_tree_map, tree_map def _a ( SCREAMING_SNAKE_CASE : Union[dict, list, tuple, torch.Tensor] ) -> List[Tuple[int, ...]]: """simple docstring""" __lowerCAmelCase: List[str] = [] if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): for v in tree.values(): shapes.extend(_fetch_dims(SCREAMING_SNAKE_CASE ) ) elif isinstance(SCREAMING_SNAKE_CASE , (list, tuple) ): for t in tree: shapes.extend(_fetch_dims(SCREAMING_SNAKE_CASE ) ) elif isinstance(SCREAMING_SNAKE_CASE , torch.Tensor ): shapes.append(tree.shape ) else: raise ValueError('Not supported' ) return shapes @torch.jit.ignore def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Tuple[int, ...] ) -> Tuple[int, ...]: """simple docstring""" __lowerCAmelCase: Union[str, Any] = [] for d in reversed(SCREAMING_SNAKE_CASE ): idx.append(flat_idx % d ) __lowerCAmelCase: Any = flat_idx // d return tuple(reversed(SCREAMING_SNAKE_CASE ) ) @torch.jit.ignore def _a ( SCREAMING_SNAKE_CASE : Sequence[int] , SCREAMING_SNAKE_CASE : Sequence[int] , SCREAMING_SNAKE_CASE : Sequence[int] , SCREAMING_SNAKE_CASE : Optional[Sequence[bool]] = None , SCREAMING_SNAKE_CASE : Optional[Sequence[bool]] = None , ) -> List[Tuple[slice, ...]]: """simple docstring""" def reduce_edge_list(SCREAMING_SNAKE_CASE : List[bool] ) -> None: __lowerCAmelCase: Dict = True for i in range(len(SCREAMING_SNAKE_CASE ) ): __lowerCAmelCase: int = -1 * (i + 1) l[reversed_idx] &= tally __lowerCAmelCase: List[Any] = l[reversed_idx] if start_edges is None: __lowerCAmelCase: Optional[int] = [s == 0 for s in start] reduce_edge_list(SCREAMING_SNAKE_CASE ) if end_edges is None: __lowerCAmelCase: Optional[Any] = [e == (d - 1) for e, d in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )] reduce_edge_list(SCREAMING_SNAKE_CASE ) # Base cases. Either start/end are empty and we're done, or the final, # one-dimensional tensor can be simply sliced if len(SCREAMING_SNAKE_CASE ) == 0: return [()] elif len(SCREAMING_SNAKE_CASE ) == 1: return [(slice(start[0] , end[0] + 1 ),)] __lowerCAmelCase: List[Tuple[slice, ...]] = [] __lowerCAmelCase: List[slice] = [] # Dimensions common to start and end can be selected directly for s, e in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): if s == e: path_list.append(slice(SCREAMING_SNAKE_CASE , s + 1 ) ) else: break __lowerCAmelCase: Tuple[slice, ...] = tuple(SCREAMING_SNAKE_CASE ) __lowerCAmelCase: Any = len(SCREAMING_SNAKE_CASE ) # start == end, and we're done if divergence_idx == len(SCREAMING_SNAKE_CASE ): return [path] def upper() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None __lowerCAmelCase: Union[str, Any] = start[divergence_idx] return tuple( path + (slice(SCREAMING_SNAKE_CASE , sdi + 1 ),) + s for s in _get_minimal_slice_set( start[divergence_idx + 1 :] , [d - 1 for d in dims[divergence_idx + 1 :]] , dims[divergence_idx + 1 :] , start_edges=start_edges[divergence_idx + 1 :] , end_edges=[True for _ in end_edges[divergence_idx + 1 :]] , ) ) def lower() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None __lowerCAmelCase: int = end[divergence_idx] return tuple( path + (slice(SCREAMING_SNAKE_CASE , edi + 1 ),) + s for s in _get_minimal_slice_set( [0 for _ in start[divergence_idx + 1 :]] , end[divergence_idx + 1 :] , dims[divergence_idx + 1 :] , start_edges=[True for _ in start_edges[divergence_idx + 1 :]] , end_edges=end_edges[divergence_idx + 1 :] , ) ) # If both start and end are at the edges of the subtree rooted at # divergence_idx, we can just select the whole subtree at once if start_edges[divergence_idx] and end_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] + 1 ),) ) # If just start is at the edge, we can grab almost all of the subtree, # treating only the ragged bottom edge as an edge case elif start_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] , end[divergence_idx] ),) ) slices.extend(lower() ) # Analogous to the previous case, but the top is ragged this time elif end_edges[divergence_idx]: slices.extend(upper() ) slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] + 1 ),) ) # If both sides of the range are ragged, we need to handle both sides # separately. If there's contiguous meat in between them, we can index it # in one big chunk else: slices.extend(upper() ) __lowerCAmelCase: List[Any] = end[divergence_idx] - start[divergence_idx] if middle_ground > 1: slices.append(path + (slice(start[divergence_idx] + 1 , end[divergence_idx] ),) ) slices.extend(lower() ) return slices @torch.jit.ignore def _a ( SCREAMING_SNAKE_CASE : torch.Tensor , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ) -> torch.Tensor: """simple docstring""" __lowerCAmelCase: Optional[Any] = t.shape[:no_batch_dims] __lowerCAmelCase: Any = list(_flat_idx_to_idx(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) # _get_minimal_slice_set is inclusive __lowerCAmelCase: Optional[int] = list(_flat_idx_to_idx(flat_end - 1 , SCREAMING_SNAKE_CASE ) ) # Get an ordered list of slices to perform __lowerCAmelCase: Tuple = _get_minimal_slice_set( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) __lowerCAmelCase: Optional[int] = [t[s] for s in slices] return torch.cat([s.view((-1,) + t.shape[no_batch_dims:] ) for s in sliced_tensors] ) def _a ( SCREAMING_SNAKE_CASE : Callable , SCREAMING_SNAKE_CASE : Dict[str, Any] , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : bool = False , SCREAMING_SNAKE_CASE : Any = None , SCREAMING_SNAKE_CASE : bool = False , ) -> Any: """simple docstring""" if not (len(SCREAMING_SNAKE_CASE ) > 0): raise ValueError('Must provide at least one input' ) __lowerCAmelCase: int = [shape[:no_batch_dims] for shape in _fetch_dims(SCREAMING_SNAKE_CASE )] __lowerCAmelCase: Union[str, Any] = tuple([max(SCREAMING_SNAKE_CASE ) for s in zip(*SCREAMING_SNAKE_CASE )] ) def _prep_inputs(SCREAMING_SNAKE_CASE : torch.Tensor ) -> torch.Tensor: if not low_mem: if not sum(t.shape[:no_batch_dims] ) == no_batch_dims: __lowerCAmelCase: Union[str, Any] = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) __lowerCAmelCase: Tuple = t.reshape(-1 , *t.shape[no_batch_dims:] ) else: __lowerCAmelCase: Optional[Any] = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) return t __lowerCAmelCase: Dict[str, Any] = tensor_tree_map(_prep_inputs , SCREAMING_SNAKE_CASE ) __lowerCAmelCase: Optional[Any] = None if _out is not None: __lowerCAmelCase: List[str] = tensor_tree_map(lambda SCREAMING_SNAKE_CASE : t.view([-1] + list(t.shape[no_batch_dims:] ) ) , _out ) __lowerCAmelCase: Any = 1 for d in orig_batch_dims: flat_batch_dim *= d __lowerCAmelCase: List[str] = flat_batch_dim // chunk_size + (flat_batch_dim % chunk_size != 0) def _select_chunk(SCREAMING_SNAKE_CASE : torch.Tensor ) -> torch.Tensor: return t[i : i + chunk_size] if t.shape[0] != 1 else t __lowerCAmelCase: Any = 0 __lowerCAmelCase: Optional[int] = prepped_outputs for _ in range(SCREAMING_SNAKE_CASE ): # Chunk the input if not low_mem: __lowerCAmelCase: str = _select_chunk else: __lowerCAmelCase: Dict = partial( _chunk_slice , flat_start=SCREAMING_SNAKE_CASE , flat_end=min(SCREAMING_SNAKE_CASE , i + chunk_size ) , no_batch_dims=len(SCREAMING_SNAKE_CASE ) , ) __lowerCAmelCase: Dict[str, Any] = tensor_tree_map(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Run the layer on the chunk __lowerCAmelCase: Optional[int] = layer(**SCREAMING_SNAKE_CASE ) # Allocate space for the output if out is None: __lowerCAmelCase: Optional[int] = tensor_tree_map(lambda SCREAMING_SNAKE_CASE : t.new_zeros((flat_batch_dim,) + t.shape[1:] ) , SCREAMING_SNAKE_CASE ) # Put the chunk in its pre-allocated space if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): def assign(SCREAMING_SNAKE_CASE : dict , SCREAMING_SNAKE_CASE : dict ) -> None: for k, v in da.items(): if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): assign(SCREAMING_SNAKE_CASE , da[k] ) else: if _add_into_out: v[i : i + chunk_size] += da[k] else: __lowerCAmelCase: Union[str, Any] = da[k] assign(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) elif isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): for xa, xa in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): if _add_into_out: xa[i : i + chunk_size] += xa else: __lowerCAmelCase: Dict = xa elif isinstance(SCREAMING_SNAKE_CASE , torch.Tensor ): if _add_into_out: out[i : i + chunk_size] += output_chunk else: __lowerCAmelCase: Dict = output_chunk else: raise ValueError('Not supported' ) i += chunk_size __lowerCAmelCase: Union[str, Any] = tensor_tree_map(lambda SCREAMING_SNAKE_CASE : t.view(orig_batch_dims + t.shape[1:] ) , SCREAMING_SNAKE_CASE ) return out class A_ : def __init__( self : Union[str, Any] , UpperCAmelCase : int = 5_1_2 , ) -> str: __lowerCAmelCase: str = max_chunk_size __lowerCAmelCase: Optional[int] = None __lowerCAmelCase: Optional[tuple] = None def UpperCAmelCase ( self : List[Any] , UpperCAmelCase : Callable , UpperCAmelCase : tuple , UpperCAmelCase : int ) -> int: logging.info('Tuning chunk size...' ) if min_chunk_size >= self.max_chunk_size: return min_chunk_size __lowerCAmelCase: List[int] = [2**l for l in range(int(math.log(self.max_chunk_size , 2 ) ) + 1 )] __lowerCAmelCase: Tuple = [c for c in candidates if c > min_chunk_size] __lowerCAmelCase: int = [min_chunk_size] + candidates candidates[-1] += 4 def test_chunk_size(UpperCAmelCase : int ) -> bool: try: with torch.no_grad(): fn(*UpperCAmelCase , chunk_size=UpperCAmelCase ) return True except RuntimeError: return False __lowerCAmelCase: Optional[Any] = 0 __lowerCAmelCase: Any = len(UpperCAmelCase ) - 1 while i > min_viable_chunk_size_index: __lowerCAmelCase: Tuple = test_chunk_size(candidates[i] ) if not viable: __lowerCAmelCase: Dict = (min_viable_chunk_size_index + i) // 2 else: __lowerCAmelCase: Optional[int] = i __lowerCAmelCase: Union[str, Any] = (i + len(UpperCAmelCase ) - 1) // 2 return candidates[min_viable_chunk_size_index] def UpperCAmelCase ( self : Optional[int] , UpperCAmelCase : Iterable , UpperCAmelCase : Iterable ) -> bool: __lowerCAmelCase: Tuple = True for aa, aa in zip(UpperCAmelCase , UpperCAmelCase ): assert type(UpperCAmelCase ) == type(UpperCAmelCase ) if isinstance(UpperCAmelCase , (list, tuple) ): consistent &= self._compare_arg_caches(UpperCAmelCase , UpperCAmelCase ) elif isinstance(UpperCAmelCase , UpperCAmelCase ): __lowerCAmelCase: Union[str, Any] = [v for _, v in sorted(aa.items() , key=lambda UpperCAmelCase : x[0] )] __lowerCAmelCase: Tuple = [v for _, v in sorted(aa.items() , key=lambda UpperCAmelCase : x[0] )] consistent &= self._compare_arg_caches(UpperCAmelCase , UpperCAmelCase ) else: consistent &= aa == aa return consistent def UpperCAmelCase ( self : Dict , UpperCAmelCase : Callable , UpperCAmelCase : tuple , UpperCAmelCase : int , ) -> int: __lowerCAmelCase: List[str] = True __lowerCAmelCase: tuple = tree_map(lambda UpperCAmelCase : a.shape if isinstance(UpperCAmelCase , torch.Tensor ) else a , UpperCAmelCase , UpperCAmelCase ) if self.cached_arg_data is not None: # If args have changed shape/value, we need to re-tune assert len(self.cached_arg_data ) == len(UpperCAmelCase ) __lowerCAmelCase: Union[str, Any] = self._compare_arg_caches(self.cached_arg_data , UpperCAmelCase ) else: # Otherwise, we can reuse the precomputed value __lowerCAmelCase: str = False if not consistent: __lowerCAmelCase: Dict = self._determine_favorable_chunk_size( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , ) __lowerCAmelCase: Tuple = arg_data assert self.cached_chunk_size is not None return self.cached_chunk_size
322
def _a ( SCREAMING_SNAKE_CASE : str ) -> Union[str, Any]: """simple docstring""" __lowerCAmelCase: str = len(SCREAMING_SNAKE_CASE ) __lowerCAmelCase: List[Any] = sum(SCREAMING_SNAKE_CASE ) __lowerCAmelCase: str = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): __lowerCAmelCase: Tuple = True for i in range(1 , s + 1 ): __lowerCAmelCase: Any = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): __lowerCAmelCase: Optional[int] = dp[i][j - 1] if arr[i - 1] <= j: __lowerCAmelCase: Union[str, Any] = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: __lowerCAmelCase: Tuple = s - 2 * j break return diff
322
1
import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin __UpperCamelCase : int = get_tests_dir("fixtures/test_sentencepiece_bpe_char.model") @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( __magic_name__ , unittest.TestCase ): UpperCamelCase__ = SpeechTaTokenizer UpperCamelCase__ = False UpperCamelCase__ = True def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing a = SpeechTaTokenizer(__magic_name__ ) a = AddedToken("""<mask>""" , lstrip=__magic_name__ , rstrip=__magic_name__ ) a = mask_token tokenizer.add_special_tokens({"""mask_token""": mask_token} ) tokenizer.add_tokens(["""<ctc_blank>"""] ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase__ ( self :Optional[int] , __magic_name__ :Any ): '''simple docstring''' a = """this is a test""" a = """this is a test""" return input_text, output_text def lowerCamelCase__ ( self :Optional[int] , __magic_name__ :Tuple , __magic_name__ :Optional[Any]=False , __magic_name__ :Any=20 , __magic_name__ :int=5 ): '''simple docstring''' a , a = self.get_input_output_texts(__magic_name__ ) a = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) a = tokenizer.decode(__magic_name__ , clean_up_tokenization_spaces=__magic_name__ ) return text, ids def lowerCamelCase__ ( self :str ): '''simple docstring''' a = """<pad>""" a = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__magic_name__ ) , __magic_name__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__magic_name__ ) , __magic_name__ ) def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' a = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-4] , """œ""" ) self.assertEqual(vocab_keys[-2] , """<mask>""" ) self.assertEqual(vocab_keys[-1] , """<ctc_blank>""" ) self.assertEqual(len(__magic_name__ ) , 81 ) def lowerCamelCase__ ( self :Optional[int] ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 79 ) def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' a = self.get_tokenizers(do_lower_case=__magic_name__ ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): a = tokenizer.vocab_size a = len(__magic_name__ ) self.assertNotEqual(__magic_name__ , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) a = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] a = tokenizer.add_tokens(__magic_name__ ) a = tokenizer.vocab_size a = len(__magic_name__ ) self.assertNotEqual(__magic_name__ , 0 ) self.assertEqual(__magic_name__ , __magic_name__ ) self.assertEqual(__magic_name__ , len(__magic_name__ ) ) self.assertEqual(__magic_name__ , all_size + len(__magic_name__ ) ) a = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=__magic_name__ ) self.assertGreaterEqual(len(__magic_name__ ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) a = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} a = tokenizer.add_special_tokens(__magic_name__ ) a = tokenizer.vocab_size a = len(__magic_name__ ) self.assertNotEqual(__magic_name__ , 0 ) self.assertEqual(__magic_name__ , __magic_name__ ) self.assertEqual(__magic_name__ , len(__magic_name__ ) ) self.assertEqual(__magic_name__ , all_size_a + len(__magic_name__ ) ) a = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=__magic_name__ ) self.assertGreaterEqual(len(__magic_name__ ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) def lowerCamelCase__ ( self :str ): '''simple docstring''' pass def lowerCamelCase__ ( self :Any ): '''simple docstring''' pass def lowerCamelCase__ ( self :str ): '''simple docstring''' a = self.get_tokenizer() a = tokenizer.tokenize("""This is a test""" ) # fmt: off self.assertListEqual(__magic_name__ , [SPIECE_UNDERLINE, """T""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """a""", SPIECE_UNDERLINE, """t""", """e""", """s""", """t"""] ) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(__magic_name__ ) , [4, 32, 11, 10, 12, 4, 10, 12, 4, 7, 4, 6, 5, 12, 6] , ) a = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( __magic_name__ , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """92000""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) a = tokenizer.convert_tokens_to_ids(__magic_name__ ) # fmt: off self.assertListEqual(__magic_name__ , [4, 30, 4, 20, 7, 12, 4, 25, 8, 13, 9, 4, 10, 9, 4, 3, 23, 4, 7, 9, 14, 4, 6, 11, 10, 12, 4, 10, 12, 4, 19, 7, 15, 12, 73, 26] ) # fmt: on a = tokenizer.convert_ids_to_tokens(__magic_name__ ) self.assertListEqual( __magic_name__ , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """<unk>""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) @slow def lowerCamelCase__ ( self :Any ): '''simple docstring''' a = [ """Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides """ """general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural """ """Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained """ """models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.""", """BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly """ """conditioning on both left and right context in all layers.""", """The quick brown fox jumps over the lazy dog.""", ] # fmt: off a = { """input_ids""": [ [4, 32, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 64, 19, 8, 13, 18, 5, 13, 15, 22, 4, 28, 9, 8, 20, 9, 4, 7, 12, 4, 24, 22, 6, 8, 13, 17, 11, 39, 6, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 7, 9, 14, 4, 24, 22, 6, 8, 13, 17, 11, 39, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 39, 25, 5, 13, 6, 63, 4, 24, 13, 8, 27, 10, 14, 5, 12, 4, 21, 5, 9, 5, 13, 7, 15, 39, 24, 16, 13, 24, 8, 12, 5, 4, 7, 13, 17, 11, 10, 6, 5, 17, 6, 16, 13, 5, 12, 4, 64, 40, 47, 54, 32, 23, 4, 53, 49, 32, 23, 4, 54, 8, 40, 47, 54, 32, 7, 23, 4, 69, 52, 43, 23, 4, 51, 10, 12, 6, 10, 15, 40, 5, 13, 6, 23, 4, 69, 52, 48, 5, 6, 26, 26, 26, 63, 4, 19, 8, 13, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 61, 9, 14, 5, 13, 12, 6, 7, 9, 14, 10, 9, 21, 4, 64, 48, 52, 61, 63, 4, 7, 9, 14, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 53, 5, 9, 5, 13, 7, 6, 10, 8, 9, 4, 64, 48, 52, 53, 63, 4, 20, 10, 6, 11, 4, 8, 27, 5, 13, 4, 6, 11, 10, 13, 6, 22, 39, 6, 20, 8, 4, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 4, 18, 8, 14, 5, 15, 12, 4, 10, 9, 4, 8, 9, 5, 4, 11, 16, 9, 14, 13, 5, 14, 4, 24, 15, 16, 12, 4, 15, 7, 9, 21, 16, 7, 21, 5, 12, 4, 7, 9, 14, 4, 14, 5, 5, 24, 4, 10, 9, 6, 5, 13, 8, 24, 5, 13, 7, 25, 10, 15, 10, 6, 22, 4, 25, 5, 6, 20, 5, 5, 9, 4, 58, 7, 37, 23, 4, 49, 22, 32, 8, 13, 17, 11, 4, 7, 9, 14, 4, 32, 5, 9, 12, 8, 13, 55, 15, 8, 20, 26, 2], [4, 40, 47, 54, 32, 4, 10, 12, 4, 14, 5, 12, 10, 21, 9, 5, 14, 4, 6, 8, 4, 24, 13, 5, 39, 6, 13, 7, 10, 9, 4, 14, 5, 5, 24, 4, 25, 10, 14, 10, 13, 5, 17, 6, 10, 8, 9, 7, 15, 4, 13, 5, 24, 13, 5, 12, 5, 9, 6, 7, 6, 10, 8, 9, 12, 4, 19, 13, 8, 18, 4, 16, 9, 15, 7, 25, 5, 15, 5, 14, 4, 6, 5, 37, 6, 4, 25, 22, 4, 46, 8, 10, 9, 6, 15, 22, 4, 17, 8, 9, 14, 10, 6, 10, 8, 9, 10, 9, 21, 4, 8, 9, 4, 25, 8, 6, 11, 4, 15, 5, 19, 6, 4, 7, 9, 14, 4, 13, 10, 21, 11, 6, 4, 17, 8, 9, 6, 5, 37, 6, 4, 10, 9, 4, 7, 15, 15, 4, 15, 7, 22, 5, 13, 12, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 32, 11, 5, 4, 45, 16, 10, 17, 28, 4, 25, 13, 8, 20, 9, 4, 19, 8, 37, 4, 46, 16, 18, 24, 12, 4, 8, 27, 5, 13, 4, 6, 11, 5, 4, 15, 7, 57, 22, 4, 14, 8, 21, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], """attention_mask""": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=__magic_name__ , model_name="""microsoft/speecht5_asr""" , revision="""c5ef64c71905caeccde0e4462ef3f9077224c524""" , sequences=__magic_name__ , )
370
__UpperCamelCase : Dict = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" def __A ( ) -> None: a = input("""Enter message: """ ) a = input("""Enter key [alphanumeric]: """ ) a = input("""Encrypt/Decrypt [e/d]: """ ) if mode.lower().startswith("""e""" ): a = """encrypt""" a = encrypt_message(__lowerCamelCase , __lowerCamelCase ) elif mode.lower().startswith("""d""" ): a = """decrypt""" a = decrypt_message(__lowerCamelCase , __lowerCamelCase ) print(f'\n{mode.title()}ed message:' ) print(__lowerCamelCase ) def __A ( __lowerCamelCase , __lowerCamelCase ) -> str: return translate_message(__lowerCamelCase , __lowerCamelCase , """encrypt""" ) def __A ( __lowerCamelCase , __lowerCamelCase ) -> str: return translate_message(__lowerCamelCase , __lowerCamelCase , """decrypt""" ) def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> str: a = [] a = 0 a = key.upper() for symbol in message: a = LETTERS.find(symbol.upper() ) if num != -1: if mode == "encrypt": num += LETTERS.find(key[key_index] ) elif mode == "decrypt": num -= LETTERS.find(key[key_index] ) num %= len(__lowerCamelCase ) if symbol.isupper(): translated.append(LETTERS[num] ) elif symbol.islower(): translated.append(LETTERS[num].lower() ) key_index += 1 if key_index == len(__lowerCamelCase ): a = 0 else: translated.append(__lowerCamelCase ) return "".join(__lowerCamelCase ) if __name__ == "__main__": main()
347
0
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig snake_case_ = [ """openmmlab/upernet-convnext-tiny""", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring snake_case_ = """UperNetConfig""" class A_ ( nn.Module ): """simple docstring""" def __init__( self :Tuple , lowercase_ :int , lowercase_ :int , lowercase_ :Union[int, Tuple[int, int]] , lowercase_ :Union[int, Tuple[int, int], str] = 0 , lowercase_ :bool = False , lowercase_ :Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() UpperCAmelCase = nn.Convad( in_channels=lowercase_ , out_channels=lowercase_ , kernel_size=lowercase_ , padding=lowercase_ , bias=lowercase_ , dilation=lowercase_ , ) UpperCAmelCase = nn.BatchNormad(lowercase_ ) UpperCAmelCase = nn.ReLU() def UpperCAmelCase__ ( self :Optional[Any] , lowercase_ :torch.Tensor ) -> torch.Tensor: UpperCAmelCase = self.conv(lowercase_ ) UpperCAmelCase = self.batch_norm(lowercase_ ) UpperCAmelCase = self.activation(lowercase_ ) return output class A_ ( nn.Module ): """simple docstring""" def __init__( self :Optional[Any] , lowercase_ :int , lowercase_ :int , lowercase_ :int ) -> None: super().__init__() UpperCAmelCase = [ nn.AdaptiveAvgPoolad(lowercase_ ), UperNetConvModule(lowercase_ , lowercase_ , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowercase_ ) , lowercase_ ) def UpperCAmelCase__ ( self :List[str] , lowercase_ :torch.Tensor ) -> torch.Tensor: UpperCAmelCase = input for layer in self.layers: UpperCAmelCase = layer(lowercase_ ) return hidden_state class A_ ( nn.Module ): """simple docstring""" def __init__( self :Optional[Any] , lowercase_ :Tuple[int, ...] , lowercase_ :int , lowercase_ :int , lowercase_ :bool ) -> None: super().__init__() UpperCAmelCase = pool_scales UpperCAmelCase = align_corners UpperCAmelCase = in_channels UpperCAmelCase = channels UpperCAmelCase = [] for i, pool_scale in enumerate(lowercase_ ): UpperCAmelCase = UperNetPyramidPoolingBlock(pool_scale=lowercase_ , in_channels=lowercase_ , channels=lowercase_ ) self.blocks.append(lowercase_ ) self.add_module(str(lowercase_ ) , lowercase_ ) def UpperCAmelCase__ ( self :List[Any] , lowercase_ :torch.Tensor ) -> List[torch.Tensor]: UpperCAmelCase = [] for ppm in self.blocks: UpperCAmelCase = ppm(lowercase_ ) UpperCAmelCase = nn.functional.interpolate( lowercase_ , size=x.size()[2:] , mode='bilinear' , align_corners=self.align_corners ) ppm_outs.append(lowercase_ ) return ppm_outs class A_ ( nn.Module ): """simple docstring""" def __init__( self :Dict , lowercase_ :Optional[Any] , lowercase_ :Optional[int] ) -> Any: super().__init__() UpperCAmelCase = config UpperCAmelCase = config.pool_scales # e.g. (1, 2, 3, 6) UpperCAmelCase = in_channels UpperCAmelCase = config.hidden_size UpperCAmelCase = False UpperCAmelCase = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module UpperCAmelCase = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) UpperCAmelCase = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module UpperCAmelCase = nn.ModuleList() UpperCAmelCase = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer UpperCAmelCase = UperNetConvModule(lowercase_ , self.channels , kernel_size=1 ) UpperCAmelCase = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowercase_ ) self.fpn_convs.append(lowercase_ ) UpperCAmelCase = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def UpperCAmelCase__ ( self :Optional[Any] ) -> List[Any]: self.apply(self._init_weights ) def UpperCAmelCase__ ( self :str , lowercase_ :Union[str, Any] ) -> str: if isinstance(lowercase_ , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def UpperCAmelCase__ ( self :Dict , lowercase_ :int ) -> int: UpperCAmelCase = inputs[-1] UpperCAmelCase = [x] psp_outs.extend(self.psp_modules(lowercase_ ) ) UpperCAmelCase = torch.cat(lowercase_ , dim=1 ) UpperCAmelCase = self.bottleneck(lowercase_ ) return output def UpperCAmelCase__ ( self :str , lowercase_ :torch.Tensor ) -> torch.Tensor: # build laterals UpperCAmelCase = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowercase_ ) ) # build top-down path UpperCAmelCase = len(lowercase_ ) for i in range(used_backbone_levels - 1 , 0 , -1 ): UpperCAmelCase = laterals[i - 1].shape[2:] UpperCAmelCase = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowercase_ , mode='bilinear' , align_corners=self.align_corners ) # build outputs UpperCAmelCase = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): UpperCAmelCase = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode='bilinear' , align_corners=self.align_corners ) UpperCAmelCase = torch.cat(lowercase_ , dim=1 ) UpperCAmelCase = self.fpn_bottleneck(lowercase_ ) UpperCAmelCase = self.classifier(lowercase_ ) return output class A_ ( nn.Module ): """simple docstring""" def __init__( self :Optional[Any] , lowercase_ :Optional[Any] , lowercase_ :int = 2 , lowercase_ :int = 3 , lowercase_ :Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() UpperCAmelCase = config UpperCAmelCase = config.auxiliary_in_channels UpperCAmelCase = config.auxiliary_channels UpperCAmelCase = config.auxiliary_num_convs UpperCAmelCase = config.auxiliary_concat_input UpperCAmelCase = in_index UpperCAmelCase = (kernel_size // 2) * dilation UpperCAmelCase = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowercase_ , padding=lowercase_ , dilation=lowercase_ ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowercase_ , padding=lowercase_ , dilation=lowercase_ ) ) if self.num_convs == 0: UpperCAmelCase = nn.Identity() else: UpperCAmelCase = nn.Sequential(*lowercase_ ) if self.concat_input: UpperCAmelCase = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowercase_ , padding=kernel_size // 2 ) UpperCAmelCase = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def UpperCAmelCase__ ( self :List[str] ) -> Dict: self.apply(self._init_weights ) def UpperCAmelCase__ ( self :int , lowercase_ :Any ) -> List[Any]: if isinstance(lowercase_ , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def UpperCAmelCase__ ( self :Dict , lowercase_ :torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps UpperCAmelCase = encoder_hidden_states[self.in_index] UpperCAmelCase = self.convs(lowercase_ ) if self.concat_input: UpperCAmelCase = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) UpperCAmelCase = self.classifier(lowercase_ ) return output class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __UpperCamelCase = UperNetConfig __UpperCamelCase = """pixel_values""" __UpperCamelCase = True def UpperCAmelCase__ ( self :Tuple , lowercase_ :List[str] ) -> Union[str, Any]: if isinstance(lowercase_ , lowercase_ ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def UpperCAmelCase__ ( self :int ) -> Tuple: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def UpperCAmelCase__ ( self :Any , lowercase_ :Dict , lowercase_ :Union[str, Any]=False ) -> Optional[int]: if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase = value snake_case_ = R""" Parameters: This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. config ([`UperNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ snake_case_ = R""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details. output_attentions (`bool`, *optional*): Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See `attentions` under returned tensors for more detail. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( """UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes.""" , SCREAMING_SNAKE_CASE_ , ) class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self :int , lowercase_ :Tuple ) -> int: super().__init__(lowercase_ ) UpperCAmelCase = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) UpperCAmelCase = UperNetHead(lowercase_ , in_channels=self.backbone.channels ) UpperCAmelCase = UperNetFCNHead(lowercase_ ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format('batch_size, sequence_length' ) ) @replace_return_docstrings(output_type=lowercase_ , config_class=_CONFIG_FOR_DOC ) def UpperCAmelCase__ ( self :Optional[int] , lowercase_ :Optional[torch.Tensor] = None , lowercase_ :Optional[bool] = None , lowercase_ :Optional[bool] = None , lowercase_ :Optional[torch.Tensor] = None , lowercase_ :Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: UpperCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) UpperCAmelCase = output_attentions if output_attentions is not None else self.config.output_attentions UpperCAmelCase = self.backbone.forward_with_filtered_kwargs( lowercase_ , output_hidden_states=lowercase_ , output_attentions=lowercase_ ) UpperCAmelCase = outputs.feature_maps UpperCAmelCase = self.decode_head(lowercase_ ) UpperCAmelCase = nn.functional.interpolate(lowercase_ , size=pixel_values.shape[2:] , mode='bilinear' , align_corners=lowercase_ ) UpperCAmelCase = None if self.auxiliary_head is not None: UpperCAmelCase = self.auxiliary_head(lowercase_ ) UpperCAmelCase = nn.functional.interpolate( lowercase_ , size=pixel_values.shape[2:] , mode='bilinear' , align_corners=lowercase_ ) UpperCAmelCase = None if labels is not None: if self.config.num_labels == 1: raise ValueError('The number of labels should be greater than one' ) else: # compute weighted loss UpperCAmelCase = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) UpperCAmelCase = loss_fct(lowercase_ , lowercase_ ) UpperCAmelCase = loss_fct(lowercase_ , lowercase_ ) UpperCAmelCase = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: UpperCAmelCase = (logits,) + outputs[1:] else: UpperCAmelCase = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowercase_ , logits=lowercase_ , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
78
"""simple docstring""" import os import time import numpy as np import onnxruntime as ort snake_case_ = """1""" snake_case_ = """0""" snake_case_ = """1""" snake_case_ = ort.SessionOptions() snake_case_ = ort.GraphOptimizationLevel.ORT_DISABLE_ALL print("""Create inference session...""") snake_case_ = ["""TensorrtExecutionProvider""", """CUDAExecutionProvider"""] snake_case_ = ort.InferenceSession("""model.onnx""", sess_options=sess_opt, providers=execution_provider) snake_case_ = ort.RunOptions() snake_case_ = 128 snake_case_ = 1 snake_case_ = np.ones((batch, sequence), dtype=np.intaa) snake_case_ = np.ones((batch, sequence), dtype=np.intaa) snake_case_ = np.ones((batch, sequence), dtype=np.intaa) print("""Warm up phase...""") sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print("""Start inference...""") snake_case_ = time.time() snake_case_ = 2000 snake_case_ = {} for iter in range(max_iters): snake_case_ = sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print("""Average Inference Time = {:.3f} ms""".format((time.time() - start_time) * 1000 / max_iters))
78
1
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A = { "configuration_autoformer": [ "AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "AutoformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "AutoformerForPrediction", "AutoformerModel", "AutoformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
273
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig __A = logging.get_logger(__name__) # General docstring __A = "ResNetConfig" # Base docstring __A = "microsoft/resnet-50" __A = [1, 2048, 7, 7] # Image classification docstring __A = "microsoft/resnet-50" __A = "tiger cat" __A = [ "microsoft/resnet-50", # See all resnet models at https://huggingface.co/models?filter=resnet ] class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__(self : Optional[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int = 3 , UpperCAmelCase_ : int = 1 , UpperCAmelCase_ : str = "relu") ->Any: '''simple docstring''' super().__init__() lowerCamelCase__: Dict =nn.Convad( UpperCAmelCase_ , UpperCAmelCase_ , kernel_size=UpperCAmelCase_ , stride=UpperCAmelCase_ , padding=kernel_size // 2 , bias=UpperCAmelCase_) lowerCamelCase__: Any =nn.BatchNormad(UpperCAmelCase_) lowerCamelCase__: Optional[int] =ACTaFN[activation] if activation is not None else nn.Identity() def SCREAMING_SNAKE_CASE_ (self : Tuple , UpperCAmelCase_ : Tensor) ->Tensor: '''simple docstring''' lowerCamelCase__: List[Any] =self.convolution(UpperCAmelCase_) lowerCamelCase__: List[str] =self.normalization(UpperCAmelCase_) lowerCamelCase__: Optional[Any] =self.activation(UpperCAmelCase_) return hidden_state class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__(self : Any , UpperCAmelCase_ : ResNetConfig) ->str: '''simple docstring''' super().__init__() lowerCamelCase__: Tuple =ResNetConvLayer( config.num_channels , config.embedding_size , kernel_size=7 , stride=2 , activation=config.hidden_act) lowerCamelCase__: Optional[int] =nn.MaxPoolad(kernel_size=3 , stride=2 , padding=1) lowerCamelCase__: Optional[Any] =config.num_channels def SCREAMING_SNAKE_CASE_ (self : Union[str, Any] , UpperCAmelCase_ : Tensor) ->Tensor: '''simple docstring''' lowerCamelCase__: Optional[Any] =pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration.") lowerCamelCase__: Dict =self.embedder(UpperCAmelCase_) lowerCamelCase__: str =self.pooler(UpperCAmelCase_) return embedding class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__(self : List[str] , UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int = 2) ->Any: '''simple docstring''' super().__init__() lowerCamelCase__: Optional[Any] =nn.Convad(UpperCAmelCase_ , UpperCAmelCase_ , kernel_size=1 , stride=UpperCAmelCase_ , bias=UpperCAmelCase_) lowerCamelCase__: Union[str, Any] =nn.BatchNormad(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Optional[int] , UpperCAmelCase_ : Tensor) ->Tensor: '''simple docstring''' lowerCamelCase__: Optional[int] =self.convolution(UpperCAmelCase_) lowerCamelCase__: Any =self.normalization(UpperCAmelCase_) return hidden_state class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__(self : Tuple , UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int = 1 , UpperCAmelCase_ : str = "relu") ->Tuple: '''simple docstring''' super().__init__() lowerCamelCase__: Tuple =in_channels != out_channels or stride != 1 lowerCamelCase__: str =( ResNetShortCut(UpperCAmelCase_ , UpperCAmelCase_ , stride=UpperCAmelCase_) if should_apply_shortcut else nn.Identity() ) lowerCamelCase__: Tuple =nn.Sequential( ResNetConvLayer(UpperCAmelCase_ , UpperCAmelCase_ , stride=UpperCAmelCase_) , ResNetConvLayer(UpperCAmelCase_ , UpperCAmelCase_ , activation=UpperCAmelCase_) , ) lowerCamelCase__: Optional[Any] =ACTaFN[activation] def SCREAMING_SNAKE_CASE_ (self : Tuple , UpperCAmelCase_ : Any) ->Optional[Any]: '''simple docstring''' lowerCamelCase__: Any =hidden_state lowerCamelCase__: List[str] =self.layer(UpperCAmelCase_) lowerCamelCase__: str =self.shortcut(UpperCAmelCase_) hidden_state += residual lowerCamelCase__: Dict =self.activation(UpperCAmelCase_) return hidden_state class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__(self : List[str] , UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int = 1 , UpperCAmelCase_ : str = "relu" , UpperCAmelCase_ : int = 4) ->Tuple: '''simple docstring''' super().__init__() lowerCamelCase__: Union[str, Any] =in_channels != out_channels or stride != 1 lowerCamelCase__: List[str] =out_channels // reduction lowerCamelCase__: Optional[Any] =( ResNetShortCut(UpperCAmelCase_ , UpperCAmelCase_ , stride=UpperCAmelCase_) if should_apply_shortcut else nn.Identity() ) lowerCamelCase__: Dict =nn.Sequential( ResNetConvLayer(UpperCAmelCase_ , UpperCAmelCase_ , kernel_size=1) , ResNetConvLayer(UpperCAmelCase_ , UpperCAmelCase_ , stride=UpperCAmelCase_) , ResNetConvLayer(UpperCAmelCase_ , UpperCAmelCase_ , kernel_size=1 , activation=UpperCAmelCase_) , ) lowerCamelCase__: Tuple =ACTaFN[activation] def SCREAMING_SNAKE_CASE_ (self : int , UpperCAmelCase_ : List[str]) ->int: '''simple docstring''' lowerCamelCase__: str =hidden_state lowerCamelCase__: Optional[Any] =self.layer(UpperCAmelCase_) lowerCamelCase__: List[Any] =self.shortcut(UpperCAmelCase_) hidden_state += residual lowerCamelCase__: Tuple =self.activation(UpperCAmelCase_) return hidden_state class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__(self : str , UpperCAmelCase_ : ResNetConfig , UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int = 2 , UpperCAmelCase_ : int = 2 , ) ->Dict: '''simple docstring''' super().__init__() lowerCamelCase__: List[Any] =ResNetBottleNeckLayer if config.layer_type == "bottleneck" else ResNetBasicLayer lowerCamelCase__: List[str] =nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(UpperCAmelCase_ , UpperCAmelCase_ , stride=UpperCAmelCase_ , activation=config.hidden_act) , *[layer(UpperCAmelCase_ , UpperCAmelCase_ , activation=config.hidden_act) for _ in range(depth - 1)] , ) def SCREAMING_SNAKE_CASE_ (self : Tuple , UpperCAmelCase_ : Tensor) ->Tensor: '''simple docstring''' lowerCamelCase__: List[Any] =input for layer in self.layers: lowerCamelCase__: Any =layer(UpperCAmelCase_) return hidden_state class _SCREAMING_SNAKE_CASE ( nn.Module ): '''simple docstring''' def __init__(self : int , UpperCAmelCase_ : ResNetConfig) ->Any: '''simple docstring''' super().__init__() lowerCamelCase__: Tuple =nn.ModuleList([]) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( UpperCAmelCase_ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , )) lowerCamelCase__: int =zip(config.hidden_sizes , config.hidden_sizes[1:]) for (in_channels, out_channels), depth in zip(UpperCAmelCase_ , config.depths[1:]): self.stages.append(ResNetStage(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , depth=UpperCAmelCase_)) def SCREAMING_SNAKE_CASE_ (self : Tuple , UpperCAmelCase_ : Tensor , UpperCAmelCase_ : bool = False , UpperCAmelCase_ : bool = True) ->BaseModelOutputWithNoAttention: '''simple docstring''' lowerCamelCase__: str =() if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: lowerCamelCase__: Union[str, Any] =hidden_states + (hidden_state,) lowerCamelCase__: str =stage_module(UpperCAmelCase_) if output_hidden_states: lowerCamelCase__: Optional[int] =hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None) return BaseModelOutputWithNoAttention( last_hidden_state=UpperCAmelCase_ , hidden_states=UpperCAmelCase_ , ) class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = ResNetConfig lowercase_ = "resnet" lowercase_ = "pixel_values" lowercase_ = True def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : int) ->Any: '''simple docstring''' if isinstance(UpperCAmelCase_ , nn.Convad): nn.init.kaiming_normal_(module.weight , mode="fan_out" , nonlinearity="relu") elif isinstance(UpperCAmelCase_ , (nn.BatchNormad, nn.GroupNorm)): nn.init.constant_(module.weight , 1) nn.init.constant_(module.bias , 0) def SCREAMING_SNAKE_CASE_ (self : Any , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[Any]=False) ->Any: '''simple docstring''' if isinstance(UpperCAmelCase_ , UpperCAmelCase_): lowerCamelCase__: List[str] =value __A = R"\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`ResNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" __A = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConvNextImageProcessor.__call__`] for details.\n\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "The bare ResNet model outputting raw features without any specific head on top." , __SCREAMING_SNAKE_CASE , ) class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__(self : Union[str, Any] , UpperCAmelCase_ : str) ->int: '''simple docstring''' super().__init__(UpperCAmelCase_) lowerCamelCase__: str =config lowerCamelCase__: str =ResNetEmbeddings(UpperCAmelCase_) lowerCamelCase__: Optional[Any] =ResNetEncoder(UpperCAmelCase_) lowerCamelCase__: Optional[Any] =nn.AdaptiveAvgPoolad((1, 1)) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCAmelCase_) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def SCREAMING_SNAKE_CASE_ (self : Optional[int] , UpperCAmelCase_ : Tensor , UpperCAmelCase_ : Optional[bool] = None , UpperCAmelCase_ : Optional[bool] = None) ->BaseModelOutputWithPoolingAndNoAttention: '''simple docstring''' lowerCamelCase__: Optional[int] =( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCamelCase__: Optional[int] =return_dict if return_dict is not None else self.config.use_return_dict lowerCamelCase__: int =self.embedder(UpperCAmelCase_) lowerCamelCase__: int =self.encoder( UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_) lowerCamelCase__: int =encoder_outputs[0] lowerCamelCase__: Tuple =self.pooler(UpperCAmelCase_) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCAmelCase_ , pooler_output=UpperCAmelCase_ , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( "\n ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , __SCREAMING_SNAKE_CASE , ) class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__(self : Tuple , UpperCAmelCase_ : Union[str, Any]) ->Any: '''simple docstring''' super().__init__(UpperCAmelCase_) lowerCamelCase__: int =config.num_labels lowerCamelCase__: Optional[Any] =ResNetModel(UpperCAmelCase_) # classification head lowerCamelCase__: Optional[Any] =nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCAmelCase_) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def SCREAMING_SNAKE_CASE_ (self : Union[str, Any] , UpperCAmelCase_ : Optional[torch.FloatTensor] = None , UpperCAmelCase_ : Optional[torch.LongTensor] = None , UpperCAmelCase_ : Optional[bool] = None , UpperCAmelCase_ : Optional[bool] = None , ) ->ImageClassifierOutputWithNoAttention: '''simple docstring''' lowerCamelCase__: Optional[int] =return_dict if return_dict is not None else self.config.use_return_dict lowerCamelCase__: List[Any] =self.resnet(UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_) lowerCamelCase__: int =outputs.pooler_output if return_dict else outputs[1] lowerCamelCase__: Dict =self.classifier(UpperCAmelCase_) lowerCamelCase__: Optional[Any] =None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowerCamelCase__: Dict ="regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowerCamelCase__: Tuple ="single_label_classification" else: lowerCamelCase__: Optional[int] ="multi_label_classification" if self.config.problem_type == "regression": lowerCamelCase__: Dict =MSELoss() if self.num_labels == 1: lowerCamelCase__: str =loss_fct(logits.squeeze() , labels.squeeze()) else: lowerCamelCase__: int =loss_fct(UpperCAmelCase_ , UpperCAmelCase_) elif self.config.problem_type == "single_label_classification": lowerCamelCase__: List[Any] =CrossEntropyLoss() lowerCamelCase__: int =loss_fct(logits.view(-1 , self.num_labels) , labels.view(-1)) elif self.config.problem_type == "multi_label_classification": lowerCamelCase__: List[str] =BCEWithLogitsLoss() lowerCamelCase__: int =loss_fct(UpperCAmelCase_ , UpperCAmelCase_) if not return_dict: lowerCamelCase__: List[str] =(logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=UpperCAmelCase_ , logits=UpperCAmelCase_ , hidden_states=outputs.hidden_states) @add_start_docstrings( "\n ResNet backbone, to be used with frameworks like DETR and MaskFormer.\n " , __SCREAMING_SNAKE_CASE , ) class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__(self : str , UpperCAmelCase_ : List[Any]) ->Dict: '''simple docstring''' super().__init__(UpperCAmelCase_) super()._init_backbone(UpperCAmelCase_) lowerCamelCase__: int =[config.embedding_size] + config.hidden_sizes lowerCamelCase__: List[Any] =ResNetEmbeddings(UpperCAmelCase_) lowerCamelCase__: Optional[int] =ResNetEncoder(UpperCAmelCase_) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCAmelCase_) @replace_return_docstrings(output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC) def SCREAMING_SNAKE_CASE_ (self : Union[str, Any] , UpperCAmelCase_ : Tensor , UpperCAmelCase_ : Optional[bool] = None , UpperCAmelCase_ : Optional[bool] = None) ->BackboneOutput: '''simple docstring''' lowerCamelCase__: Optional[Any] =return_dict if return_dict is not None else self.config.use_return_dict lowerCamelCase__: Union[str, Any] =( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCamelCase__: int =self.embedder(UpperCAmelCase_) lowerCamelCase__: Optional[Any] =self.encoder(UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_) lowerCamelCase__: Any =outputs.hidden_states lowerCamelCase__: int =() for idx, stage in enumerate(self.stage_names): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: lowerCamelCase__: Dict =(feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=UpperCAmelCase_ , hidden_states=outputs.hidden_states if output_hidden_states else None , attentions=UpperCAmelCase_ , )
273
1