code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import operator as op def lowerCAmelCase_ (lowerCAmelCase__: List[str] ): """simple docstring""" UpperCAmelCase_: Dict = [] UpperCAmelCase_: str = lambda lowerCAmelCase__ , lowerCAmelCase__ : int(x / y ) # noqa: E731 integer division operation UpperCAmelCase_: Union[str, Any] = { """^""": op.pow, """*""": op.mul, """/""": div, """+""": op.add, """-""": op.sub, } # operators & their respective operation # print table header print("""Symbol""".center(8 ) , """Action""".center(1_2 ) , """Stack""" , sep=""" | """ ) print("""-""" * (3_0 + len(lowerCAmelCase__ )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(lowerCAmelCase__ ) # append x to stack # output in tabular format print(x.rjust(8 ) , ("""push(""" + x + """)""").ljust(1_2 ) , """,""".join(lowerCAmelCase__ ) , sep=""" | """ ) else: UpperCAmelCase_: List[Any] = stack.pop() # pop stack # output in tabular format print("""""".rjust(8 ) , ("""pop(""" + b + """)""").ljust(1_2 ) , """,""".join(lowerCAmelCase__ ) , sep=""" | """ ) UpperCAmelCase_: Tuple = stack.pop() # pop stack # output in tabular format print("""""".rjust(8 ) , ("""pop(""" + a + """)""").ljust(1_2 ) , """,""".join(lowerCAmelCase__ ) , sep=""" | """ ) stack.append( str(opr[x](int(lowerCAmelCase__ ) , int(lowerCAmelCase__ ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) , ("""push(""" + a + x + b + """)""").ljust(1_2 ) , """,""".join(lowerCAmelCase__ ) , sep=""" | """ , ) return int(stack[0] ) if __name__ == "__main__": a : Optional[int] = input('\n\nEnter a Postfix Equation (space separated) = ').split(' ') print('\n\tResult = ', solve(Postfix))
147
from collections import namedtuple a : List[Any] = namedtuple('from_to', 'from_ to') a : Tuple = { 'cubicmeter': from_to(1, 1), 'litre': from_to(0.0_0_1, 1_000), 'kilolitre': from_to(1, 1), 'gallon': from_to(0.0_0_4_5_4, 2_6_4.1_7_2), 'cubicyard': from_to(0.7_6_4_5_5, 1.3_0_7_9_5), 'cubicfoot': from_to(0.0_2_8, 3_5.3_1_4_7), 'cup': from_to(0.0_0_0_2_3_6_5_8_8, 4_2_2_6.7_5), } def lowerCAmelCase_ (lowerCAmelCase__: float , lowerCAmelCase__: str , lowerCAmelCase__: str ): """simple docstring""" if from_type not in METRIC_CONVERSION: raise ValueError( F'Invalid \'from_type\' value: {from_type!r} Supported values are:\n' + """, """.join(lowerCAmelCase__ ) ) if to_type not in METRIC_CONVERSION: raise ValueError( F'Invalid \'to_type\' value: {to_type!r}. Supported values are:\n' + """, """.join(lowerCAmelCase__ ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
147
1
"""simple docstring""" # Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best hparam into the config # - generate model_cards - useful if you have multiple models from the same paper import argparse import json import os import re from collections import OrderedDict from os.path import basename, dirname import fairseq import torch from fairseq import hub_utils from fairseq.data.dictionary import Dictionary from transformers import FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() SCREAMING_SNAKE_CASE = 2 # based on the results of a search on a range of `num_beams`, `length_penalty` and `early_stopping` # values against wmt19 test data to obtain the best BLEU scores, we will use the following defaults: # # * `num_beams`: 5 (higher scores better, but requires more memory/is slower, can be adjusted by users) # * `early_stopping`: `False` consistently scored better # * `length_penalty` varied, so will assign the best one depending on the model SCREAMING_SNAKE_CASE = { # fairseq: "wmt19-ru-en": {"length_penalty": 1.1}, "wmt19-en-ru": {"length_penalty": 1.15}, "wmt19-en-de": {"length_penalty": 1.0}, "wmt19-de-en": {"length_penalty": 1.1}, # allenai: "wmt16-en-de-dist-12-1": {"length_penalty": 0.6}, "wmt16-en-de-dist-6-1": {"length_penalty": 0.6}, "wmt16-en-de-12-1": {"length_penalty": 0.8}, "wmt19-de-en-6-6-base": {"length_penalty": 0.6}, "wmt19-de-en-6-6-big": {"length_penalty": 0.6}, } # this remaps the different models to their organization names SCREAMING_SNAKE_CASE = {} for m in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: SCREAMING_SNAKE_CASE = "facebook" for m in [ "wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1", "wmt19-de-en-6-6-base", "wmt19-de-en-6-6-big", ]: SCREAMING_SNAKE_CASE = "allenai" def _SCREAMING_SNAKE_CASE ( lowercase_ ) -> Union[str, Any]: # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} A__ = dict((re.sub(R"@@$" , "" , lowercase_ ), v) if k.endswith("@@" ) else (re.sub(R"$" , "</w>" , lowercase_ ), v) for k, v in d.items() ) A__ = "<s> <pad> </s> <unk>".split() # restore the special tokens for k in keep_keys: del da[f"""{k}</w>"""] A__ = d[k] # restore return da def _SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ ) -> int: # prep assert os.path.exists(lowercase_ ) os.makedirs(lowercase_ , exist_ok=lowercase_ ) print(f"""Writing results to {pytorch_dump_folder_path}""" ) # handle various types of models A__ = basename(lowercase_ ) A__ = dirname(lowercase_ ) A__ = fairseq.model_parallel.models.transformer.ModelParallelTransformerModel A__ = cls.hub_models() A__ = {"bpe": "fastbpe", "tokenizer": "moses"} A__ = "." # note: since the model dump is old, fairseq has upgraded its model some # time later, and it does a whole lot of rewrites and splits on the saved # weights, therefore we can't use torch.load() directly on the model file. # see: upgrade_state_dict(state_dict) in fairseq_model.py print(f"""using checkpoint {checkpoint_file}""" ) A__ = hub_utils.from_pretrained( lowercase_ , lowercase_ , lowercase_ , archive_map=lowercase_ , **lowercase_ ) A__ = vars(chkpt["args"]["model"] ) A__ = args["source_lang"] A__ = args["target_lang"] A__ = dirname(lowercase_ ) A__ = basename(lowercase_ ) # dicts A__ = os.path.join(lowercase_ , f"""dict.{src_lang}.txt""" ) A__ = os.path.join(lowercase_ , f"""dict.{tgt_lang}.txt""" ) A__ = Dictionary.load(lowercase_ ) A__ = rewrite_dict_keys(src_dict.indices ) A__ = len(lowercase_ ) A__ = os.path.join(lowercase_ , "vocab-src.json" ) print(f"""Generating {src_vocab_file} of {src_vocab_size} of {src_lang} records""" ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(lowercase_ , ensure_ascii=lowercase_ , indent=lowercase_ ) ) # detect whether this is a do_lower_case situation, which can be derived by checking whether we # have at least one uppercase letter in the source vocab A__ = True for k in src_vocab.keys(): if not k.islower(): A__ = False break A__ = Dictionary.load(lowercase_ ) A__ = rewrite_dict_keys(tgt_dict.indices ) A__ = len(lowercase_ ) A__ = os.path.join(lowercase_ , "vocab-tgt.json" ) print(f"""Generating {tgt_vocab_file} of {tgt_vocab_size} of {tgt_lang} records""" ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(lowercase_ , ensure_ascii=lowercase_ , indent=lowercase_ ) ) # merges_file (bpecodes) A__ = os.path.join(lowercase_ , VOCAB_FILES_NAMES["merges_file"] ) for fn in ["bpecodes", "code"]: # older fairseq called the merges file "code" A__ = os.path.join(lowercase_ , lowercase_ ) if os.path.exists(lowercase_ ): break with open(lowercase_ , encoding="utf-8" ) as fin: A__ = fin.read() A__ = re.sub(R" \d+$" , "" , lowercase_ , 0 , re.M ) # remove frequency number print(f"""Generating {merges_file}""" ) with open(lowercase_ , "w" , encoding="utf-8" ) as fout: fout.write(lowercase_ ) # model config A__ = os.path.join(lowercase_ , "config.json" ) # validate bpe/tokenizer config, as currently it's hardcoded to moses+fastbpe - # may have to modify the tokenizer if a different type is used by a future model assert args["bpe"] == "fastbpe", f"""need to extend tokenizer to support bpe={args["bpe"]}""" assert args["tokenizer"] == "moses", f"""need to extend tokenizer to support bpe={args["tokenizer"]}""" A__ = { "architectures": ["FSMTForConditionalGeneration"], "model_type": "fsmt", "activation_dropout": args["activation_dropout"], "activation_function": "relu", "attention_dropout": args["attention_dropout"], "d_model": args["decoder_embed_dim"], "dropout": args["dropout"], "init_std": 0.0_2, "max_position_embeddings": args["max_source_positions"], "num_hidden_layers": args["encoder_layers"], "src_vocab_size": src_vocab_size, "tgt_vocab_size": tgt_vocab_size, "langs": [src_lang, tgt_lang], "encoder_attention_heads": args["encoder_attention_heads"], "encoder_ffn_dim": args["encoder_ffn_embed_dim"], "encoder_layerdrop": args["encoder_layerdrop"], "encoder_layers": args["encoder_layers"], "decoder_attention_heads": args["decoder_attention_heads"], "decoder_ffn_dim": args["decoder_ffn_embed_dim"], "decoder_layerdrop": args["decoder_layerdrop"], "decoder_layers": args["decoder_layers"], "bos_token_id": 0, "pad_token_id": 1, "eos_token_id": 2, "is_encoder_decoder": True, "scale_embedding": not args["no_scale_embedding"], "tie_word_embeddings": args["share_all_embeddings"], } # good hparam defaults to start with A__ = 5 A__ = False if model_dir in best_score_hparams and "length_penalty" in best_score_hparams[model_dir]: A__ = best_score_hparams[model_dir]["length_penalty"] else: A__ = 1.0 print(f"""Generating {fsmt_model_config_file}""" ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(lowercase_ , ensure_ascii=lowercase_ , indent=lowercase_ ) ) # tokenizer config A__ = os.path.join(lowercase_ , lowercase_ ) A__ = { "langs": [src_lang, tgt_lang], "model_max_length": 10_24, "do_lower_case": do_lower_case, } print(f"""Generating {fsmt_tokenizer_config_file}""" ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(lowercase_ , ensure_ascii=lowercase_ , indent=lowercase_ ) ) # model A__ = chkpt["models"][0] A__ = model.state_dict() # rename keys to start with 'model.' A__ = OrderedDict(("model." + k, v) for k, v in model_state_dict.items() ) # remove unneeded keys A__ = [ "model.model", "model.encoder.version", "model.decoder.version", "model.encoder_embed_tokens.weight", "model.decoder_embed_tokens.weight", "model.encoder.embed_positions._float_tensor", "model.decoder.embed_positions._float_tensor", ] for k in ignore_keys: model_state_dict.pop(lowercase_ , lowercase_ ) A__ = FSMTConfig.from_pretrained(lowercase_ ) A__ = FSMTForConditionalGeneration(lowercase_ ) # check that it loads ok model_new.load_state_dict(lowercase_ , strict=lowercase_ ) # save A__ = os.path.join(lowercase_ , lowercase_ ) print(f"""Generating {pytorch_weights_dump_path}""" ) torch.save(lowercase_ , lowercase_ ) print("Conversion is done!" ) print("\nLast step is to upload the files to s3" ) print(f"""cd {data_root}""" ) print(f"""transformers-cli upload {model_dir}""" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--fsmt_checkpoint_path", default=None, type=str, required=True, help=( "Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts," " bpecodes, etc." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_fsmt_checkpoint_to_pytorch(args.fsmt_checkpoint_path, args.pytorch_dump_folder_path)
230
"""simple docstring""" from __future__ import annotations SCREAMING_SNAKE_CASE = "#" class UpperCAmelCase_ : def __init__( self : Dict ) -> None: '''simple docstring''' A__ = {} def __magic_name__ ( self : Optional[Any] , snake_case_ : str ) -> None: '''simple docstring''' A__ = self._trie for char in text: if char not in trie: A__ = {} A__ = trie[char] A__ = True def __magic_name__ ( self : List[Any] , snake_case_ : str ) -> tuple | list: '''simple docstring''' A__ = self._trie for char in prefix: if char in trie: A__ = trie[char] else: return [] return self._elements(snake_case_ ) def __magic_name__ ( self : Union[str, Any] , snake_case_ : dict ) -> tuple: '''simple docstring''' A__ = [] for c, v in d.items(): A__ = [" "] if c == END else [(c + s) for s in self._elements(snake_case_ )] result.extend(snake_case_ ) return tuple(snake_case_ ) SCREAMING_SNAKE_CASE = Trie() SCREAMING_SNAKE_CASE = ("depart", "detergent", "daring", "dog", "deer", "deal") for word in words: trie.insert_word(word) def _SCREAMING_SNAKE_CASE ( lowercase_ ) -> tuple: A__ = trie.find_word(lowercase_ ) return tuple(string + word for word in suffixes ) def _SCREAMING_SNAKE_CASE ( ) -> None: print(autocomplete_using_trie("de" ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
230
1
'''simple docstring''' import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed __lowerCAmelCase = 'true' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=82 , _SCREAMING_SNAKE_CASE=16 ): set_seed(42 ) _snake_case = RegressionModel() _snake_case = deepcopy(lowercase__ ) _snake_case = RegressionDataset(length=lowercase__ ) _snake_case = DataLoader(lowercase__ , batch_size=lowercase__ ) model.to(accelerator.device ) _snake_case = accelerator.prepare(lowercase__ , lowercase__ ) return model, ddp_model, dataloader def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ): _snake_case = AutoTokenizer.from_pretrained("""hf-internal-testing/mrpc-bert-base-cased""" ) _snake_case = load_dataset("""glue""" , """mrpc""" , split="""validation""" ) def tokenize_function(_SCREAMING_SNAKE_CASE ): _snake_case = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=lowercase__ , max_length=lowercase__ ) return outputs with accelerator.main_process_first(): _snake_case = dataset.map( lowercase__ , batched=lowercase__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) _snake_case = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(_SCREAMING_SNAKE_CASE ): if use_longest: return tokenizer.pad(lowercase__ , padding="""longest""" , return_tensors="""pt""" ) return tokenizer.pad(lowercase__ , padding="""max_length""" , max_length=128 , return_tensors="""pt""" ) return DataLoader(lowercase__ , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=16 ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = Accelerator(dispatch_batches=lowercase__ , split_batches=lowercase__ ) _snake_case = get_dataloader(lowercase__ , not dispatch_batches ) _snake_case = AutoModelForSequenceClassification.from_pretrained( """hf-internal-testing/mrpc-bert-base-cased""" , return_dict=lowercase__ ) _snake_case = accelerator.prepare(lowercase__ , lowercase__ ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = [] for batch in dataloader: _snake_case = batch.values() with torch.no_grad(): _snake_case = model(lowercase__ ) _snake_case = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) _snake_case = [], [] for logit, targ in logits_and_targets: logits.append(lowercase__ ) targs.append(lowercase__ ) _snake_case = torch.cat(lowercase__ ), torch.cat(lowercase__ ) return logits, targs def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=82 , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=16 ): _snake_case = get_basic_setup(lowercase__ , lowercase__ , lowercase__ ) _snake_case = generate_predictions(lowercase__ , lowercase__ , lowercase__ ) assert ( len(lowercase__ ) == num_samples ), f"""Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(lowercase__ )}""" def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = False , _SCREAMING_SNAKE_CASE = False ): _snake_case = evaluate.load("""glue""" , """mrpc""" ) _snake_case = get_mrpc_setup(lowercase__ , lowercase__ ) # First do baseline _snake_case = setup['''no'''] model.to(lowercase__ ) model.eval() for batch in dataloader: batch.to(lowercase__ ) with torch.inference_mode(): _snake_case = model(**lowercase__ ) _snake_case = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=lowercase__ , references=batch["""labels"""] ) _snake_case = metric.compute() # Then do distributed _snake_case = setup['''ddp'''] model.eval() for batch in dataloader: with torch.inference_mode(): _snake_case = model(**lowercase__ ) _snake_case = outputs.logits.argmax(dim=-1 ) _snake_case = batch['''labels'''] _snake_case = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=lowercase__ , references=lowercase__ ) _snake_case = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f"""Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n""" def __SCREAMING_SNAKE_CASE ( ): _snake_case = Accelerator(split_batches=lowercase__ , dispatch_batches=lowercase__ ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("""**Testing gather_for_metrics**""" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`""" ) test_mrpc(lowercase__ , lowercase__ ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("""**Test torch metrics**""" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: _snake_case = Accelerator(split_batches=lowercase__ , dispatch_batches=lowercase__ ) if accelerator.is_local_main_process: print(f"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99""" ) test_torch_metrics(lowercase__ , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("""**Test last batch is not dropped when perfectly divisible**""" ) _snake_case = Accelerator() test_torch_metrics(lowercase__ , 512 ) accelerator.state._reset_state() def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
341
import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __lowerCAmelCase : Optional[int] ='\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' __lowerCAmelCase : Optional[Any] ='\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' __lowerCAmelCase : Dict ='\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _UpperCamelCase ( lowercase__ , lowercase__ ): return float((preds == labels).mean() ) def _UpperCamelCase ( lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : Optional[Any] = simple_accuracy(lowercase__ , lowercase__ ) __SCREAMING_SNAKE_CASE : List[str] = float(fa_score(y_true=lowercase__ , y_pred=lowercase__ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : Optional[int] = np.array(lowercase__ ) __SCREAMING_SNAKE_CASE : str = np.array(lowercase__ ) __SCREAMING_SNAKE_CASE : str = en_sentvecs.shape[0] # mean centering __SCREAMING_SNAKE_CASE : Tuple = en_sentvecs - np.mean(lowercase__ , axis=0 ) __SCREAMING_SNAKE_CASE : Optional[int] = in_sentvecs - np.mean(lowercase__ , axis=0 ) __SCREAMING_SNAKE_CASE : str = cdist(lowercase__ , lowercase__ , '''cosine''' ) __SCREAMING_SNAKE_CASE : int = np.array(range(lowercase__ ) ) __SCREAMING_SNAKE_CASE : Optional[Any] = sim.argsort(axis=1 )[:, :10] __SCREAMING_SNAKE_CASE : str = np.any(preds == actual[:, None] , axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): '''simple docstring''' def __magic_name__( self :Tuple ) -> Tuple: if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( '''You should supply a configuration name selected in ''' '''["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ''' '''"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ''' '''"wiki-ner"]''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int64''' ) if self.config_name != '''cvit-mkb-clsr''' else datasets.Sequence(datasets.Value('''float32''' ) ), '''references''': datasets.Value('''int64''' ) if self.config_name != '''cvit-mkb-clsr''' else datasets.Sequence(datasets.Value('''float32''' ) ), } ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' if self.config_name != '''cvit-mkb-clsr''' else None , ) def __magic_name__( self :List[str] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple ) -> str: if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(lowerCAmelCase__ , lowerCAmelCase__ )} elif self.config_name in ["wiki-ner"]: return acc_and_fa(lowerCAmelCase__ , lowerCAmelCase__ ) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(lowerCAmelCase__ , lowerCAmelCase__ )} else: raise KeyError( '''You should supply a configuration name selected in ''' '''["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ''' '''"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ''' '''"wiki-ner"]''' )
9
0
"""simple docstring""" def lowercase_ ( ) -> int: return 1 def lowercase_ ( __UpperCAmelCase ) -> int: return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def lowercase_ ( __UpperCAmelCase ) -> int: return 0 if x < 0 else five_pence(x - 5 ) + two_pence(__UpperCAmelCase ) def lowercase_ ( __UpperCAmelCase ) -> int: return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(__UpperCAmelCase ) def lowercase_ ( __UpperCAmelCase ) -> int: return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(__UpperCAmelCase ) def lowercase_ ( __UpperCAmelCase ) -> int: return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(__UpperCAmelCase ) def lowercase_ ( __UpperCAmelCase ) -> int: return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(__UpperCAmelCase ) def lowercase_ ( __UpperCAmelCase ) -> int: return 0 if x < 0 else two_pound(x - 200 ) + one_pound(__UpperCAmelCase ) def lowercase_ ( __UpperCAmelCase = 200 ) -> int: return two_pound(__UpperCAmelCase ) if __name__ == "__main__": print(solution(int(input().strip())))
212
"""simple docstring""" from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _lowerCamelCase ( a_ ): _lowerCamelCase :Dict = ["image_processor", "tokenizer"] _lowerCamelCase :Dict = "BlipImageProcessor" _lowerCamelCase :Any = ("BertTokenizer", "BertTokenizerFast") def __init__( self : Dict , UpperCamelCase : Optional[int] , UpperCamelCase : Union[str, Any] ) -> str: """simple docstring""" lowerCAmelCase__ : Optional[Any] = False super().__init__(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : List[Any] = self.image_processor def __call__( self : int , UpperCamelCase : ImageInput = None , UpperCamelCase : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , UpperCamelCase : bool = True , UpperCamelCase : Union[bool, str, PaddingStrategy] = False , UpperCamelCase : Union[bool, str, TruncationStrategy] = None , UpperCamelCase : Optional[int] = None , UpperCamelCase : int = 0 , UpperCamelCase : Optional[int] = None , UpperCamelCase : Optional[bool] = None , UpperCamelCase : bool = False , UpperCamelCase : bool = False , UpperCamelCase : bool = False , UpperCamelCase : bool = False , UpperCamelCase : bool = False , UpperCamelCase : bool = True , UpperCamelCase : Optional[Union[str, TensorType]] = None , **UpperCamelCase : Optional[int] , ) -> BatchEncoding: """simple docstring""" if images is None and text is None: raise ValueError("""You have to specify either images or text.""" ) # Get only text if images is None: lowerCAmelCase__ : Any = self.tokenizer lowerCAmelCase__ : Optional[int] = self.tokenizer( text=UpperCamelCase , add_special_tokens=UpperCamelCase , padding=UpperCamelCase , truncation=UpperCamelCase , max_length=UpperCamelCase , stride=UpperCamelCase , pad_to_multiple_of=UpperCamelCase , return_attention_mask=UpperCamelCase , return_overflowing_tokens=UpperCamelCase , return_special_tokens_mask=UpperCamelCase , return_offsets_mapping=UpperCamelCase , return_token_type_ids=UpperCamelCase , return_length=UpperCamelCase , verbose=UpperCamelCase , return_tensors=UpperCamelCase , **UpperCamelCase , ) return text_encoding # add pixel_values lowerCAmelCase__ : Tuple = self.image_processor(UpperCamelCase , return_tensors=UpperCamelCase ) if text is not None: lowerCAmelCase__ : Optional[int] = self.tokenizer( text=UpperCamelCase , add_special_tokens=UpperCamelCase , padding=UpperCamelCase , truncation=UpperCamelCase , max_length=UpperCamelCase , stride=UpperCamelCase , pad_to_multiple_of=UpperCamelCase , return_attention_mask=UpperCamelCase , return_overflowing_tokens=UpperCamelCase , return_special_tokens_mask=UpperCamelCase , return_offsets_mapping=UpperCamelCase , return_token_type_ids=UpperCamelCase , return_length=UpperCamelCase , verbose=UpperCamelCase , return_tensors=UpperCamelCase , **UpperCamelCase , ) else: lowerCAmelCase__ : Tuple = None if text_encoding is not None: encoding_image_processor.update(UpperCamelCase ) return encoding_image_processor def _lowerCAmelCase ( self : int , *UpperCamelCase : Union[str, Any] , **UpperCamelCase : List[str] ) -> Dict: """simple docstring""" return self.tokenizer.batch_decode(*UpperCamelCase , **UpperCamelCase ) def _lowerCAmelCase ( self : List[Any] , *UpperCamelCase : Tuple , **UpperCamelCase : List[str] ) -> List[Any]: """simple docstring""" return self.tokenizer.decode(*UpperCamelCase , **UpperCamelCase ) @property def _lowerCAmelCase ( self : List[Any] ) -> Tuple: """simple docstring""" lowerCAmelCase__ : Optional[int] = self.tokenizer.model_input_names lowerCAmelCase__ : Union[str, Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
212
1
"""simple docstring""" import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def snake_case ( self : int )-> str: lowerCamelCase__ : Optional[int] =AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) lowerCamelCase__ : Optional[Any] =AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(lowerCamelCase ) lowerCamelCase__ : Any =-1 lowerCamelCase__ : Optional[int] =ids_tensor((1, 5), vocab_size=model.config.vocab_size ).to(lowerCamelCase ) lowerCamelCase__ : str =model.generate(lowerCamelCase, max_new_tokens=10, do_sample=lowerCamelCase ) lowerCamelCase__ : List[str] =tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: lowerCamelCase__ : Tuple =TextStreamer(lowerCamelCase ) model.generate(lowerCamelCase, max_new_tokens=10, do_sample=lowerCamelCase, streamer=lowerCamelCase ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer lowerCamelCase__ : Tuple =cs.out[:-1] self.assertEqual(lowerCamelCase, lowerCamelCase ) def snake_case ( self : List[str] )-> Optional[Any]: lowerCamelCase__ : Optional[Any] =AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) lowerCamelCase__ : int =AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(lowerCamelCase ) lowerCamelCase__ : Any =-1 lowerCamelCase__ : List[str] =ids_tensor((1, 5), vocab_size=model.config.vocab_size ).to(lowerCamelCase ) lowerCamelCase__ : List[Any] =model.generate(lowerCamelCase, max_new_tokens=10, do_sample=lowerCamelCase ) lowerCamelCase__ : str =tokenizer.decode(greedy_ids[0] ) lowerCamelCase__ : Optional[int] =TextIteratorStreamer(lowerCamelCase ) lowerCamelCase__ : int ={'''input_ids''': input_ids, '''max_new_tokens''': 10, '''do_sample''': False, '''streamer''': streamer} lowerCamelCase__ : int =Thread(target=model.generate, kwargs=lowerCamelCase ) thread.start() lowerCamelCase__ : List[str] ='''''' for new_text in streamer: streamer_text += new_text self.assertEqual(lowerCamelCase, lowerCamelCase ) def snake_case ( self : int )-> str: lowerCamelCase__ : Union[str, Any] =AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) lowerCamelCase__ : Any =AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(lowerCamelCase ) lowerCamelCase__ : Optional[Any] =-1 lowerCamelCase__ : List[Any] =ids_tensor((1, 5), vocab_size=model.config.vocab_size ).to(lowerCamelCase ) lowerCamelCase__ : Any =model.generate(lowerCamelCase, max_new_tokens=10, do_sample=lowerCamelCase ) lowerCamelCase__ : Optional[Any] =greedy_ids[:, input_ids.shape[1] :] lowerCamelCase__ : Tuple =tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: lowerCamelCase__ : Dict =TextStreamer(lowerCamelCase, skip_prompt=lowerCamelCase ) model.generate(lowerCamelCase, max_new_tokens=10, do_sample=lowerCamelCase, streamer=lowerCamelCase ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer lowerCamelCase__ : Tuple =cs.out[:-1] self.assertEqual(lowerCamelCase, lowerCamelCase ) def snake_case ( self : Any )-> Any: # Tests that we can pass `decode_kwargs` to the streamer to control how the tokens are decoded. Must be tested # with actual models -- the dummy models' tokenizers are not aligned with their models, and # `skip_special_tokens=True` has no effect on them lowerCamelCase__ : Tuple =AutoTokenizer.from_pretrained('''distilgpt2''' ) lowerCamelCase__ : List[str] =AutoModelForCausalLM.from_pretrained('''distilgpt2''' ).to(lowerCamelCase ) lowerCamelCase__ : List[Any] =-1 lowerCamelCase__ : Any =torch.ones((1, 5), device=lowerCamelCase ).long() * model.config.bos_token_id with CaptureStdout() as cs: lowerCamelCase__ : str =TextStreamer(lowerCamelCase, skip_special_tokens=lowerCamelCase ) model.generate(lowerCamelCase, max_new_tokens=1, do_sample=lowerCamelCase, streamer=lowerCamelCase ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token lowerCamelCase__ : List[Any] =cs.out[:-1] # Remove the final "\n" lowerCamelCase__ : List[str] =tokenizer(lowerCamelCase, return_tensors='''pt''' ) self.assertEqual(streamer_text_tokenized.input_ids.shape, (1, 1) ) def snake_case ( self : Optional[Any] )-> Optional[Any]: lowerCamelCase__ : str =AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) lowerCamelCase__ : Optional[int] =AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(lowerCamelCase ) lowerCamelCase__ : Optional[Any] =-1 lowerCamelCase__ : int =ids_tensor((1, 5), vocab_size=model.config.vocab_size ).to(lowerCamelCase ) lowerCamelCase__ : Tuple =TextIteratorStreamer(lowerCamelCase, timeout=0.001 ) lowerCamelCase__ : Optional[Any] ={'''input_ids''': input_ids, '''max_new_tokens''': 10, '''do_sample''': False, '''streamer''': streamer} lowerCamelCase__ : List[str] =Thread(target=model.generate, kwargs=lowerCamelCase ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(lowerCamelCase ): lowerCamelCase__ : str ='''''' for new_text in streamer: streamer_text += new_text
238
"""simple docstring""" import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import VideoMAEConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEModel, ) from transformers.models.videomae.modeling_videomae import VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Optional[Any], lowerCamelCase : Any, lowerCamelCase : List[Any]=13, lowerCamelCase : Any=10, lowerCamelCase : Optional[Any]=3, lowerCamelCase : Union[str, Any]=2, lowerCamelCase : Dict=2, lowerCamelCase : Tuple=2, lowerCamelCase : List[str]=True, lowerCamelCase : Optional[int]=True, lowerCamelCase : Dict=32, lowerCamelCase : Any=5, lowerCamelCase : Dict=4, lowerCamelCase : Any=37, lowerCamelCase : Union[str, Any]="gelu", lowerCamelCase : Dict=0.1, lowerCamelCase : Union[str, Any]=0.1, lowerCamelCase : Dict=10, lowerCamelCase : str=0.02, lowerCamelCase : List[Any]=0.9, lowerCamelCase : List[Any]=None, )-> str: lowerCamelCase__ : List[str] =parent lowerCamelCase__ : Any =batch_size lowerCamelCase__ : str =image_size lowerCamelCase__ : Optional[Any] =num_channels lowerCamelCase__ : Optional[int] =patch_size lowerCamelCase__ : List[str] =tubelet_size lowerCamelCase__ : Optional[Any] =num_frames lowerCamelCase__ : Any =is_training lowerCamelCase__ : List[Any] =use_labels lowerCamelCase__ : Union[str, Any] =hidden_size lowerCamelCase__ : List[str] =num_hidden_layers lowerCamelCase__ : str =num_attention_heads lowerCamelCase__ : List[Any] =intermediate_size lowerCamelCase__ : Any =hidden_act lowerCamelCase__ : int =hidden_dropout_prob lowerCamelCase__ : Optional[int] =attention_probs_dropout_prob lowerCamelCase__ : Optional[Any] =type_sequence_label_size lowerCamelCase__ : int =initializer_range lowerCamelCase__ : Optional[Any] =mask_ratio lowerCamelCase__ : Any =scope # in VideoMAE, the number of tokens equals num_frames/tubelet_size * num_patches per frame lowerCamelCase__ : Optional[Any] =(image_size // patch_size) ** 2 lowerCamelCase__ : Any =(num_frames // tubelet_size) * self.num_patches_per_frame # use this variable to define bool_masked_pos lowerCamelCase__ : List[Any] =int(mask_ratio * self.seq_length ) def snake_case ( self : Dict )-> Union[str, Any]: lowerCamelCase__ : str =floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Any =None if self.use_labels: lowerCamelCase__ : Union[str, Any] =ids_tensor([self.batch_size], self.type_sequence_label_size ) lowerCamelCase__ : Optional[Any] =self.get_config() return config, pixel_values, labels def snake_case ( self : Union[str, Any] )-> Optional[int]: return VideoMAEConfig( image_size=self.image_size, patch_size=self.patch_size, num_channels=self.num_channels, num_frames=self.num_frames, tubelet_size=self.tubelet_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, is_decoder=lowerCamelCase, initializer_range=self.initializer_range, ) def snake_case ( self : Dict, lowerCamelCase : Tuple, lowerCamelCase : Optional[Any], lowerCamelCase : Any )-> Union[str, Any]: lowerCamelCase__ : List[str] =VideoMAEModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() lowerCamelCase__ : int =model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self : Any, lowerCamelCase : str, lowerCamelCase : Optional[int], lowerCamelCase : str )-> Dict: lowerCamelCase__ : int =VideoMAEForPreTraining(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch lowerCamelCase__ : Optional[int] =torch.ones((self.num_masks,) ) lowerCamelCase__ : List[str] =torch.cat([mask, torch.zeros(self.seq_length - mask.size(0 ) )] ) lowerCamelCase__ : int =mask.expand(self.batch_size, -1 ).bool() lowerCamelCase__ : Any =model(lowerCamelCase, lowerCamelCase ) # model only returns predictions for masked patches lowerCamelCase__ : Optional[int] =mask.sum().item() lowerCamelCase__ : Dict =3 * self.tubelet_size * self.patch_size**2 self.parent.assertEqual(result.logits.shape, (self.batch_size, num_masked_patches, decoder_num_labels) ) def snake_case ( self : Optional[Any] )-> Tuple: lowerCamelCase__ : Tuple =self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Dict =config_and_inputs lowerCamelCase__ : List[str] ={'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' _a = ( (VideoMAEModel, VideoMAEForPreTraining, VideoMAEForVideoClassification) if is_torch_available() else () ) _a = ( {'feature-extraction': VideoMAEModel, 'video-classification': VideoMAEForVideoClassification} if is_torch_available() else {} ) _a = False _a = False _a = False _a = False def snake_case ( self : List[Any] )-> Tuple: lowerCamelCase__ : int =VideoMAEModelTester(self ) lowerCamelCase__ : Optional[int] =ConfigTester(self, config_class=lowerCamelCase, has_text_modality=lowerCamelCase, hidden_size=37 ) def snake_case ( self : Any, lowerCamelCase : List[Any], lowerCamelCase : Optional[Any], lowerCamelCase : List[str]=False )-> Tuple: lowerCamelCase__ : str =copy.deepcopy(lowerCamelCase ) if model_class == VideoMAEForPreTraining: # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch lowerCamelCase__ : Any =torch.ones((self.model_tester.num_masks,) ) lowerCamelCase__ : Dict =torch.cat([mask, torch.zeros(self.model_tester.seq_length - mask.size(0 ) )] ) lowerCamelCase__ : Optional[int] =mask.expand(self.model_tester.batch_size, -1 ).bool() lowerCamelCase__ : int =bool_masked_pos.to(lowerCamelCase ) if return_labels: if model_class in [ *get_values(lowerCamelCase ), ]: lowerCamelCase__ : List[str] =torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase ) return inputs_dict def snake_case ( self : List[Any] )-> int: self.config_tester.run_common_tests() @unittest.skip(reason='''VideoMAE does not use inputs_embeds''' ) def snake_case ( self : List[str] )-> Tuple: pass def snake_case ( self : Union[str, Any] )-> Union[str, Any]: lowerCamelCase__ , lowerCamelCase__ : Optional[int] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[str] =model_class(lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings(), (nn.Module) ) lowerCamelCase__ : Optional[Any] =model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase, nn.Linear ) ) def snake_case ( self : Optional[int] )-> Optional[Any]: lowerCamelCase__ , lowerCamelCase__ : Any =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] =model_class(lowerCamelCase ) lowerCamelCase__ : Dict =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : Tuple =[*signature.parameters.keys()] lowerCamelCase__ : List[str] =['''pixel_values'''] self.assertListEqual(arg_names[:1], lowerCamelCase ) def snake_case ( self : Tuple )-> Optional[int]: lowerCamelCase__ : Any =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def snake_case ( self : List[Any] )-> Union[str, Any]: lowerCamelCase__ : Optional[int] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase ) @slow def snake_case ( self : List[Any] )-> Dict: for model_name in VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : str =VideoMAEModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def snake_case ( self : List[str] )-> Optional[int]: if not self.has_attentions: pass else: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] =self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple =True for model_class in self.all_model_classes: lowerCamelCase__ : Any =self.model_tester.seq_length - self.model_tester.num_masks lowerCamelCase__ : Any =( num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length ) lowerCamelCase__ : Optional[int] =True lowerCamelCase__ : Optional[int] =False lowerCamelCase__ : Optional[int] =True lowerCamelCase__ : int =model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : Union[str, Any] =model(**self._prepare_for_class(lowerCamelCase, lowerCamelCase ) ) lowerCamelCase__ : str =outputs.attentions self.assertEqual(len(lowerCamelCase ), self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCamelCase__ : Tuple =True lowerCamelCase__ : Union[str, Any] =model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : List[str] =model(**self._prepare_for_class(lowerCamelCase, lowerCamelCase ) ) lowerCamelCase__ : int =outputs.attentions self.assertEqual(len(lowerCamelCase ), self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ), [self.model_tester.num_attention_heads, seq_len, seq_len], ) lowerCamelCase__ : Union[str, Any] =len(lowerCamelCase ) # Check attention is always last and order is fine lowerCamelCase__ : List[Any] =True lowerCamelCase__ : Union[str, Any] =True lowerCamelCase__ : Dict =model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : Any =model(**self._prepare_for_class(lowerCamelCase, lowerCamelCase ) ) self.assertEqual(out_len + 1, len(lowerCamelCase ) ) lowerCamelCase__ : Optional[Any] =outputs.attentions self.assertEqual(len(lowerCamelCase ), self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ), [self.model_tester.num_attention_heads, seq_len, seq_len], ) def snake_case ( self : str )-> int: def check_hidden_states_output(lowerCamelCase : Optional[Any], lowerCamelCase : List[str], lowerCamelCase : Optional[Any] ): lowerCamelCase__ : List[Any] =model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : Optional[Any] =model(**self._prepare_for_class(lowerCamelCase, lowerCamelCase ) ) lowerCamelCase__ : Dict =outputs.hidden_states lowerCamelCase__ : Any =self.model_tester.num_hidden_layers + 1 self.assertEqual(len(lowerCamelCase ), lowerCamelCase ) lowerCamelCase__ : Any =self.model_tester.seq_length - self.model_tester.num_masks lowerCamelCase__ : str =num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ), [seq_length, self.model_tester.hidden_size], ) lowerCamelCase__ , lowerCamelCase__ : List[str] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Union[str, Any] =True check_hidden_states_output(lowerCamelCase, lowerCamelCase, lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : int =True check_hidden_states_output(lowerCamelCase, lowerCamelCase, lowerCamelCase ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def snake_case ( self : Optional[int] )-> int: pass def snake_case__ ( ): """simple docstring""" lowerCamelCase__ : int =hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''' , filename='''eating_spaghetti.npy''' , repo_type='''dataset''' ) lowerCamelCase__ : str =np.load(__lowerCamelCase ) return list(__lowerCamelCase ) @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case ( self : List[str] )-> List[Any]: # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5], image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def snake_case ( self : Optional[Any] )-> Dict: lowerCamelCase__ : str =VideoMAEForVideoClassification.from_pretrained('''MCG-NJU/videomae-base-finetuned-kinetics''' ).to( lowerCamelCase ) lowerCamelCase__ : Optional[Any] =self.default_image_processor lowerCamelCase__ : List[str] =prepare_video() lowerCamelCase__ : Union[str, Any] =image_processor(lowerCamelCase, return_tensors='''pt''' ).to(lowerCamelCase ) # forward pass with torch.no_grad(): lowerCamelCase__ : Tuple =model(**lowerCamelCase ) # verify the logits lowerCamelCase__ : Union[str, Any] =torch.Size((1, 400) ) self.assertEqual(outputs.logits.shape, lowerCamelCase ) lowerCamelCase__ : Tuple =torch.tensor([0.3_669, -0.0_688, -0.2_421] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3], lowerCamelCase, atol=1E-4 ) ) @slow def snake_case ( self : Any )-> Tuple: lowerCamelCase__ : Tuple =VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''' ).to(lowerCamelCase ) lowerCamelCase__ : Optional[int] =self.default_image_processor lowerCamelCase__ : Dict =prepare_video() lowerCamelCase__ : Dict =image_processor(lowerCamelCase, return_tensors='''pt''' ).to(lowerCamelCase ) # add boolean mask, indicating which patches to mask lowerCamelCase__ : str =hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''', filename='''bool_masked_pos.pt''' ) lowerCamelCase__ : Dict =torch.load(lowerCamelCase ) # forward pass with torch.no_grad(): lowerCamelCase__ : Union[str, Any] =model(**lowerCamelCase ) # verify the logits lowerCamelCase__ : Dict =torch.Size([1, 1408, 1536] ) lowerCamelCase__ : Union[str, Any] =torch.tensor( [[0.7_994, 0.9_612, 0.8_508], [0.7_401, 0.8_958, 0.8_302], [0.5_862, 0.7_468, 0.7_325]], device=lowerCamelCase ) self.assertEqual(outputs.logits.shape, lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3], lowerCamelCase, atol=1E-4 ) ) # verify the loss (`config.norm_pix_loss` = `True`) lowerCamelCase__ : Optional[int] =torch.tensor([0.5_142], device=lowerCamelCase ) self.assertTrue(torch.allclose(outputs.loss, lowerCamelCase, atol=1E-4 ) ) # verify the loss (`config.norm_pix_loss` = `False`) lowerCamelCase__ : Union[str, Any] =VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''', norm_pix_loss=lowerCamelCase ).to( lowerCamelCase ) with torch.no_grad(): lowerCamelCase__ : Union[str, Any] =model(**lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =torch.tensor(torch.tensor([0.6_469] ), device=lowerCamelCase ) self.assertTrue(torch.allclose(outputs.loss, lowerCamelCase, atol=1E-4 ) )
238
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_fnet import FNetTokenizer else: snake_case__ : Optional[Any] = None snake_case__ : Optional[int] = logging.get_logger(__name__) snake_case__ : Any = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} snake_case__ : Optional[Any] = { '''vocab_file''': { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/spiece.model''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/tokenizer.json''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/tokenizer.json''', }, } snake_case__ : Dict = { '''google/fnet-base''': 512, '''google/fnet-large''': 512, } snake_case__ : Tuple = '''▁''' class __SCREAMING_SNAKE_CASE ( lowerCamelCase_ ): '''simple docstring''' lowerCamelCase_ :Optional[int] = VOCAB_FILES_NAMES lowerCamelCase_ :int = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ :List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ :Optional[Any] = ['''input_ids''', '''token_type_ids'''] lowerCamelCase_ :Union[str, Any] = FNetTokenizer def __init__( self , snake_case_=None , snake_case_=None , snake_case_=False , snake_case_=True , snake_case_=True , snake_case_="<unk>" , snake_case_="[SEP]" , snake_case_="<pad>" , snake_case_="[CLS]" , snake_case_="[MASK]" , **snake_case_ , ): '''simple docstring''' UpperCAmelCase_ : Any = ( AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ , normalized=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else mask_token ) super().__init__( snake_case_ , tokenizer_file=snake_case_ , do_lower_case=snake_case_ , remove_space=snake_case_ , keep_accents=snake_case_ , unk_token=snake_case_ , sep_token=snake_case_ , pad_token=snake_case_ , cls_token=snake_case_ , mask_token=snake_case_ , **snake_case_ , ) UpperCAmelCase_ : Dict = do_lower_case UpperCAmelCase_ : Optional[Any] = remove_space UpperCAmelCase_ : Any = keep_accents UpperCAmelCase_ : Tuple = vocab_file UpperCAmelCase_ : List[str] = False if not self.vocab_file else True def _UpperCamelCase ( self , snake_case_ , snake_case_ = None ): '''simple docstring''' UpperCAmelCase_ : str = [self.sep_token_id] UpperCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _UpperCamelCase ( self , snake_case_ , snake_case_ = None ): '''simple docstring''' UpperCAmelCase_ : Tuple = [self.sep_token_id] UpperCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _UpperCamelCase ( self , snake_case_ , snake_case_ = None ): '''simple docstring''' if not os.path.isdir(snake_case_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase_ : Any = os.path.join( snake_case_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ): copyfile(self.vocab_file , snake_case_ ) return (out_vocab_file,)
274
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS snake_case__ : Optional[Any] = logging.get_logger(__name__) snake_case__ : List[Any] = { '''linear''': get_linear_schedule_with_warmup, '''cosine''': get_cosine_schedule_with_warmup, '''cosine_w_restarts''': get_cosine_with_hard_restarts_schedule_with_warmup, '''polynomial''': get_polynomial_decay_schedule_with_warmup, '''constant''': get_constant_schedule, '''constant_w_warmup''': get_constant_schedule_with_warmup, } class __SCREAMING_SNAKE_CASE ( lowerCamelCase_ ): '''simple docstring''' def __init__( self , snake_case_=None , snake_case_=None , *snake_case_ , **snake_case_ ): '''simple docstring''' super().__init__(*snake_case_ , **snake_case_ ) if config is None: assert isinstance(self.model , snake_case_ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F''' {self.model.__class__}''' ) UpperCAmelCase_ : Tuple = self.model.config else: UpperCAmelCase_ : Optional[Any] = config UpperCAmelCase_ : Optional[Any] = data_args UpperCAmelCase_ : Dict = self.config.tgt_vocab_size if isinstance(self.config , snake_case_ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F'''The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for''' ' padding..' ) if self.args.label_smoothing == 0: UpperCAmelCase_ : Dict = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss UpperCAmelCase_ : Union[str, Any] = label_smoothed_nll_loss def _UpperCamelCase ( self , snake_case_ ): '''simple docstring''' if self.optimizer is None: UpperCAmelCase_ : Optional[Any] = ['bias', 'LayerNorm.weight'] UpperCAmelCase_ : Union[str, Any] = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] UpperCAmelCase_ : Optional[Any] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: UpperCAmelCase_ : List[str] = Adafactor UpperCAmelCase_ : int = {'scale_parameter': False, 'relative_step': False} else: UpperCAmelCase_ : Union[str, Any] = AdamW UpperCAmelCase_ : Optional[int] = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } UpperCAmelCase_ : Optional[int] = self.args.learning_rate if self.sharded_ddp: UpperCAmelCase_ : Optional[Any] = OSS( params=snake_case_ , optim=snake_case_ , **snake_case_ , ) else: UpperCAmelCase_ : Tuple = optimizer_cls(snake_case_ , **snake_case_ ) if self.lr_scheduler is None: UpperCAmelCase_ : int = self._get_lr_scheduler(snake_case_ ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def _UpperCamelCase ( self , snake_case_ ): '''simple docstring''' UpperCAmelCase_ : int = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": UpperCAmelCase_ : List[Any] = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": UpperCAmelCase_ : Optional[Any] = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: UpperCAmelCase_ : int = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=snake_case_ ) return scheduler def _UpperCamelCase ( self ): '''simple docstring''' if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def _UpperCamelCase ( self , snake_case_ , snake_case_ , snake_case_ ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token UpperCAmelCase_ : Any = model(**snake_case_ , use_cache=snake_case_ )[0] UpperCAmelCase_ : int = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models UpperCAmelCase_ , UpperCAmelCase_ : List[str] = model(**snake_case_ , labels=snake_case_ , use_cache=snake_case_ )[:2] else: # compute label smoothed loss UpperCAmelCase_ : List[str] = model(**snake_case_ , use_cache=snake_case_ )[0] UpperCAmelCase_ : Optional[int] = torch.nn.functional.log_softmax(snake_case_ , dim=-1 ) UpperCAmelCase_ , UpperCAmelCase_ : Dict = self.loss_fn(snake_case_ , snake_case_ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def _UpperCamelCase ( self , snake_case_ , snake_case_ ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = inputs.pop('labels' ) UpperCAmelCase_ , UpperCAmelCase_ : Dict = self._compute_loss(snake_case_ , snake_case_ , snake_case_ ) return loss def _UpperCamelCase ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ = None , ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = self._prepare_inputs(snake_case_ ) UpperCAmelCase_ : Union[str, Any] = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: UpperCAmelCase_ : Tuple = self.model.generate( inputs['input_ids'] , attention_mask=inputs['attention_mask'] , **snake_case_ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: UpperCAmelCase_ : Tuple = self._pad_tensors_to_max_len(snake_case_ , gen_kwargs['max_length'] ) UpperCAmelCase_ : List[str] = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = self._compute_loss(snake_case_ , snake_case_ , snake_case_ ) UpperCAmelCase_ : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) UpperCAmelCase_ : str = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: UpperCAmelCase_ : List[Any] = self._pad_tensors_to_max_len(snake_case_ , gen_kwargs['max_length'] ) return (loss, logits, labels) def _UpperCamelCase ( self , snake_case_ , snake_case_ ): '''simple docstring''' UpperCAmelCase_ : Any = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F''' padded to `max_length`={max_length}''' ) UpperCAmelCase_ : Tuple = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) UpperCAmelCase_ : Dict = tensor return padded_tensor
274
1
from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline _UpperCAmelCase : Dict = logging.get_logger(__name__) class lowercase ( _SCREAMING_SNAKE_CASE ): def __UpperCamelCase ( self , A_ ) -> int: """simple docstring""" if isinstance(A_ , A_ ): UpperCamelCase = [label.strip() for label in labels.split(',' ) if label.strip()] return labels def __call__( self , A_ , A_ , A_ ) -> List[Any]: """simple docstring""" if len(A_ ) == 0 or len(A_ ) == 0: raise ValueError('You must include at least one label and at least one sequence.' ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( 'The provided hypothesis_template "{}" was not able to be formatted with the target labels. ' 'Make sure the passed template includes formatting syntax such as {{}} where the label should go.' ).format(A_ ) ) if isinstance(A_ , A_ ): UpperCamelCase = [sequences] UpperCamelCase = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(A_ )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(_SCREAMING_SNAKE_CASE ) class lowercase ( _SCREAMING_SNAKE_CASE ): def __init__( self , A_=ZeroShotClassificationArgumentHandler() , *A_ , **A_ ) -> Optional[Any]: """simple docstring""" UpperCamelCase = args_parser super().__init__(*A_ , **A_ ) if self.entailment_id == -1: logger.warning( 'Failed to determine \'entailment\' label id from the label2id mapping in the model config. Setting to ' '-1. Define a descriptive label2id mapping in the model config to ensure correct outputs.' ) @property def __UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" for label, ind in self.model.config.labelaid.items(): if label.lower().startswith('entail' ): return ind return -1 def __UpperCamelCase ( self , A_ , A_=True , A_=True , A_=TruncationStrategy.ONLY_FIRST , **A_ ) -> Any: """simple docstring""" UpperCamelCase = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( 'Tokenizer was not supporting padding necessary for zero-shot, attempting to use ' ' `pad_token=eos_token`' ) UpperCamelCase = self.tokenizer.eos_token try: UpperCamelCase = self.tokenizer( A_ , add_special_tokens=A_ , return_tensors=A_ , padding=A_ , truncation=A_ , ) except Exception as e: if "too short" in str(A_ ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. UpperCamelCase = self.tokenizer( A_ , add_special_tokens=A_ , return_tensors=A_ , padding=A_ , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def __UpperCamelCase ( self , **A_ ) -> Tuple: """simple docstring""" if kwargs.get('multi_class' , A_ ) is not None: UpperCamelCase = kwargs['multi_class'] logger.warning( 'The `multi_class` argument has been deprecated and renamed to `multi_label`. ' '`multi_class` will be removed in a future version of Transformers.' ) UpperCamelCase = {} if "candidate_labels" in kwargs: UpperCamelCase = self._args_parser._parse_labels(kwargs['candidate_labels'] ) if "hypothesis_template" in kwargs: UpperCamelCase = kwargs['hypothesis_template'] UpperCamelCase = {} if "multi_label" in kwargs: UpperCamelCase = kwargs['multi_label'] return preprocess_params, {}, postprocess_params def __call__( self , A_ , *A_ , **A_ , ) -> Tuple: """simple docstring""" if len(A_ ) == 0: pass elif len(A_ ) == 1 and "candidate_labels" not in kwargs: UpperCamelCase = args[0] else: raise ValueError(F'''Unable to understand extra arguments {args}''' ) return super().__call__(A_ , **A_ ) def __UpperCamelCase ( self , A_ , A_=None , A_="This example is {}." ) -> Dict: """simple docstring""" UpperCamelCase , UpperCamelCase = self._args_parser(A_ , A_ , A_ ) for i, (candidate_label, sequence_pair) in enumerate(zip(A_ , A_ ) ): UpperCamelCase = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(A_ ) - 1, **model_input, } def __UpperCamelCase ( self , A_ ) -> Optional[Any]: """simple docstring""" UpperCamelCase = inputs['candidate_label'] UpperCamelCase = inputs['sequence'] UpperCamelCase = {k: inputs[k] for k in self.tokenizer.model_input_names} UpperCamelCase = self.model(**A_ ) UpperCamelCase = { 'candidate_label': candidate_label, 'sequence': sequence, 'is_last': inputs['is_last'], **outputs, } return model_outputs def __UpperCamelCase ( self , A_ , A_=False ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = [outputs['candidate_label'] for outputs in model_outputs] UpperCamelCase = [outputs['sequence'] for outputs in model_outputs] UpperCamelCase = np.concatenate([output['logits'].numpy() for output in model_outputs] ) UpperCamelCase = logits.shape[0] UpperCamelCase = len(A_ ) UpperCamelCase = N // n UpperCamelCase = logits.reshape((num_sequences, n, -1) ) if multi_label or len(A_ ) == 1: # softmax over the entailment vs. contradiction dim for each label independently UpperCamelCase = self.entailment_id UpperCamelCase = -1 if entailment_id == 0 else 0 UpperCamelCase = reshaped_outputs[..., [contradiction_id, entailment_id]] UpperCamelCase = np.exp(A_ ) / np.exp(A_ ).sum(-1 , keepdims=A_ ) UpperCamelCase = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels UpperCamelCase = reshaped_outputs[..., self.entailment_id] UpperCamelCase = np.exp(A_ ) / np.exp(A_ ).sum(-1 , keepdims=A_ ) UpperCamelCase = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
222
import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def A ( lowercase ) -> Optional[Any]: '''simple docstring''' UpperCamelCase = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(lowercase , lowercase ) def A ( lowercase ) -> List[Any]: '''simple docstring''' UpperCamelCase = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: UpperCamelCase = s_dict.pop(lowercase ) elif "subsample" in key: UpperCamelCase = s_dict.pop(lowercase ) def A ( lowercase ) -> Tuple: '''simple docstring''' UpperCamelCase , UpperCamelCase = emb.weight.shape UpperCamelCase = nn.Linear(lowercase , lowercase , bias=lowercase ) UpperCamelCase = emb.weight.data return lin_layer def A ( lowercase , lowercase ) -> Dict: '''simple docstring''' UpperCamelCase = torch.load(lowercase , map_location='cpu' ) UpperCamelCase = mam_aaa['args'] UpperCamelCase = mam_aaa['model'] UpperCamelCase = state_dict['decoder.output_projection.weight'] remove_ignore_keys_(lowercase ) rename_keys(lowercase ) UpperCamelCase = state_dict['decoder.embed_tokens.weight'].shape[0] UpperCamelCase = args.share_decoder_input_output_embed UpperCamelCase = [int(lowercase ) for i in args.conv_kernel_sizes.split(',' )] UpperCamelCase = SpeechaTextConfig( vocab_size=lowercase , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='relu' , num_conv_layers=len(lowercase ) , conv_channels=args.conv_channels , conv_kernel_sizes=lowercase , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=lowercase , num_beams=5 , max_length=200 , use_cache=lowercase , decoder_start_token_id=2 , early_stopping=lowercase , ) UpperCamelCase = SpeechaTextForConditionalGeneration(lowercase ) UpperCamelCase , UpperCamelCase = model.model.load_state_dict(lowercase , strict=lowercase ) if len(lowercase ) > 0 and not set(lowercase ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( 'Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,' f''' but all the following weights are missing {missing}''' ) if tie_embeds: UpperCamelCase = make_linear_from_emb(model.model.decoder.embed_tokens ) else: UpperCamelCase = lm_head_weights model.save_pretrained(lowercase ) if __name__ == "__main__": _UpperCAmelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("--fairseq_path", type=str, help="Path to the fairseq model (.pt) file.") parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") _UpperCAmelCase : Dict = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
222
1
from typing import Dict, Optional import numpy as np import datasets __UpperCAmelCase = "\nIoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union\nbetween the predicted segmentation and the ground truth. For binary (two classes) or multi-class segmentation,\nthe mean IoU of the image is calculated by taking the IoU of each class and averaging them.\n" __UpperCAmelCase = "\nArgs:\n predictions (`List[ndarray]`):\n List of predicted segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n references (`List[ndarray]`):\n List of ground truth segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n num_labels (`int`):\n Number of classes (categories).\n ignore_index (`int`):\n Index that will be ignored during evaluation.\n nan_to_num (`int`, *optional*):\n If specified, NaN values will be replaced by the number defined by the user.\n label_map (`dict`, *optional*):\n If specified, dictionary mapping old label indices to new label indices.\n reduce_labels (`bool`, *optional*, defaults to `False`):\n Whether or not to reduce all label values of segmentation maps by 1. Usually used for datasets where 0 is used for background,\n and background itself is not included in all classes of a dataset (e.g. ADE20k). The background label will be replaced by 255.\n\nReturns:\n `Dict[str, float | ndarray]` comprising various elements:\n - *mean_iou* (`float`):\n Mean Intersection-over-Union (IoU averaged over all categories).\n - *mean_accuracy* (`float`):\n Mean accuracy (averaged over all categories).\n - *overall_accuracy* (`float`):\n Overall accuracy on all images.\n - *per_category_accuracy* (`ndarray` of shape `(num_labels,)`):\n Per category accuracy.\n - *per_category_iou* (`ndarray` of shape `(num_labels,)`):\n Per category IoU.\n\nExamples:\n\n >>> import numpy as np\n\n >>> mean_iou = datasets.load_metric(\"mean_iou\")\n\n >>> # suppose one has 3 different segmentation maps predicted\n >>> predicted_1 = np.array([[1, 2], [3, 4], [5, 255]])\n >>> actual_1 = np.array([[0, 3], [5, 4], [6, 255]])\n\n >>> predicted_2 = np.array([[2, 7], [9, 2], [3, 6]])\n >>> actual_2 = np.array([[1, 7], [9, 2], [3, 6]])\n\n >>> predicted_3 = np.array([[2, 2, 3], [8, 2, 4], [3, 255, 2]])\n >>> actual_3 = np.array([[1, 2, 2], [8, 2, 1], [3, 255, 1]])\n\n >>> predicted = [predicted_1, predicted_2, predicted_3]\n >>> ground_truth = [actual_1, actual_2, actual_3]\n\n >>> results = mean_iou.compute(predictions=predicted, references=ground_truth, num_labels=10, ignore_index=255, reduce_labels=False)\n >>> print(results) # doctest: +NORMALIZE_WHITESPACE\n {'mean_iou': 0.47750000000000004, 'mean_accuracy': 0.5916666666666666, 'overall_accuracy': 0.5263157894736842, 'per_category_iou': array([0. , 0. , 0.375, 0.4 , 0.5 , 0. , 0.5 , 1. , 1. , 1. ]), 'per_category_accuracy': array([0. , 0. , 0.75 , 0.66666667, 1. , 0. , 0.5 , 1. , 1. , 1. ])}\n" __UpperCAmelCase = "\\n@software{MMSegmentation_Contributors_OpenMMLab_Semantic_Segmentation_2020,\nauthor = {{MMSegmentation Contributors}},\nlicense = {Apache-2.0},\nmonth = {7},\ntitle = {{OpenMMLab Semantic Segmentation Toolbox and Benchmark}},\nurl = {https://github.com/open-mmlab/mmsegmentation},\nyear = {2020}\n}" def lowercase__ ( __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Optional[Any] , __snake_case : Any , __snake_case : Dict = None , __snake_case : Optional[Any] = False , ): '''simple docstring''' if label_map is not None: for old_id, new_id in label_map.items(): UpperCAmelCase_ : str = new_id # turn into Numpy arrays UpperCAmelCase_ : int = np.array(lowerCamelCase_ ) UpperCAmelCase_ : int = np.array(lowerCamelCase_ ) if reduce_labels: UpperCAmelCase_ : List[str] = 255 UpperCAmelCase_ : Any = label - 1 UpperCAmelCase_ : Union[str, Any] = 255 UpperCAmelCase_ : Optional[Any] = label != ignore_index UpperCAmelCase_ : Tuple = np.not_equal(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = pred_label[mask] UpperCAmelCase_ : Dict = np.array(lowerCamelCase_ )[mask] UpperCAmelCase_ : Optional[Any] = pred_label[pred_label == label] UpperCAmelCase_ : Optional[Any] = np.histogram(lowerCamelCase_ , bins=lowerCamelCase_ , range=(0, num_labels - 1) )[0] UpperCAmelCase_ : Optional[int] = np.histogram(lowerCamelCase_ , bins=lowerCamelCase_ , range=(0, num_labels - 1) )[0] UpperCAmelCase_ : Tuple = np.histogram(lowerCamelCase_ , bins=lowerCamelCase_ , range=(0, num_labels - 1) )[0] UpperCAmelCase_ : List[str] = area_pred_label + area_label - area_intersect return area_intersect, area_union, area_pred_label, area_label def lowercase__ ( __snake_case : str , __snake_case : Tuple , __snake_case : Any , __snake_case : int , __snake_case : int = None , __snake_case : Optional[int] = False , ): '''simple docstring''' UpperCAmelCase_ : str = np.zeros((num_labels,) , dtype=np.floataa ) UpperCAmelCase_ : Any = np.zeros((num_labels,) , dtype=np.floataa ) UpperCAmelCase_ : List[Any] = np.zeros((num_labels,) , dtype=np.floataa ) UpperCAmelCase_ : Tuple = np.zeros((num_labels,) , dtype=np.floataa ) for result, gt_seg_map in zip(lowerCamelCase_ , lowerCamelCase_ ): UpperCAmelCase_ : List[str] = intersect_and_union( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) total_area_intersect += area_intersect total_area_union += area_union total_area_pred_label += area_pred_label total_area_label += area_label return total_area_intersect, total_area_union, total_area_pred_label, total_area_label def lowercase__ ( __snake_case : List[Any] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : Tuple , __snake_case : int = None , __snake_case : str = None , __snake_case : Any = False , ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = total_intersect_and_union( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # compute metrics UpperCAmelCase_ : Tuple = {} UpperCAmelCase_ : List[Any] = total_area_intersect.sum() / total_area_label.sum() UpperCAmelCase_ : List[str] = total_area_intersect / total_area_union UpperCAmelCase_ : Optional[Any] = total_area_intersect / total_area_label UpperCAmelCase_ : Tuple = np.nanmean(lowerCamelCase_ ) UpperCAmelCase_ : Dict = np.nanmean(lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = all_acc UpperCAmelCase_ : int = iou UpperCAmelCase_ : Union[str, Any] = acc if nan_to_num is not None: UpperCAmelCase_ : Dict = {metric: np.nan_to_num(lowerCamelCase_ , nan=lowerCamelCase_ ) for metric, metric_value in metrics.items()} return metrics @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase (datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( # 1st Seq - height dim, 2nd - width dim { 'predictions': datasets.Sequence(datasets.Sequence(datasets.Value('uint16' ) ) ), 'references': datasets.Sequence(datasets.Sequence(datasets.Value('uint16' ) ) ), } ) , reference_urls=[ 'https://github.com/open-mmlab/mmsegmentation/blob/71c201b1813267d78764f306a297ca717827c4bf/mmseg/core/evaluation/metrics.py' ] , ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = False , ) -> List[Any]: UpperCAmelCase_ : Any = mean_iou( results=_UpperCamelCase , gt_seg_maps=_UpperCamelCase , num_labels=_UpperCamelCase , ignore_index=_UpperCamelCase , nan_to_num=_UpperCamelCase , label_map=_UpperCamelCase , reduce_labels=_UpperCamelCase , ) return iou_result
368
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : Any = (DDPMParallelScheduler,) def __UpperCAmelCase ( self , **_UpperCamelCase ) -> Any: UpperCAmelCase_ : List[str] = { 'num_train_timesteps': 1_0_0_0, 'beta_start': 0.00_01, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**_UpperCamelCase ) return config def __UpperCAmelCase ( self ) -> Any: for timesteps in [1, 5, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=_UpperCamelCase ) def __UpperCAmelCase ( self ) -> int: for beta_start, beta_end in zip([0.00_01, 0.0_01, 0.01, 0.1] , [0.0_02, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=_UpperCamelCase , beta_end=_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Any: for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Optional[Any]: for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Optional[int]: for clip_sample in [True, False]: self.check_over_configs(clip_sample=_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Union[str, Any]: self.check_over_configs(thresholding=_UpperCamelCase ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=_UpperCamelCase , prediction_type=_UpperCamelCase , sample_max_value=_UpperCamelCase , ) def __UpperCAmelCase ( self ) -> Any: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Union[str, Any]: for t in [0, 5_0_0, 9_9_9]: self.check_over_forward(time_step=_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Dict: UpperCAmelCase_ : Dict = self.scheduler_classes[0] UpperCAmelCase_ : int = self.get_scheduler_config() UpperCAmelCase_ : Union[str, Any] = scheduler_class(**_UpperCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(4_8_7 ) - 0.0_09_79 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(9_9_9 ) - 0.02 ) ) < 1E-5 def __UpperCAmelCase ( self ) -> Dict: UpperCAmelCase_ : List[Any] = self.scheduler_classes[0] UpperCAmelCase_ : Any = self.get_scheduler_config() UpperCAmelCase_ : Union[str, Any] = scheduler_class(**_UpperCamelCase ) UpperCAmelCase_ : Optional[Any] = len(_UpperCamelCase ) UpperCAmelCase_ : Any = self.dummy_model() UpperCAmelCase_ : List[Any] = self.dummy_sample_deter UpperCAmelCase_ : Union[str, Any] = self.dummy_sample_deter + 0.1 UpperCAmelCase_ : str = self.dummy_sample_deter - 0.1 UpperCAmelCase_ : Tuple = samplea.shape[0] UpperCAmelCase_ : Tuple = torch.stack([samplea, samplea, samplea] , dim=0 ) UpperCAmelCase_ : str = torch.arange(_UpperCamelCase )[0:3, None].repeat(1 , _UpperCamelCase ) UpperCAmelCase_ : Any = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) UpperCAmelCase_ : Dict = scheduler.batch_step_no_noise(_UpperCamelCase , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) UpperCAmelCase_ : List[Any] = torch.sum(torch.abs(_UpperCamelCase ) ) UpperCAmelCase_ : Any = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_sum.item() - 11_53.18_33 ) < 1E-2 assert abs(result_mean.item() - 0.50_05 ) < 1E-3 def __UpperCAmelCase ( self ) -> Union[str, Any]: UpperCAmelCase_ : List[str] = self.scheduler_classes[0] UpperCAmelCase_ : List[Any] = self.get_scheduler_config() UpperCAmelCase_ : Any = scheduler_class(**_UpperCamelCase ) UpperCAmelCase_ : Optional[int] = len(_UpperCamelCase ) UpperCAmelCase_ : Optional[int] = self.dummy_model() UpperCAmelCase_ : Optional[int] = self.dummy_sample_deter UpperCAmelCase_ : Union[str, Any] = torch.manual_seed(0 ) for t in reversed(range(_UpperCamelCase ) ): # 1. predict noise residual UpperCAmelCase_ : Union[str, Any] = model(_UpperCamelCase , _UpperCamelCase ) # 2. predict previous mean of sample x_t-1 UpperCAmelCase_ : str = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , generator=_UpperCamelCase ).prev_sample UpperCAmelCase_ : str = pred_prev_sample UpperCAmelCase_ : Optional[int] = torch.sum(torch.abs(_UpperCamelCase ) ) UpperCAmelCase_ : Optional[int] = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_sum.item() - 2_58.96_06 ) < 1E-2 assert abs(result_mean.item() - 0.33_72 ) < 1E-3 def __UpperCAmelCase ( self ) -> int: UpperCAmelCase_ : List[Any] = self.scheduler_classes[0] UpperCAmelCase_ : Optional[int] = self.get_scheduler_config(prediction_type='v_prediction' ) UpperCAmelCase_ : Tuple = scheduler_class(**_UpperCamelCase ) UpperCAmelCase_ : Tuple = len(_UpperCamelCase ) UpperCAmelCase_ : Tuple = self.dummy_model() UpperCAmelCase_ : List[str] = self.dummy_sample_deter UpperCAmelCase_ : Any = torch.manual_seed(0 ) for t in reversed(range(_UpperCamelCase ) ): # 1. predict noise residual UpperCAmelCase_ : Optional[Any] = model(_UpperCamelCase , _UpperCamelCase ) # 2. predict previous mean of sample x_t-1 UpperCAmelCase_ : Any = scheduler.step(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , generator=_UpperCamelCase ).prev_sample UpperCAmelCase_ : List[str] = pred_prev_sample UpperCAmelCase_ : Optional[int] = torch.sum(torch.abs(_UpperCamelCase ) ) UpperCAmelCase_ : Any = torch.mean(torch.abs(_UpperCamelCase ) ) assert abs(result_sum.item() - 2_02.02_96 ) < 1E-2 assert abs(result_mean.item() - 0.26_31 ) < 1E-3 def __UpperCAmelCase ( self ) -> List[Any]: UpperCAmelCase_ : Any = self.scheduler_classes[0] UpperCAmelCase_ : Optional[Any] = self.get_scheduler_config() UpperCAmelCase_ : str = scheduler_class(**_UpperCamelCase ) UpperCAmelCase_ : Optional[int] = [1_0_0, 8_7, 5_0, 1, 0] scheduler.set_timesteps(timesteps=_UpperCamelCase ) UpperCAmelCase_ : Tuple = scheduler.timesteps for i, timestep in enumerate(_UpperCamelCase ): if i == len(_UpperCamelCase ) - 1: UpperCAmelCase_ : List[str] = -1 else: UpperCAmelCase_ : int = timesteps[i + 1] UpperCAmelCase_ : int = scheduler.previous_timestep(_UpperCamelCase ) UpperCAmelCase_ : Optional[Any] = prev_t.item() self.assertEqual(_UpperCamelCase , _UpperCamelCase ) def __UpperCAmelCase ( self ) -> List[Any]: UpperCAmelCase_ : str = self.scheduler_classes[0] UpperCAmelCase_ : Tuple = self.get_scheduler_config() UpperCAmelCase_ : Union[str, Any] = scheduler_class(**_UpperCamelCase ) UpperCAmelCase_ : Optional[int] = [1_0_0, 8_7, 5_0, 5_1, 0] with self.assertRaises(_UpperCamelCase , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Tuple: UpperCAmelCase_ : Any = self.scheduler_classes[0] UpperCAmelCase_ : List[str] = self.get_scheduler_config() UpperCAmelCase_ : List[str] = scheduler_class(**_UpperCamelCase ) UpperCAmelCase_ : Tuple = [1_0_0, 8_7, 5_0, 1, 0] UpperCAmelCase_ : Optional[Any] = len(_UpperCamelCase ) with self.assertRaises(_UpperCamelCase , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=_UpperCamelCase , timesteps=_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Dict: UpperCAmelCase_ : List[str] = self.scheduler_classes[0] UpperCAmelCase_ : Any = self.get_scheduler_config() UpperCAmelCase_ : Optional[Any] = scheduler_class(**_UpperCamelCase ) UpperCAmelCase_ : Dict = [scheduler.config.num_train_timesteps] with self.assertRaises( _UpperCamelCase , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=_UpperCamelCase )
145
0
"""simple docstring""" from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { """t5-small""": """https://huggingface.co/t5-small/resolve/main/config.json""", """t5-base""": """https://huggingface.co/t5-base/resolve/main/config.json""", """t5-large""": """https://huggingface.co/t5-large/resolve/main/config.json""", """t5-3b""": """https://huggingface.co/t5-3b/resolve/main/config.json""", """t5-11b""": """https://huggingface.co/t5-11b/resolve/main/config.json""", } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """t5""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = {"""hidden_size""": """d_model""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , lowercase=32128 , lowercase=512 , lowercase=64 , lowercase=2048 , lowercase=6 , lowercase=None , lowercase=8 , lowercase=32 , lowercase=128 , lowercase=0.1 , lowercase=1E-6 , lowercase=1.0 , lowercase="relu" , lowercase=True , lowercase=True , lowercase=0 , lowercase=1 , **lowercase , ): _lowerCamelCase : Optional[int] = vocab_size _lowerCamelCase : Any = d_model _lowerCamelCase : List[str] = d_kv _lowerCamelCase : Optional[int] = d_ff _lowerCamelCase : Union[str, Any] = num_layers _lowerCamelCase : str = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry _lowerCamelCase : Union[str, Any] = num_heads _lowerCamelCase : Any = relative_attention_num_buckets _lowerCamelCase : Tuple = relative_attention_max_distance _lowerCamelCase : List[str] = dropout_rate _lowerCamelCase : Optional[Any] = layer_norm_epsilon _lowerCamelCase : Dict = initializer_factor _lowerCamelCase : Dict = feed_forward_proj _lowerCamelCase : Union[str, Any] = use_cache _lowerCamelCase : Tuple = self.feed_forward_proj.split('-' ) _lowerCamelCase : Tuple = act_info[-1] _lowerCamelCase : str = act_info[0] == 'gated' if len(lowercase ) > 1 and act_info[0] != "gated" or len(lowercase ) > 2: raise ValueError( F'''`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.''' 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": _lowerCamelCase : int = 'gelu_new' super().__init__( pad_token_id=lowercase , eos_token_id=lowercase , is_encoder_decoder=lowercase , **lowercase , ) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' @property def A_ ( self ): _lowerCamelCase : Any = { 'input_ids': {0: 'batch', 1: 'encoder_sequence'}, 'attention_mask': {0: 'batch', 1: 'encoder_sequence'}, } if self.use_past: _lowerCamelCase : int = 'past_encoder_sequence + sequence' _lowerCamelCase : Union[str, Any] = {0: 'batch'} _lowerCamelCase : Optional[Any] = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: _lowerCamelCase : Tuple = {0: 'batch', 1: 'decoder_sequence'} _lowerCamelCase : Dict = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(lowercase , direction='inputs' ) return common_inputs @property def A_ ( self ): return 13
96
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : Any = data _lowerCamelCase : Node | None = None class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : str = None _lowerCamelCase : str = None def __iter__( self ): _lowerCamelCase : List[str] = self.head while self.head: yield node.data _lowerCamelCase : Optional[int] = node.next if node == self.head: break def __len__( self ): return sum(1 for _ in self ) def __repr__( self ): return "->".join(str(lowercase ) for item in iter(self ) ) def A_ ( self , lowercase ): self.insert_nth(len(self ) , lowercase ) def A_ ( self , lowercase ): self.insert_nth(0 , lowercase ) def A_ ( self , lowercase , lowercase ): if index < 0 or index > len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : List[Any] = Node(lowercase ) if self.head is None: _lowerCamelCase : str = new_node # first node points itself _lowerCamelCase : Union[str, Any] = new_node elif index == 0: # insert at head _lowerCamelCase : List[str] = self.head _lowerCamelCase : str = new_node else: _lowerCamelCase : Union[str, Any] = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : Union[str, Any] = temp.next _lowerCamelCase : List[str] = new_node if index == len(self ) - 1: # insert at tail _lowerCamelCase : Any = new_node def A_ ( self ): return self.delete_nth(0 ) def A_ ( self ): return self.delete_nth(len(self ) - 1 ) def A_ ( self , lowercase = 0 ): if not 0 <= index < len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : Any = self.head if self.head == self.tail: # just one node _lowerCamelCase : List[str] = None elif index == 0: # delete head node _lowerCamelCase : List[str] = self.tail.next.next _lowerCamelCase : Optional[int] = self.head.next else: _lowerCamelCase : Dict = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : int = temp.next _lowerCamelCase : Optional[int] = temp.next.next if index == len(self ) - 1: # delete at tail _lowerCamelCase : List[Any] = temp return delete_node.data def A_ ( self ): return len(self ) == 0 def _snake_case ( ): _lowerCamelCase : Union[str, Any] = CircularLinkedList() assert len(lowercase__ ) == 0 assert circular_linked_list.is_empty() is True assert str(lowercase__ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(lowercase__ ) == i circular_linked_list.insert_nth(lowercase__ , i + 1 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
96
1
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import is_torch_available, logging lowerCAmelCase : Optional[Any] =logging.get_logger(__name__) lowerCAmelCase : str ={ '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/resolve/main/config.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/config.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/config.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/config.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/config.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/config.json''', } class a_ ( _lowerCAmelCase ): __A = "bloom" __A = ["past_key_values"] __A = { "num_hidden_layers": "n_layer", "num_attention_heads": "n_head", } def __init__( self : Tuple , lowercase : Union[str, Any]=250_880 , lowercase : str=64 , lowercase : Optional[Any]=2 , lowercase : str=8 , lowercase : Optional[int]=1e-5 , lowercase : int=0.02 , lowercase : str=True , lowercase : Optional[int]=1 , lowercase : Tuple=2 , lowercase : Any=False , lowercase : Dict=0.0 , lowercase : int=0.0 , lowercase : List[str]=1 , lowercase : Optional[int]=False , **lowercase : Dict , ): """simple docstring""" lowercase_ :int = vocab_size # Backward compatibility with n_embed kwarg lowercase_ :Dict = kwargs.pop("n_embed" , lowercase ) lowercase_ :Tuple = hidden_size if n_embed is None else n_embed lowercase_ :Optional[Any] = n_layer lowercase_ :Any = n_head lowercase_ :Any = layer_norm_epsilon lowercase_ :Dict = initializer_range lowercase_ :Optional[int] = use_cache lowercase_ :Dict = pretraining_tp lowercase_ :Dict = apply_residual_connection_post_layernorm lowercase_ :int = hidden_dropout lowercase_ :int = attention_dropout lowercase_ :Union[str, Any] = bos_token_id lowercase_ :Tuple = eos_token_id lowercase_ :Tuple = slow_but_exact super().__init__(bos_token_id=lowercase , eos_token_id=lowercase , **lowercase ) class a_ ( _lowerCAmelCase ): __A = version.parse("1.12" ) def __init__( self : Tuple , lowercase : PretrainedConfig , lowercase : str = "default" , lowercase : List[PatchingSpec] = None , lowercase : bool = False , ): """simple docstring""" super().__init__(lowercase , task=lowercase , patching_specs=lowercase , use_past=lowercase ) if not getattr(self._config , "pad_token_id" , lowercase ): # TODO: how to do that better? lowercase_ :int = 0 @property def lowercase__ ( self : Any ): """simple docstring""" lowercase_ :List[Any] = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: # BLOOM stores values on dynamic axis 2. For more details see: https://github.com/huggingface/transformers/pull/18344 self.fill_with_past_key_values_(lowercase , direction="inputs" , inverted_values_shape=lowercase ) lowercase_ :Any = {0: "batch", 1: "past_sequence + sequence"} else: lowercase_ :Any = {0: "batch", 1: "sequence"} return common_inputs @property def lowercase__ ( self : str ): """simple docstring""" return self._config.n_layer @property def lowercase__ ( self : Any ): """simple docstring""" return self._config.n_head @property def lowercase__ ( self : int ): """simple docstring""" return 1e-3 def lowercase__ ( self : int , lowercase : "PreTrainedTokenizer" , lowercase : int = -1 , lowercase : int = -1 , lowercase : bool = False , lowercase : Optional["TensorType"] = None , ): """simple docstring""" lowercase_ :Tuple = super(lowercase , self ).generate_dummy_inputs( lowercase , batch_size=lowercase , seq_length=lowercase , is_pair=lowercase , framework=lowercase ) # We need to order the input in the way they appears in the forward() lowercase_ :int = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch lowercase_ , lowercase_ :Any = common_inputs["input_ids"].shape # Not using the same length for past_key_values lowercase_ :Tuple = seqlen + 2 lowercase_ :str = self._config.hidden_size // self.num_attention_heads lowercase_ :str = ( batch * self.num_attention_heads, head_dim, past_key_values_length, ) lowercase_ :Optional[int] = ( batch * self.num_attention_heads, past_key_values_length, head_dim, ) lowercase_ :Any = [ (torch.zeros(lowercase ), torch.zeros(lowercase )) for _ in range(self.num_layers ) ] lowercase_ :str = common_inputs["attention_mask"] if self.use_past: lowercase_ :Optional[int] = ordered_inputs["attention_mask"].dtype lowercase_ :Any = torch.cat( [ordered_inputs["attention_mask"], torch.ones(lowercase , lowercase , dtype=lowercase )] , dim=1 ) return ordered_inputs @property def lowercase__ ( self : Optional[Any] ): """simple docstring""" return 13
147
'''simple docstring''' def UpperCAmelCase_ ( __lowerCamelCase : dict ): lowercase_ :set[int] = set() # To detect a back edge, keep track of vertices currently in the recursion stack lowercase_ :set[int] = set() return any( node not in visited and depth_first_search(__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ) for node in graph ) def UpperCAmelCase_ ( __lowerCamelCase : dict ,__lowerCamelCase : int ,__lowerCamelCase : set ,__lowerCamelCase : set ): visited.add(__lowerCamelCase ) rec_stk.add(__lowerCamelCase ) for node in graph[vertex]: if node not in visited: if depth_first_search(__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(__lowerCamelCase ) return False if __name__ == "__main__": from doctest import testmod testmod()
147
1
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class A ( _UpperCAmelCase ): """simple docstring""" lowerCamelCase = (DPMSolverSinglestepScheduler,) lowerCamelCase = (('num_inference_steps', 25),) def snake_case__ ( self : Tuple,**lowercase_ : Dict )-> Optional[int]: '''simple docstring''' A__ = { 'num_train_timesteps': 1_0_0_0, 'beta_start': 0.0_001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'solver_order': 2, 'prediction_type': 'epsilon', 'thresholding': False, 'sample_max_value': 1.0, 'algorithm_type': 'dpmsolver++', 'solver_type': 'midpoint', 'lambda_min_clipped': -float('inf' ), 'variance_type': None, } config.update(**lowercase_ ) return config def snake_case__ ( self : str,lowercase_ : Optional[Any]=0,**lowercase_ : Any )-> List[Any]: '''simple docstring''' A__ = dict(self.forward_default_kwargs ) A__ = kwargs.pop('num_inference_steps',lowercase_ ) A__ = self.dummy_sample A__ = 0.1 * sample A__ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: A__ = self.get_scheduler_config(**lowercase_ ) A__ = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals A__ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) A__ = scheduler_class.from_pretrained(lowercase_ ) new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals A__ = dummy_past_residuals[: new_scheduler.config.solver_order] A__ , A__ = sample, sample for t in range(lowercase_,time_step + scheduler.config.solver_order + 1 ): A__ = scheduler.step(lowercase_,lowercase_,lowercase_,**lowercase_ ).prev_sample A__ = new_scheduler.step(lowercase_,lowercase_,lowercase_,**lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def snake_case__ ( self : List[str] )-> List[Any]: '''simple docstring''' pass def snake_case__ ( self : Tuple,lowercase_ : Union[str, Any]=0,**lowercase_ : Union[str, Any] )-> Union[str, Any]: '''simple docstring''' A__ = dict(self.forward_default_kwargs ) A__ = kwargs.pop('num_inference_steps',lowercase_ ) A__ = self.dummy_sample A__ = 0.1 * sample A__ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: A__ = self.get_scheduler_config() A__ = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals (must be after setting timesteps) A__ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) A__ = scheduler_class.from_pretrained(lowercase_ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residual (must be after setting timesteps) A__ = dummy_past_residuals[: new_scheduler.config.solver_order] A__ = scheduler.step(lowercase_,lowercase_,lowercase_,**lowercase_ ).prev_sample A__ = new_scheduler.step(lowercase_,lowercase_,lowercase_,**lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def snake_case__ ( self : Optional[Any],lowercase_ : Optional[int]=None,**lowercase_ : int )-> int: '''simple docstring''' if scheduler is None: A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(**lowercase_ ) A__ = scheduler_class(**lowercase_ ) A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(**lowercase_ ) A__ = scheduler_class(**lowercase_ ) A__ = 1_0 A__ = self.dummy_model() A__ = self.dummy_sample_deter scheduler.set_timesteps(lowercase_ ) for i, t in enumerate(scheduler.timesteps ): A__ = model(lowercase_,lowercase_ ) A__ = scheduler.step(lowercase_,lowercase_,lowercase_ ).prev_sample return sample def snake_case__ ( self : Any )-> str: '''simple docstring''' A__ = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) A__ = 5_0 A__ = self.dummy_model() A__ = self.dummy_sample_deter scheduler.set_timesteps(lowercase_ ) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:] ): A__ = model(lowercase_,lowercase_ ) A__ = scheduler.step(lowercase_,lowercase_,lowercase_ ).prev_sample A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_mean.item() - 0.2_574 ) < 1E-3 def snake_case__ ( self : Optional[Any] )-> List[Any]: '''simple docstring''' for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=lowercase_ ) def snake_case__ ( self : int )-> Optional[Any]: '''simple docstring''' A__ = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) A__ = self.full_loop(scheduler=lowercase_ ) A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_mean.item() - 0.2_791 ) < 1E-3 A__ = DEISMultistepScheduler.from_config(scheduler.config ) A__ = DPMSolverMultistepScheduler.from_config(scheduler.config ) A__ = UniPCMultistepScheduler.from_config(scheduler.config ) A__ = DPMSolverSinglestepScheduler.from_config(scheduler.config ) A__ = self.full_loop(scheduler=lowercase_ ) A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_mean.item() - 0.2_791 ) < 1E-3 def snake_case__ ( self : Tuple )-> Any: '''simple docstring''' self.check_over_configs(thresholding=lowercase_ ) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=lowercase_,prediction_type=lowercase_,sample_max_value=lowercase_,algorithm_type='dpmsolver++',solver_order=lowercase_,solver_type=lowercase_,) def snake_case__ ( self : List[Any] )-> int: '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowercase_ ) def snake_case__ ( self : Dict )-> List[Any]: '''simple docstring''' for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=lowercase_,solver_type=lowercase_,prediction_type=lowercase_,algorithm_type=lowercase_,) A__ = self.full_loop( solver_order=lowercase_,solver_type=lowercase_,prediction_type=lowercase_,algorithm_type=lowercase_,) assert not torch.isnan(lowercase_ ).any(), "Samples have nan numbers" def snake_case__ ( self : Optional[int] )-> Tuple: '''simple docstring''' self.check_over_configs(lower_order_final=lowercase_ ) self.check_over_configs(lower_order_final=lowercase_ ) def snake_case__ ( self : Tuple )-> Optional[int]: '''simple docstring''' self.check_over_configs(lambda_min_clipped=-float('inf' ) ) self.check_over_configs(lambda_min_clipped=-5.1 ) def snake_case__ ( self : Optional[Any] )-> Tuple: '''simple docstring''' self.check_over_configs(variance_type=lowercase_ ) self.check_over_configs(variance_type='learned_range' ) def snake_case__ ( self : str )-> Any: '''simple docstring''' for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=lowercase_,time_step=0 ) def snake_case__ ( self : Tuple )-> Tuple: '''simple docstring''' A__ = self.full_loop() A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_mean.item() - 0.2_791 ) < 1E-3 def snake_case__ ( self : Any )-> Union[str, Any]: '''simple docstring''' A__ = self.full_loop(use_karras_sigmas=lowercase_ ) A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_mean.item() - 0.2_248 ) < 1E-3 def snake_case__ ( self : Union[str, Any] )-> Tuple: '''simple docstring''' A__ = self.full_loop(prediction_type='v_prediction' ) A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_mean.item() - 0.1_453 ) < 1E-3 def snake_case__ ( self : Tuple )-> int: '''simple docstring''' A__ = self.full_loop(prediction_type='v_prediction',use_karras_sigmas=lowercase_ ) A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_mean.item() - 0.0_649 ) < 1E-3 def snake_case__ ( self : List[Any] )-> int: '''simple docstring''' A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(thresholding=lowercase_,dynamic_thresholding_ratio=0 ) A__ = scheduler_class(**lowercase_ ) A__ = 1_0 A__ = self.dummy_model() A__ = self.dummy_sample_deter.half() scheduler.set_timesteps(lowercase_ ) for i, t in enumerate(scheduler.timesteps ): A__ = model(lowercase_,lowercase_ ) A__ = scheduler.step(lowercase_,lowercase_,lowercase_ ).prev_sample assert sample.dtype == torch.floataa
7
def _SCREAMING_SNAKE_CASE ( a ) -> bool: return str(a ) == str(a )[::-1] def _SCREAMING_SNAKE_CASE ( a ) -> int: return int(a ) + int(str(a )[::-1] ) def _SCREAMING_SNAKE_CASE ( a = 1_00_00 ) -> int: __A : int = [] for num in range(1 , a ): __A : List[str] = 0 __A : List[Any] = num while iterations < 50: __A : str = sum_reverse(a ) iterations += 1 if is_palindrome(a ): break else: lychrel_nums.append(a ) return len(a ) if __name__ == "__main__": print(F"""{solution() = }""")
280
0
"""simple docstring""" def _UpperCAmelCase ( __lowerCamelCase : int ) -> int: assert ( isinstance(a__ , a__ ) and number_of_steps > 0 ), f'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 _snake_case , _snake_case = 1, 1 for _ in range(number_of_steps - 1 ): _snake_case , _snake_case = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
359
"""simple docstring""" from __future__ import annotations class lowerCAmelCase__ : def __init__( self : Optional[int] , _lowerCamelCase : int = 0 ): _snake_case = key def lowercase ( self : Tuple , _lowerCamelCase : str , _lowerCamelCase : int ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) _snake_case = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(_lowerCamelCase ) ^ key ) for ch in content] def lowercase ( self : List[Any] , _lowerCamelCase : str , _lowerCamelCase : int ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) _snake_case = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(_lowerCamelCase ) ^ key ) for ch in content] def lowercase ( self : Union[str, Any] , _lowerCamelCase : str , _lowerCamelCase : int = 0 ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) _snake_case = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned _snake_case = '''''' for ch in content: ans += chr(ord(_lowerCamelCase ) ^ key ) return ans def lowercase ( self : List[str] , _lowerCamelCase : str , _lowerCamelCase : int = 0 ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) _snake_case = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned _snake_case = '''''' for ch in content: ans += chr(ord(_lowerCamelCase ) ^ key ) return ans def lowercase ( self : List[Any] , _lowerCamelCase : str , _lowerCamelCase : int = 0 ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) try: with open(_lowerCamelCase ) as fin, open('''encrypt.out''' , '''w+''' ) as fout: # actual encrypt-process for line in fin: fout.write(self.encrypt_string(_lowerCamelCase , _lowerCamelCase ) ) except OSError: return False return True def lowercase ( self : Optional[Any] , _lowerCamelCase : str , _lowerCamelCase : int ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) try: with open(_lowerCamelCase ) as fin, open('''decrypt.out''' , '''w+''' ) as fout: # actual encrypt-process for line in fin: fout.write(self.decrypt_string(_lowerCamelCase , _lowerCamelCase ) ) except OSError: return False return True # Tests # crypt = XORCipher() # key = 67 # # test encrypt # print(crypt.encrypt("hallo welt",key)) # # test decrypt # print(crypt.decrypt(crypt.encrypt("hallo welt",key), key)) # # test encrypt_string # print(crypt.encrypt_string("hallo welt",key)) # # test decrypt_string # print(crypt.decrypt_string(crypt.encrypt_string("hallo welt",key),key)) # if (crypt.encrypt_file("test.txt",key)): # print("encrypt successful") # else: # print("encrypt unsuccessful") # if (crypt.decrypt_file("encrypt.out",key)): # print("decrypt successful") # else: # print("decrypt unsuccessful")
40
0
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ = logging.get_logger(__name__) UpperCAmelCase__ = { "microsoft/markuplm-base": "https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json", "microsoft/markuplm-large": "https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json", } class lowercase_ ( lowercase ): '''simple docstring''' __snake_case = '''markuplm''' def __init__( self : Union[str, Any] , __UpperCAmelCase : int=30_522 , __UpperCAmelCase : Optional[Any]=768 , __UpperCAmelCase : List[str]=12 , __UpperCAmelCase : Dict=12 , __UpperCAmelCase : Tuple=3_072 , __UpperCAmelCase : Optional[Any]="gelu" , __UpperCAmelCase : List[Any]=0.1 , __UpperCAmelCase : int=0.1 , __UpperCAmelCase : str=512 , __UpperCAmelCase : Union[str, Any]=2 , __UpperCAmelCase : Tuple=0.02 , __UpperCAmelCase : Dict=1e-1_2 , __UpperCAmelCase : int=0 , __UpperCAmelCase : List[str]=0 , __UpperCAmelCase : Optional[Any]=2 , __UpperCAmelCase : List[Any]=256 , __UpperCAmelCase : Dict=1_024 , __UpperCAmelCase : Optional[int]=216 , __UpperCAmelCase : str=1_001 , __UpperCAmelCase : int=32 , __UpperCAmelCase : Union[str, Any]=50 , __UpperCAmelCase : Any="absolute" , __UpperCAmelCase : Any=True , __UpperCAmelCase : List[str]=None , **__UpperCAmelCase : Union[str, Any] , ) ->str: """simple docstring""" super().__init__( pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase , ) a = vocab_size a = hidden_size a = num_hidden_layers a = num_attention_heads a = hidden_act a = intermediate_size a = hidden_dropout_prob a = attention_probs_dropout_prob a = max_position_embeddings a = type_vocab_size a = initializer_range a = layer_norm_eps a = position_embedding_type a = use_cache a = classifier_dropout # additional properties a = max_depth a = max_xpath_tag_unit_embeddings a = max_xpath_subs_unit_embeddings a = tag_pad_id a = subs_pad_id a = xpath_unit_hidden_size
0
from __future__ import annotations UpperCAmelCase__ = list[list[int]] # assigning initial values to the grid UpperCAmelCase__ = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution UpperCAmelCase__ = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def _a ( a :Matrix , a :int , a :int , a :int ) -> bool: for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def _a ( a :Matrix ) -> tuple[int, int] | None: for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def _a ( a :Matrix ) -> Matrix | None: if location := find_empty_location(a ): a , a = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(a , a , a , a ): a = digit if sudoku(a ) is not None: return grid a = 0 return None def _a ( a :Matrix ) -> None: for row in grid: for cell in row: print(a , end=''' ''' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print("\nExample grid:\n" + "=" * 20) print_solution(example_grid) print("\nExample grid solution:") UpperCAmelCase__ = sudoku(example_grid) if solution is not None: print_solution(solution) else: print("Cannot find a solution.")
0
1
"""simple docstring""" import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __UpperCAmelCase = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class _SCREAMING_SNAKE_CASE ( A__ , unittest.TestCase ): UpperCAmelCase_ :List[str] = XLMProphetNetTokenizer UpperCAmelCase_ :Dict = False UpperCAmelCase_ :List[str] = True def __lowerCAmelCase ( self ) -> Optional[Any]: super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase_ :Optional[int] = XLMProphetNetTokenizer(__A , keep_accents=__A ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self ) -> Union[str, Any]: lowerCAmelCase_ :Optional[Any] = """[PAD]""" lowerCAmelCase_ :Any = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__A ) , __A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__A ) , __A ) def __lowerCAmelCase ( self ) -> Optional[Any]: lowerCAmelCase_ :List[str] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """[PAD]""" ) self.assertEqual(vocab_keys[1] , """[CLS]""" ) self.assertEqual(vocab_keys[-1] , """j""" ) self.assertEqual(len(__A ) , 1012 ) def __lowerCAmelCase ( self ) -> str: self.assertEqual(self.get_tokenizer().vocab_size , 1012 ) def __lowerCAmelCase ( self ) -> Optional[int]: lowerCAmelCase_ :Optional[Any] = XLMProphetNetTokenizer(__A , keep_accents=__A ) lowerCAmelCase_ :Dict = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(__A , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__A ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) lowerCAmelCase_ :List[Any] = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( __A , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) lowerCAmelCase_ :Optional[int] = tokenizer.convert_tokens_to_ids(__A ) self.assertListEqual( __A , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) lowerCAmelCase_ :Any = tokenizer.convert_ids_to_tokens(__A ) self.assertListEqual( __A , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """[UNK]""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """[UNK]""", """.""", ] , ) @cached_property def __lowerCAmelCase ( self ) -> Any: return XLMProphetNetTokenizer.from_pretrained("""microsoft/xprophetnet-large-wiki100-cased""" ) @slow def __lowerCAmelCase ( self ) -> Optional[Any]: lowerCAmelCase_ :List[Any] = """Hello World!""" lowerCAmelCase_ :Optional[Any] = [3_5389, 6672, 49, 2] self.assertListEqual(__A , self.big_tokenizer.encode(__A ) ) @slow def __lowerCAmelCase ( self ) -> Any: # fmt: off lowerCAmelCase_ :Optional[int] = {"""input_ids""": [[1_1073, 8_2783, 18, 26, 8_2783, 549, 5_1540, 248, 1_7209, 1301, 217, 20, 21_5186, 1325, 147, 1_7209, 1301, 217, 20, 5_6370, 53, 12_2020, 20, 1_6477, 27, 8_7355, 4548, 20, 4728, 7_8392, 17, 15_9969, 18, 26, 2_4491, 629, 15, 538, 2_2704, 5439, 15, 2788, 2_4491, 9885, 15, 4_3534, 605, 15, 814, 1_8403, 3_3200, 29, 15, 4_3534, 2_4458, 1_2410, 111, 2_4966, 8_3669, 9637, 14_4068, 26, 850, 2_2346, 27, 147, 2_4966, 8_3669, 8_3490, 26, 3_9113, 735, 27, 689, 656, 2800, 1339, 4600, 53, 12_2020, 11_5785, 34, 816, 1339, 4_6887, 18, 147, 5_3905, 1951, 4_2238, 4_1170, 1_7732, 834, 436, 15, 2_7523, 9_8733, 217, 147, 5542, 4981, 930, 1_7347, 16, 2], [2_0091, 629, 94, 8_2786, 58, 490, 20, 1528, 84, 5_3905, 344, 8_0592, 11_0128, 1_8822, 5267, 1306, 62, 15_2537, 308, 7997, 401, 12_4427, 549, 3_5442, 225, 109, 1_5055, 2_5748, 147, 7119, 4_3712, 34, 767, 13_5366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 6_3784, 11_9466, 17, 14_7808, 8_8214, 18, 656, 81, 32, 3296, 1_0280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__A , model_name="""microsoft/xprophetnet-large-wiki100-cased""" , revision="""1acad1643ddd54a44df6a1b797ada8373685d90e""" , )
365
"""simple docstring""" # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class _SCREAMING_SNAKE_CASE ( A__ , A__ , A__ , unittest.TestCase ): UpperCAmelCase_ :int = StableDiffusionControlNetImgaImgPipeline UpperCAmelCase_ :str = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"height", "width"} UpperCAmelCase_ :Tuple = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCAmelCase_ :List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({"control_image"} ) UpperCAmelCase_ :Optional[Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def __lowerCAmelCase ( self ) -> List[str]: torch.manual_seed(0 ) lowerCAmelCase_ :Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) lowerCAmelCase_ :List[Any] = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[Any] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" , clip_sample=__A , set_alpha_to_one=__A , ) torch.manual_seed(0 ) lowerCAmelCase_ :List[str] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) lowerCAmelCase_ :Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) lowerCAmelCase_ :List[Any] = CLIPTextModel(__A ) lowerCAmelCase_ :int = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) lowerCAmelCase_ :Union[str, Any] = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def __lowerCAmelCase ( self , __A , __A=0 ) -> List[str]: if str(__A ).startswith("""mps""" ): lowerCAmelCase_ :Tuple = torch.manual_seed(__A ) else: lowerCAmelCase_ :Optional[int] = torch.Generator(device=__A ).manual_seed(__A ) lowerCAmelCase_ :List[Any] = 2 lowerCAmelCase_ :int = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__A , device=torch.device(__A ) , ) lowerCAmelCase_ :Optional[int] = floats_tensor(control_image.shape , rng=random.Random(__A ) ).to(__A ) lowerCAmelCase_ :Tuple = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase_ :List[Any] = Image.fromarray(np.uinta(__A ) ).convert("""RGB""" ).resize((64, 64) ) lowerCAmelCase_ :Union[str, Any] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def __lowerCAmelCase ( self ) -> int: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def __lowerCAmelCase ( self ) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def __lowerCAmelCase ( self ) -> List[str]: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) class _SCREAMING_SNAKE_CASE ( A__ , A__ , unittest.TestCase ): UpperCAmelCase_ :List[str] = StableDiffusionControlNetImgaImgPipeline UpperCAmelCase_ :int = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"height", "width"} UpperCAmelCase_ :str = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCAmelCase_ :int = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def __lowerCAmelCase ( self ) -> Optional[int]: torch.manual_seed(0 ) lowerCAmelCase_ :Dict = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(__A ): if isinstance(__A , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) lowerCAmelCase_ :List[Any] = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__A ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[Any] = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__A ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[Any] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" , clip_sample=__A , set_alpha_to_one=__A , ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) lowerCAmelCase_ :Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) lowerCAmelCase_ :str = CLIPTextModel(__A ) lowerCAmelCase_ :str = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) lowerCAmelCase_ :Optional[Any] = MultiControlNetModel([controlneta, controlneta] ) lowerCAmelCase_ :List[Any] = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def __lowerCAmelCase ( self , __A , __A=0 ) -> str: if str(__A ).startswith("""mps""" ): lowerCAmelCase_ :Optional[Any] = torch.manual_seed(__A ) else: lowerCAmelCase_ :List[Any] = torch.Generator(device=__A ).manual_seed(__A ) lowerCAmelCase_ :Optional[Any] = 2 lowerCAmelCase_ :Optional[int] = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__A , device=torch.device(__A ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__A , device=torch.device(__A ) , ), ] lowerCAmelCase_ :int = floats_tensor(control_image[0].shape , rng=random.Random(__A ) ).to(__A ) lowerCAmelCase_ :Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase_ :List[Any] = Image.fromarray(np.uinta(__A ) ).convert("""RGB""" ).resize((64, 64) ) lowerCAmelCase_ :List[str] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def __lowerCAmelCase ( self ) -> Optional[Any]: lowerCAmelCase_ :List[str] = self.get_dummy_components() lowerCAmelCase_ :Tuple = self.pipeline_class(**__A ) pipe.to(__A ) lowerCAmelCase_ :Union[str, Any] = 1_0.0 lowerCAmelCase_ :Union[str, Any] = 4 lowerCAmelCase_ :Tuple = self.get_dummy_inputs(__A ) lowerCAmelCase_ :List[str] = steps lowerCAmelCase_ :int = scale lowerCAmelCase_ :Union[str, Any] = pipe(**__A )[0] lowerCAmelCase_ :Any = self.get_dummy_inputs(__A ) lowerCAmelCase_ :str = steps lowerCAmelCase_ :str = scale lowerCAmelCase_ :Tuple = pipe(**__A , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] lowerCAmelCase_ :Optional[Any] = self.get_dummy_inputs(__A ) lowerCAmelCase_ :Union[str, Any] = steps lowerCAmelCase_ :Union[str, Any] = scale lowerCAmelCase_ :str = pipe(**__A , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] lowerCAmelCase_ :List[str] = self.get_dummy_inputs(__A ) lowerCAmelCase_ :Optional[int] = steps lowerCAmelCase_ :Tuple = scale lowerCAmelCase_ :str = pipe(**__A , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 def __lowerCAmelCase ( self ) -> Dict: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def __lowerCAmelCase ( self ) -> Tuple: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def __lowerCAmelCase ( self ) -> Optional[int]: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) def __lowerCAmelCase ( self ) -> List[str]: lowerCAmelCase_ :str = self.get_dummy_components() lowerCAmelCase_ :Tuple = self.pipeline_class(**__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(__A ) except NotImplementedError: pass @slow @require_torch_gpu class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self ) -> str: lowerCAmelCase_ :Any = ControlNetModel.from_pretrained("""lllyasviel/sd-controlnet-canny""" ) lowerCAmelCase_ :int = StableDiffusionControlNetImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , safety_checker=__A , controlnet=__A ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__A ) lowerCAmelCase_ :List[str] = torch.Generator(device="""cpu""" ).manual_seed(0 ) lowerCAmelCase_ :List[Any] = """evil space-punk bird""" lowerCAmelCase_ :List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png""" ).resize((512, 512) ) lowerCAmelCase_ :int = load_image( """https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png""" ).resize((512, 512) ) lowerCAmelCase_ :Union[str, Any] = pipe( __A , __A , control_image=__A , generator=__A , output_type="""np""" , num_inference_steps=50 , strength=0.6 , ) lowerCAmelCase_ :Tuple = output.images[0] assert image.shape == (512, 512, 3) lowerCAmelCase_ :Tuple = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy""" ) assert np.abs(expected_image - image ).max() < 9E-2
1
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor a_ : List[Any] = logging.get_logger(__name__) class a ( _lowerCAmelCase ): def __init__( self , *__magic_name__ , **__magic_name__ ) -> Dict: warnings.warn( 'The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use FlavaImageProcessor instead.' , __magic_name__ , ) super().__init__(*__magic_name__ , **__magic_name__ )
168
'''simple docstring''' lowerCAmelCase : Optional[int] =256 # Modulus to hash a string lowerCAmelCase : Tuple =1_000_003 def UpperCAmelCase_ ( __lowerCamelCase : str ,__lowerCamelCase : str ): lowercase_ :Dict = len(__lowerCamelCase ) lowercase_ :int = len(__lowerCamelCase ) if p_len > t_len: return False lowercase_ :Any = 0 lowercase_ :Dict = 0 lowercase_ :int = 1 # Calculating the hash of pattern and substring of text for i in range(__lowerCamelCase ): lowercase_ :Union[str, Any] = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus lowercase_ :Tuple = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue lowercase_ :Tuple = (modulus_power * alphabet_size) % modulus for i in range(0 ,t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash lowercase_ :Tuple = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def UpperCAmelCase_ ( ): lowercase_ :List[Any] = "abc1abc12" lowercase_ :List[str] = "alskfjaldsabc1abc1abc12k23adsfabcabc" lowercase_ :int = "alskfjaldsk23adsfabcabc" assert rabin_karp(__lowerCamelCase ,__lowerCamelCase ) and not rabin_karp(__lowerCamelCase ,__lowerCamelCase ) # Test 2) lowercase_ :Union[str, Any] = "ABABX" lowercase_ :Optional[Any] = "ABABZABABYABABX" assert rabin_karp(__lowerCamelCase ,__lowerCamelCase ) # Test 3) lowercase_ :Dict = "AAAB" lowercase_ :int = "ABAAAAAB" assert rabin_karp(__lowerCamelCase ,__lowerCamelCase ) # Test 4) lowercase_ :Tuple = "abcdabcy" lowercase_ :Union[str, Any] = "abcxabcdabxabcdabcdabcy" assert rabin_karp(__lowerCamelCase ,__lowerCamelCase ) # Test 5) lowercase_ :Tuple = "Lü" lowercase_ :List[Any] = "Lüsai" assert rabin_karp(__lowerCamelCase ,__lowerCamelCase ) lowercase_ :str = "Lue" assert not rabin_karp(__lowerCamelCase ,__lowerCamelCase ) print("Success." ) if __name__ == "__main__": test_rabin_karp()
223
0
'''simple docstring''' # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch class UpperCamelCase ( TensorFormatter[Mapping, "torch.Tensor", Mapping] ): """simple docstring""" def __init__( self : int , UpperCAmelCase_ : Optional[int]=None , **UpperCAmelCase_ : Dict): """simple docstring""" super().__init__(features=UpperCAmelCase_) a : Optional[Any] = torch_tensor_kwargs import torch # noqa import torch at initialization def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] , UpperCAmelCase_ : List[Any]): """simple docstring""" import torch if isinstance(UpperCAmelCase_ , UpperCAmelCase_) and column: if all( isinstance(UpperCAmelCase_ , torch.Tensor) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column): return torch.stack(UpperCAmelCase_) return column def SCREAMING_SNAKE_CASE_ ( self : Optional[int] , UpperCAmelCase_ : Tuple): """simple docstring""" import torch if isinstance(UpperCAmelCase_ , (str, bytes, type(UpperCAmelCase_))): return value elif isinstance(UpperCAmelCase_ , (np.character, np.ndarray)) and np.issubdtype(value.dtype , np.character): return value.tolist() a : Optional[Any] = {} if isinstance(UpperCAmelCase_ , (np.number, np.ndarray)) and np.issubdtype(value.dtype , np.integer): a : List[Any] = {'dtype': torch.intaa} elif isinstance(UpperCAmelCase_ , (np.number, np.ndarray)) and np.issubdtype(value.dtype , np.floating): a : List[str] = {'dtype': torch.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(UpperCAmelCase_ , PIL.Image.Image): a : Dict = np.asarray(UpperCAmelCase_) return torch.tensor(UpperCAmelCase_ , **{**default_dtype, **self.torch_tensor_kwargs}) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] , UpperCAmelCase_ : Dict): """simple docstring""" import torch # support for torch, tf, jax etc. if hasattr(UpperCAmelCase_ , '__array__') and not isinstance(UpperCAmelCase_ , torch.Tensor): a : Optional[Any] = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(UpperCAmelCase_ , np.ndarray): if data_struct.dtype == object: # torch tensors cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(UpperCAmelCase_) for substruct in data_struct]) elif isinstance(UpperCAmelCase_ , (list, tuple)): return self._consolidate([self.recursive_tensorize(UpperCAmelCase_) for substruct in data_struct]) return self._tensorize(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : int , UpperCAmelCase_ : dict): """simple docstring""" return map_nested(self._recursive_tensorize , UpperCAmelCase_ , map_list=UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] , UpperCAmelCase_ : pa.Table): """simple docstring""" a : List[Any] = self.numpy_arrow_extractor().extract_row(UpperCAmelCase_) a : Tuple = self.python_features_decoder.decode_row(UpperCAmelCase_) return self.recursive_tensorize(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] , UpperCAmelCase_ : pa.Table): """simple docstring""" a : int = self.numpy_arrow_extractor().extract_column(UpperCAmelCase_) a : Union[str, Any] = self.python_features_decoder.decode_column(UpperCAmelCase_ , pa_table.column_names[0]) a : str = self.recursive_tensorize(UpperCAmelCase_) a : str = self._consolidate(UpperCAmelCase_) return column def SCREAMING_SNAKE_CASE_ ( self : int , UpperCAmelCase_ : pa.Table): """simple docstring""" a : int = self.numpy_arrow_extractor().extract_batch(UpperCAmelCase_) a : List[str] = self.python_features_decoder.decode_batch(UpperCAmelCase_) a : Tuple = self.recursive_tensorize(UpperCAmelCase_) for column_name in batch: a : Any = self._consolidate(batch[column_name]) return batch
345
'''simple docstring''' import torch def SCREAMING_SNAKE_CASE__ ( ) -> str: """simple docstring""" if torch.cuda.is_available(): a : int = torch.cuda.device_count() else: a : Any = 0 print(F"""Successfully ran on {num_gpus} GPUs""" ) if __name__ == "__main__": main()
345
1
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> list: if len(SCREAMING_SNAKE_CASE_ ) <= 1: return [tuple(SCREAMING_SNAKE_CASE_ )] lowerCAmelCase__ : Optional[Any] = [] def generate(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , SCREAMING_SNAKE_CASE_ ) for i in range(k - 1 ): if k % 2 == 0: # k is even lowerCAmelCase__ , lowerCAmelCase__ : str = arr[k - 1], arr[i] else: # k is odd lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = arr[k - 1], arr[0] generate(k - 1 , SCREAMING_SNAKE_CASE_ ) generate(len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) return res if __name__ == "__main__": lowerCamelCase__ = input("""Enter numbers separated by a comma:\n""").strip() lowerCamelCase__ = [int(item) for item in user_input.split(""",""")] print(heaps(arr))
212
from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class A__ ( __magic_name__ ): lowercase = 42 lowercase = 42 def __init__( self : Any , a : UNetaDModel , a : ScoreSdeVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=a , scheduler=a ) @torch.no_grad() def __call__( self : List[str] , a : int = 1 , a : int = 2_000 , a : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a : Optional[str] = "pil" , a : bool = True , **a : int , ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.unet.config.sample_size lowerCAmelCase__ : Union[str, Any] = (batch_size, 3, img_size, img_size) lowerCAmelCase__ : Tuple = self.unet lowerCAmelCase__ : Optional[Any] = randn_tensor(a , generator=a ) * self.scheduler.init_noise_sigma lowerCAmelCase__ : int = sample.to(self.device ) self.scheduler.set_timesteps(a ) self.scheduler.set_sigmas(a ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): lowerCAmelCase__ : Dict = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): lowerCAmelCase__ : Optional[Any] = self.unet(a , a ).sample lowerCAmelCase__ : Dict = self.scheduler.step_correct(a , a , generator=a ).prev_sample # prediction step lowerCAmelCase__ : Optional[int] = model(a , a ).sample lowerCAmelCase__ : Optional[Any] = self.scheduler.step_pred(a , a , a , generator=a ) lowerCAmelCase__ , lowerCAmelCase__ : str = output.prev_sample, output.prev_sample_mean lowerCAmelCase__ : Any = sample_mean.clamp(0 , 1 ) lowerCAmelCase__ : List[str] = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase__ : int = self.numpy_to_pil(a ) if not return_dict: return (sample,) return ImagePipelineOutput(images=a )
212
1
from __future__ import annotations from collections.abc import Sequence from typing import Literal def A_ ( _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Union[str, Any] = list(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Union[str, Any] = list(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Any = 0 for i in range(len(_UpperCAmelCase ) ): if lista[i] != lista[i]: count += 1 SCREAMING_SNAKE_CASE_: List[Any] = "_" if count > 1: return False else: return "".join(_UpperCAmelCase ) def A_ ( _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[str] = [] while True: SCREAMING_SNAKE_CASE_: Optional[int] = ["$"] * len(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] = [] for i in range(len(_UpperCAmelCase ) ): for j in range(i + 1 , len(_UpperCAmelCase ) ): SCREAMING_SNAKE_CASE_: Any = compare_string(binary[i] , binary[j] ) if k is False: SCREAMING_SNAKE_CASE_: str = "*" SCREAMING_SNAKE_CASE_: Dict = "*" temp.append("X" ) for i in range(len(_UpperCAmelCase ) ): if checka[i] == "$": pi.append(binary[i] ) if len(_UpperCAmelCase ) == 0: return pi SCREAMING_SNAKE_CASE_: Optional[Any] = list(set(_UpperCAmelCase ) ) def A_ ( _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: int = [] for minterm in minterms: SCREAMING_SNAKE_CASE_: Any = "" for _ in range(_UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Optional[Any] = str(minterm % 2 ) + string minterm //= 2 temp.append(_UpperCAmelCase ) return temp def A_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Optional[int] = list(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Dict = list(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Dict = 0 for i in range(len(_UpperCAmelCase ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def A_ ( _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: int = [] SCREAMING_SNAKE_CASE_: str = [0] * len(_UpperCAmelCase ) for i in range(len(chart[0] ) ): SCREAMING_SNAKE_CASE_: int = 0 SCREAMING_SNAKE_CASE_: Union[str, Any] = -1 for j in range(len(_UpperCAmelCase ) ): if chart[j][i] == 1: count += 1 SCREAMING_SNAKE_CASE_: List[Any] = j if count == 1: SCREAMING_SNAKE_CASE_: Dict = 1 for i in range(len(_UpperCAmelCase ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(_UpperCAmelCase ) ): SCREAMING_SNAKE_CASE_: Tuple = 0 temp.append(prime_implicants[i] ) while True: SCREAMING_SNAKE_CASE_: List[Any] = 0 SCREAMING_SNAKE_CASE_: Dict = -1 SCREAMING_SNAKE_CASE_: Tuple = 0 for i in range(len(_UpperCAmelCase ) ): SCREAMING_SNAKE_CASE_: List[str] = chart[i].count(1 ) if count_n > max_n: SCREAMING_SNAKE_CASE_: str = count_n SCREAMING_SNAKE_CASE_: List[Any] = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(_UpperCAmelCase ) ): SCREAMING_SNAKE_CASE_: Tuple = 0 def A_ ( _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: int = [[0 for x in range(len(_UpperCAmelCase ) )] for x in range(len(_UpperCAmelCase ) )] for i in range(len(_UpperCAmelCase ) ): SCREAMING_SNAKE_CASE_: Optional[int] = prime_implicants[i].count("_" ) for j in range(len(_UpperCAmelCase ) ): if is_for_table(prime_implicants[i] , binary[j] , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Tuple = 1 return chart def A_ ( ): SCREAMING_SNAKE_CASE_: Optional[int] = int(input("Enter the no. of variables\n" ) ) SCREAMING_SNAKE_CASE_: Optional[int] = [ float(_UpperCAmelCase ) for x in input( "Enter the decimal representation of Minterms 'Spaces Separated'\n" ).split() ] SCREAMING_SNAKE_CASE_: List[Any] = decimal_to_binary(_UpperCAmelCase , _UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Any = check(_UpperCAmelCase ) print("Prime Implicants are:" ) print(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] = prime_implicant_chart(_UpperCAmelCase , _UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Tuple = selection(_UpperCAmelCase , _UpperCAmelCase ) print("Essential Prime Implicants are:" ) print(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
127
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : Optional[int] = logging.get_logger(__name__) lowerCAmelCase : int = { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/config.json""", """google/fnet-large""": """https://huggingface.co/google/fnet-large/resolve/main/config.json""" # See all FNet models at https://huggingface.co/models?filter=fnet } class __lowercase ( UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : int = '''fnet''' def __init__( self : Optional[Any] , lowerCAmelCase__ : List[str]=3_2000 , lowerCAmelCase__ : Union[str, Any]=768 , lowerCAmelCase__ : Optional[Any]=12 , lowerCAmelCase__ : Tuple=3072 , lowerCAmelCase__ : Any="gelu_new" , lowerCAmelCase__ : Union[str, Any]=0.1 , lowerCAmelCase__ : Optional[Any]=512 , lowerCAmelCase__ : int=4 , lowerCAmelCase__ : List[Any]=0.02 , lowerCAmelCase__ : Union[str, Any]=1E-12 , lowerCAmelCase__ : Optional[int]=False , lowerCAmelCase__ : Any=512 , lowerCAmelCase__ : Optional[int]=3 , lowerCAmelCase__ : int=1 , lowerCAmelCase__ : List[Any]=2 , **lowerCAmelCase__ : Optional[int] , ): super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = vocab_size SCREAMING_SNAKE_CASE_: Optional[Any] = max_position_embeddings SCREAMING_SNAKE_CASE_: Tuple = hidden_size SCREAMING_SNAKE_CASE_: Union[str, Any] = num_hidden_layers SCREAMING_SNAKE_CASE_: Optional[int] = intermediate_size SCREAMING_SNAKE_CASE_: Union[str, Any] = hidden_act SCREAMING_SNAKE_CASE_: Any = hidden_dropout_prob SCREAMING_SNAKE_CASE_: int = initializer_range SCREAMING_SNAKE_CASE_: Optional[int] = type_vocab_size SCREAMING_SNAKE_CASE_: int = layer_norm_eps SCREAMING_SNAKE_CASE_: Union[str, Any] = use_tpu_fourier_optimizations SCREAMING_SNAKE_CASE_: int = tpu_short_seq_length
127
1
"""simple docstring""" import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope="session" ) def a__ ( ): """simple docstring""" UpperCamelCase = 10 UpperCamelCase = datasets.Features( { "tokens": datasets.Sequence(datasets.Value("string" ) ), "labels": datasets.Sequence(datasets.ClassLabel(names=["negative", "positive"] ) ), "answers": datasets.Sequence( { "text": datasets.Value("string" ), "answer_start": datasets.Value("int32" ), } ), "id": datasets.Value("int64" ), } ) UpperCamelCase = datasets.Dataset.from_dict( { "tokens": [["foo"] * 5] * n, "labels": [[1] * 5] * n, "answers": [{"answer_start": [97], "text": ["1976"]}] * 10, "id": list(range(SCREAMING_SNAKE_CASE_ ) ), } , features=SCREAMING_SNAKE_CASE_ , ) return dataset @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "file.arrow" ) dataset.map(cache_file_name=SCREAMING_SNAKE_CASE_ ) return filename # FILE_CONTENT + files lowerCAmelCase__ = '''\ Text data. Second line of data.''' @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "file.txt" UpperCamelCase = FILE_CONTENT with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) return filename @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" import bza UpperCamelCase = tmp_path_factory.mktemp("data" ) / "file.txt.bz2" UpperCamelCase = bytes(SCREAMING_SNAKE_CASE_ , "utf-8" ) with bza.open(SCREAMING_SNAKE_CASE_ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" import gzip UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "file.txt.gz" ) UpperCamelCase = bytes(SCREAMING_SNAKE_CASE_ , "utf-8" ) with gzip.open(SCREAMING_SNAKE_CASE_ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" if datasets.config.LZ4_AVAILABLE: import lza.frame UpperCamelCase = tmp_path_factory.mktemp("data" ) / "file.txt.lz4" UpperCamelCase = bytes(SCREAMING_SNAKE_CASE_ , "utf-8" ) with lza.frame.open(SCREAMING_SNAKE_CASE_ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" if datasets.config.PY7ZR_AVAILABLE: import pyazr UpperCamelCase = tmp_path_factory.mktemp("data" ) / "file.txt.7z" with pyazr.SevenZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as archive: archive.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" import tarfile UpperCamelCase = tmp_path_factory.mktemp("data" ) / "file.txt.tar" with tarfile.TarFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.add(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" import lzma UpperCamelCase = tmp_path_factory.mktemp("data" ) / "file.txt.xz" UpperCamelCase = bytes(SCREAMING_SNAKE_CASE_ , "utf-8" ) with lzma.open(SCREAMING_SNAKE_CASE_ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" import zipfile UpperCamelCase = tmp_path_factory.mktemp("data" ) / "file.txt.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd UpperCamelCase = tmp_path_factory.mktemp("data" ) / "file.txt.zst" UpperCamelCase = bytes(SCREAMING_SNAKE_CASE_ , "utf-8" ) with zstd.open(SCREAMING_SNAKE_CASE_ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "file.xml" UpperCamelCase = textwrap.dedent( "\\n <?xml version=\"1.0\" encoding=\"UTF-8\" ?>\n <tmx version=\"1.4\">\n <header segtype=\"sentence\" srclang=\"ca\" />\n <body>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang=\"ca\"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang=\"en\"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>" ) with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) return filename lowerCAmelCase__ = [ {'''col_1''': '''0''', '''col_2''': 0, '''col_3''': 0.0}, {'''col_1''': '''1''', '''col_2''': 1, '''col_3''': 1.0}, {'''col_1''': '''2''', '''col_2''': 2, '''col_3''': 2.0}, {'''col_1''': '''3''', '''col_2''': 3, '''col_3''': 3.0}, ] lowerCAmelCase__ = [ {'''col_1''': '''4''', '''col_2''': 4, '''col_3''': 4.0}, {'''col_1''': '''5''', '''col_2''': 5, '''col_3''': 5.0}, ] lowerCAmelCase__ = { '''col_1''': ['''0''', '''1''', '''2''', '''3'''], '''col_2''': [0, 1, 2, 3], '''col_3''': [0.0, 1.0, 2.0, 3.0], } lowerCAmelCase__ = [ {'''col_3''': 0.0, '''col_1''': '''0''', '''col_2''': 0}, {'''col_3''': 1.0, '''col_1''': '''1''', '''col_2''': 1}, ] lowerCAmelCase__ = [ {'''col_1''': '''s0''', '''col_2''': 0, '''col_3''': 0.0}, {'''col_1''': '''s1''', '''col_2''': 1, '''col_3''': 1.0}, {'''col_1''': '''s2''', '''col_2''': 2, '''col_3''': 2.0}, {'''col_1''': '''s3''', '''col_2''': 3, '''col_3''': 3.0}, ] @pytest.fixture(scope="session" ) def a__ ( ): """simple docstring""" return DATA_DICT_OF_LISTS @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = datasets.Dataset.from_dict(SCREAMING_SNAKE_CASE_ ) UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.arrow" ) dataset.map(cache_file_name=SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.sqlite" ) with contextlib.closing(sqlitea.connect(SCREAMING_SNAKE_CASE_ ) ) as con: UpperCamelCase = con.cursor() cur.execute("CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)" ) for item in DATA: cur.execute("INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)" , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.csv" ) with open(SCREAMING_SNAKE_CASE_ , "w" , newline="" ) as f: UpperCamelCase = csv.DictWriter(SCREAMING_SNAKE_CASE_ , fieldnames=["col_1", "col_2", "col_3"] ) writer.writeheader() for item in DATA: writer.writerow(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset2.csv" ) with open(SCREAMING_SNAKE_CASE_ , "w" , newline="" ) as f: UpperCamelCase = csv.DictWriter(SCREAMING_SNAKE_CASE_ , fieldnames=["col_1", "col_2", "col_3"] ) writer.writeheader() for item in DATA: writer.writerow(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" import bza UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset.csv.bz2" with open(SCREAMING_SNAKE_CASE_ , "rb" ) as f: UpperCamelCase = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(SCREAMING_SNAKE_CASE_ , "wb" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset.csv.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset.csv.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(csv_path.replace(".csv" , ".CSV" ) ) ) f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(csva_path.replace(".csv" , ".CSV" ) ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset_with_dir.csv.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE_ ) ) ) f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE_ ) ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.parquet" ) UpperCamelCase = pa.schema( { "col_1": pa.string(), "col_2": pa.intaa(), "col_3": pa.floataa(), } ) with open(SCREAMING_SNAKE_CASE_ , "wb" ) as f: UpperCamelCase = pq.ParquetWriter(SCREAMING_SNAKE_CASE_ , schema=SCREAMING_SNAKE_CASE_ ) UpperCamelCase = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(SCREAMING_SNAKE_CASE_ ) )] for k in DATA[0]} , schema=SCREAMING_SNAKE_CASE_ ) writer.write_table(SCREAMING_SNAKE_CASE_ ) writer.close() return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.json" ) UpperCamelCase = {"data": DATA} with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.json" ) UpperCamelCase = {"data": DATA_DICT_OF_LISTS} with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.jsonl" ) with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: for item in DATA: f.write(json.dumps(SCREAMING_SNAKE_CASE_ ) + "\n" ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset2.jsonl" ) with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: for item in DATA: f.write(json.dumps(SCREAMING_SNAKE_CASE_ ) + "\n" ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset_312.jsonl" ) with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: for item in DATA_312: f.write(json.dumps(SCREAMING_SNAKE_CASE_ ) + "\n" ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset-str.jsonl" ) with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: for item in DATA_STR: f.write(json.dumps(SCREAMING_SNAKE_CASE_ ) + "\n" ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" import gzip UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.txt.gz" ) with open(SCREAMING_SNAKE_CASE_ , "rb" ) as orig_file: with gzip.open(SCREAMING_SNAKE_CASE_ , "wb" ) as zipped_file: zipped_file.writelines(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" import gzip UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.jsonl.gz" ) with open(SCREAMING_SNAKE_CASE_ , "rb" ) as orig_file: with gzip.open(SCREAMING_SNAKE_CASE_ , "wb" ) as zipped_file: zipped_file.writelines(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset.jsonl.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset_nested.jsonl.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.join("nested" , os.path.basename(SCREAMING_SNAKE_CASE_ ) ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset_with_dir.jsonl.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE_ ) ) ) f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE_ ) ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset.jsonl.tar" with tarfile.TarFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.add(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) f.add(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset_nested.jsonl.tar" with tarfile.TarFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.add(SCREAMING_SNAKE_CASE_ , arcname=os.path.join("nested" , os.path.basename(SCREAMING_SNAKE_CASE_ ) ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = ["0", "1", "2", "3"] UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset.txt" ) with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = ["0", "1", "2", "3"] UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset2.txt" ) with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = ["0", "1", "2", "3"] UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset.abc" with open(SCREAMING_SNAKE_CASE_ , "w" ) as f: for item in data: f.write(item + "\n" ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset.text.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset_with_dir.text.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE_ ) ) ) f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.join("main_dir" , os.path.basename(SCREAMING_SNAKE_CASE_ ) ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset.ext.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename("unsupported.ext" ) ) f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename("unsupported_2.ext" ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = "\n".join(["First", "Second\u2029with Unicode new line", "Third"] ) UpperCamelCase = str(tmp_path_factory.mktemp("data" ) / "dataset_with_unicode_new_lines.txt" ) with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) return path @pytest.fixture(scope="session" ) def a__ ( ): """simple docstring""" return os.path.join("tests" , "features" , "data" , "test_image_rgb.jpg" ) @pytest.fixture(scope="session" ) def a__ ( ): """simple docstring""" return os.path.join("tests" , "features" , "data" , "test_audio_44100.wav" ) @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data" ) / "dataset.img.zip" with zipfile.ZipFile(SCREAMING_SNAKE_CASE_ , "w" ) as f: f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ) ) f.write(SCREAMING_SNAKE_CASE_ , arcname=os.path.basename(SCREAMING_SNAKE_CASE_ ).replace(".jpg" , "2.jpg" ) ) return path @pytest.fixture(scope="session" ) def a__ ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCamelCase = tmp_path_factory.mktemp("data_dir" ) (data_dir / "subdir").mkdir() with open(data_dir / "subdir" / "train.txt" , "w" ) as f: f.write("foo\n" * 10 ) with open(data_dir / "subdir" / "test.txt" , "w" ) as f: f.write("bar\n" * 10 ) # hidden file with open(data_dir / "subdir" / ".test.txt" , "w" ) as f: f.write("bar\n" * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / ".subdir" / "train.txt" , "w" ) as f: f.write("foo\n" * 10 ) with open(data_dir / ".subdir" / "test.txt" , "w" ) as f: f.write("bar\n" * 10 ) return data_dir
153
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf UpperCamelCase__ = logging.get_logger(__name__) @dataclass class a__ ( snake_case__ ): _a : List[str] = [ """no_inference""", """no_cuda""", """no_tpu""", """no_speed""", """no_memory""", """no_env_print""", """no_multi_process""", ] def __init__( self , **_A ): """simple docstring""" for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __lowerCAmelCase = deprecated_arg[3:] __lowerCAmelCase = not kwargs.pop(_A ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no-{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) __lowerCAmelCase = kwargs.pop("tpu_name" , self.tpu_name ) __lowerCAmelCase = kwargs.pop("device_idx" , self.device_idx ) __lowerCAmelCase = kwargs.pop("eager_mode" , self.eager_mode ) __lowerCAmelCase = kwargs.pop("use_xla" , self.use_xla ) super().__init__(**_A ) _a : str = field( default=snake_case__ , metadata={"""help""": """Name of TPU"""} , ) _a : int = field( default=0 , metadata={"""help""": """CPU / GPU device index. Defaults to 0."""} , ) _a : bool = field(default=snake_case__ , metadata={"""help""": """Benchmark models in eager model."""} ) _a : bool = field( default=snake_case__ , metadata={ """help""": """Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`.""" } , ) @cached_property def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" requires_backends(self , ["tf"] ) __lowerCAmelCase = None if self.tpu: try: if self.tpu_name: __lowerCAmelCase = tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: __lowerCAmelCase = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: __lowerCAmelCase = None return tpu @cached_property def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" requires_backends(self , ["tf"] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) __lowerCAmelCase = tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , "GPU" ) __lowerCAmelCase = tf.distribute.OneDeviceStrategy(device=f"""/gpu:{self.device_idx}""" ) else: tf.config.set_visible_devices([] , "GPU" ) # disable GPU __lowerCAmelCase = tf.distribute.OneDeviceStrategy(device=f"""/cpu:{self.device_idx}""" ) return strategy @property def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" requires_backends(self , ["tf"] ) return self._setup_tpu is not None @property def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" requires_backends(self , ["tf"] ) return self._setup_strategy @property def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" requires_backends(self , ["tf"] ) return tf.config.list_physical_devices("GPU" ) @property def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" requires_backends(self , ["tf"] ) if self.cuda: return len(self.gpu_list ) return 0 @property def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" return self.n_gpu > 0
92
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer a__ : List[Any] = logging.get_logger(__name__) a__ : str = {'vocab_file': 'vocab.txt'} a__ : Any = { 'vocab_file': { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt', } } a__ : Tuple = { 'YituTech/conv-bert-base': 5_1_2, 'YituTech/conv-bert-medium-small': 5_1_2, 'YituTech/conv-bert-small': 5_1_2, } a__ : str = { 'YituTech/conv-bert-base': {'do_lower_case': True}, 'YituTech/conv-bert-medium-small': {'do_lower_case': True}, 'YituTech/conv-bert-small': {'do_lower_case': True}, } class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_INIT_CONFIGURATION __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE = ConvBertTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=True , lowercase="[UNK]" , lowercase="[SEP]" , lowercase="[PAD]" , lowercase="[CLS]" , lowercase="[MASK]" , lowercase=True , lowercase=None , **lowercase , ) -> int: super().__init__( lowercase , tokenizer_file=lowercase , do_lower_case=lowercase , unk_token=lowercase , sep_token=lowercase , pad_token=lowercase , cls_token=lowercase , mask_token=lowercase , tokenize_chinese_chars=lowercase , strip_accents=lowercase , **lowercase , ) __UpperCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , lowercase ) != do_lower_case or normalizer_state.get("""strip_accents""" , lowercase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , lowercase ) != tokenize_chinese_chars ): __UpperCamelCase = getattr(lowercase , normalizer_state.pop("""type""" ) ) __UpperCamelCase = do_lower_case __UpperCamelCase = strip_accents __UpperCamelCase = tokenize_chinese_chars __UpperCamelCase = normalizer_class(**lowercase ) __UpperCamelCase = do_lower_case def __lowerCamelCase ( self , lowercase , lowercase=None ) -> Tuple: __UpperCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowerCamelCase ( self , lowercase , lowercase = None ) -> List[int]: __UpperCamelCase = [self.sep_token_id] __UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowerCamelCase ( self , lowercase , lowercase = None ) -> Tuple[str]: __UpperCamelCase = self._tokenizer.model.save(lowercase , name=lowercase ) return tuple(lowercase )
367
'''simple docstring''' import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex a__ : Optional[Any] = logging.getLogger(__name__) class UpperCAmelCase__ : def __init__( self ) -> Union[str, Any]: __UpperCamelCase = False def __lowerCamelCase ( self , lowercase , lowercase , lowercase , lowercase ) -> Optional[int]: if not self.initialized: __UpperCamelCase = RagRetriever( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) __UpperCamelCase = True def __lowerCamelCase ( self ) -> List[Any]: self.retriever.index.init_index() def __lowerCamelCase ( self , lowercase , lowercase ) -> Optional[Any]: __UpperCamelCase , __UpperCamelCase = self.retriever._main_retrieve(lowercase , lowercase ) return doc_ids, retrieved_doc_embeds class UpperCAmelCase__ ( UpperCAmelCase_): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase=None ) -> Optional[Any]: if index is not None and index.is_initialized() and len(lowercase ) > 0: raise ValueError( """When using Ray for distributed fine-tuning, """ """you'll need to provide the paths instead, """ """as the dataset and the index are loaded """ """separately. More info in examples/rag/use_own_knowledge_dataset.py """ ) super().__init__( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) __UpperCamelCase = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(lowercase , lowercase , lowercase , lowercase ) for worker in self.retrieval_workers ] ) def __lowerCamelCase ( self ) -> Optional[int]: logger.info("""initializing retrieval""" ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def __lowerCamelCase ( self , lowercase , lowercase ) -> List[str]: if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. __UpperCamelCase = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers ) - 1 )] __UpperCamelCase , __UpperCamelCase = ray.get(random_worker.retrieve.remote(lowercase , lowercase ) ) else: __UpperCamelCase , __UpperCamelCase = self._main_retrieve(lowercase , lowercase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(lowercase ) @classmethod def __lowerCamelCase ( cls , lowercase , lowercase=None , **lowercase ) -> Tuple: return super(lowercase , cls ).get_tokenizers(lowercase , lowercase , **lowercase ) @classmethod def __lowerCamelCase ( cls , lowercase , lowercase , lowercase=None , **lowercase ) -> Dict: __UpperCamelCase = kwargs.pop("""config""" , lowercase ) or RagConfig.from_pretrained(lowercase , **lowercase ) __UpperCamelCase = RagTokenizer.from_pretrained(lowercase , config=lowercase ) __UpperCamelCase = rag_tokenizer.question_encoder __UpperCamelCase = rag_tokenizer.generator if indexed_dataset is not None: __UpperCamelCase = """custom""" __UpperCamelCase = CustomHFIndex(config.retrieval_vector_size , lowercase ) else: __UpperCamelCase = cls._build_index(lowercase ) return cls( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , retrieval_workers=lowercase , index=lowercase , )
243
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import _LazyModule snake_case_ : Tuple = {'tokenization_byt5': ['ByT5Tokenizer']} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys snake_case_ : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
'''simple docstring''' from math import pi def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
83
1
'''simple docstring''' from __future__ import annotations from typing import Any class lowerCamelCase : def __init__( self, lowercase_, lowercase_, lowercase_ = 0 ) -> None: snake_case , snake_case = row, column snake_case = [[default_value for c in range(lowercase_ )] for r in range(lowercase_ )] def __str__( self ) -> str: snake_case = F'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier snake_case = 0 for row_vector in self.array: for obj in row_vector: snake_case = max(lowercase_, len(str(lowercase_ ) ) ) snake_case = F'''%{max_element_length}s''' # Make string and return def single_line(lowercase_ ) -> str: nonlocal string_format_identifier snake_case = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase_ ) for row_vector in self.array ) return s def __repr__( self ) -> str: return str(self ) def _lowerCamelCase ( self, lowercase_ ) -> bool: if not (isinstance(lowercase_, (list, tuple) ) and len(lowercase_ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self, lowercase_ ) -> Any: assert self.validate_indicies(lowercase_ ) return self.array[loc[0]][loc[1]] def __setitem__( self, lowercase_, lowercase_ ) -> None: assert self.validate_indicies(lowercase_ ) snake_case = value def __add__( self, lowercase_ ) -> Matrix: assert isinstance(lowercase_, lowercase_ ) assert self.row == another.row and self.column == another.column # Add snake_case = Matrix(self.row, self.column ) for r in range(self.row ): for c in range(self.column ): snake_case = self[r, c] + another[r, c] return result def __neg__( self ) -> Matrix: snake_case = Matrix(self.row, self.column ) for r in range(self.row ): for c in range(self.column ): snake_case = -self[r, c] return result def __sub__( self, lowercase_ ) -> Matrix: return self + (-another) def __mul__( self, lowercase_ ) -> Matrix: if isinstance(lowercase_, (int, float) ): # Scalar multiplication snake_case = Matrix(self.row, self.column ) for r in range(self.row ): for c in range(self.column ): snake_case = self[r, c] * another return result elif isinstance(lowercase_, lowercase_ ): # Matrix multiplication assert self.column == another.row snake_case = Matrix(self.row, another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: snake_case = F'''Unsupported type given for another ({type(lowercase_ )})''' raise TypeError(lowercase_ ) def _lowerCamelCase ( self ) -> Matrix: snake_case = Matrix(self.column, self.row ) for r in range(self.row ): for c in range(self.column ): snake_case = self[r, c] return result def _lowerCamelCase ( self, lowercase_, lowercase_ ) -> Any: assert isinstance(lowercase_, lowercase_ ) and isinstance(lowercase_, lowercase_ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate snake_case = v.transpose() snake_case = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def __magic_name__ ( ) -> None: # a^(-1) snake_case = Matrix(3 , 3 , 0 ) for i in range(3 ): snake_case = 1 print(F'''a^(-1) is {ainv}''' ) # u, v snake_case = Matrix(3 , 1 , 0 ) snake_case , snake_case , snake_case = 1, 2, -3 snake_case = Matrix(3 , 1 , 0 ) snake_case , snake_case , snake_case = 4, -2, 5 print(F'''u is {u}''' ) print(F'''v is {v}''' ) print(F'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(A , A )}''' ) def __magic_name__ ( ) -> None: import doctest doctest.testmod() testa()
332
'''simple docstring''' import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def __magic_name__ ( A ) -> Tuple: snake_case = [] embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight''', F'''stage{idx}.patch_embed.proj.weight''', ) ) embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias''', F'''stage{idx}.patch_embed.proj.bias''', ) ) embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight''', F'''stage{idx}.patch_embed.norm.weight''', ) ) embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias''', F'''stage{idx}.patch_embed.norm.bias''', ) ) return embed def __magic_name__ ( A , A ) -> Optional[int]: snake_case = [] attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj_q.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj_q.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj_k.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj_k.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj_v.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj_v.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj.bias''', ) ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight''', F'''stage{idx}.blocks.{cnt}.mlp.fc1.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias''', F'''stage{idx}.blocks.{cnt}.mlp.fc1.bias''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight''', F'''stage{idx}.blocks.{cnt}.mlp.fc2.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias''', F'''stage{idx}.blocks.{cnt}.mlp.fc2.bias''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight''', F'''stage{idx}.blocks.{cnt}.norm1.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias''', F'''stage{idx}.blocks.{cnt}.norm1.bias''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight''', F'''stage{idx}.blocks.{cnt}.norm2.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias''', F'''stage{idx}.blocks.{cnt}.norm2.bias''') ) return attention_weights def __magic_name__ ( A ) -> List[Any]: snake_case = [] token.append((F'''cvt.encoder.stages.{idx}.cls_token''', 'stage2.cls_token') ) return token def __magic_name__ ( ) -> Dict: snake_case = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def __magic_name__ ( A , A , A , A ) -> int: snake_case = 'imagenet-1k-id2label.json' snake_case = 1_0_0_0 snake_case = 'huggingface/label-files' snake_case = num_labels snake_case = json.load(open(cached_download(hf_hub_url(A , A , repo_type='dataset' ) ) , 'r' ) ) snake_case = {int(A ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} snake_case = snake_case = CvtConfig(num_labels=A , idalabel=A , labelaid=A ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": snake_case = [1, 2, 1_0] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": snake_case = [1, 4, 1_6] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: snake_case = [2, 2, 2_0] snake_case = [3, 1_2, 1_6] snake_case = [1_9_2, 7_6_8, 1_0_2_4] snake_case = CvtForImageClassification(A ) snake_case = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) snake_case = image_size snake_case = torch.load(A , map_location=torch.device('cpu' ) ) snake_case = OrderedDict() snake_case = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: snake_case = list_of_state_dict + cls_token(A ) snake_case = list_of_state_dict + embeddings(A ) for cnt in range(config.depth[idx] ): snake_case = list_of_state_dict + attention(A , A ) snake_case = list_of_state_dict + final() for gg in list_of_state_dict: print(A ) for i in range(len(A ) ): snake_case = original_weights[list_of_state_dict[i][1]] model.load_state_dict(A ) model.save_pretrained(A ) image_processor.save_pretrained(A ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() parser.add_argument( "--cvt_model", default="cvt-w24", type=str, help="Name of the cvt model you'd like to convert.", ) parser.add_argument( "--image_size", default=3_8_4, type=int, help="Input Image Size", ) parser.add_argument( "--cvt_file_name", default=r"cvtmodels\CvT-w24-384x384-IN-22k.pth", type=str, help="Input Image Size", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) lowerCAmelCase_ = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
332
1
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging lowercase__ : int = logging.get_logger(__name__) lowercase__ : Any = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED lowercase__ : List[str] = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } lowercase__ : Optional[int] = { '''allenai/led-base-16384''': 1_6384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def lowerCamelCase__ ( ): '''simple docstring''' snake_case_ = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) snake_case_ = bs[:] snake_case_ = 0 for b in range(2**8 ): if b not in bs: bs.append(_A ) cs.append(2**8 + n ) n += 1 snake_case_ = [chr(_A ) for n in cs] return dict(zip(_A , _A ) ) def lowerCamelCase__ ( _A ): '''simple docstring''' snake_case_ = set() snake_case_ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) snake_case_ = char return pairs class UpperCAmelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ = VOCAB_FILES_NAMES lowerCAmelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ = ["""input_ids""", """attention_mask"""] def __init__( self : str , __lowercase : Optional[Any] , __lowercase : int , __lowercase : int="replace" , __lowercase : Union[str, Any]="<s>" , __lowercase : Union[str, Any]="</s>" , __lowercase : Any="</s>" , __lowercase : Optional[Any]="<s>" , __lowercase : List[Any]="<unk>" , __lowercase : List[Any]="<pad>" , __lowercase : Union[str, Any]="<mask>" , __lowercase : str=False , **__lowercase : Optional[int] , ): """simple docstring""" snake_case_ = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else bos_token snake_case_ = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else eos_token snake_case_ = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else sep_token snake_case_ = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else cls_token snake_case_ = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else unk_token snake_case_ = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else pad_token # Mask token behave like a normal word, i.e. include the space before it snake_case_ = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else mask_token super().__init__( errors=_A , bos_token=_A , eos_token=_A , unk_token=_A , sep_token=_A , cls_token=_A , pad_token=_A , mask_token=_A , add_prefix_space=_A , **_A , ) with open(_A , encoding="utf-8" ) as vocab_handle: snake_case_ = json.load(_A ) snake_case_ = {v: k for k, v in self.encoder.items()} snake_case_ = errors # how to handle errors in decoding snake_case_ = bytes_to_unicode() snake_case_ = {v: k for k, v in self.byte_encoder.items()} with open(_A , encoding="utf-8" ) as merges_handle: snake_case_ = merges_handle.read().split("\n" )[1:-1] snake_case_ = [tuple(merge.split() ) for merge in bpe_merges] snake_case_ = dict(zip(_A , range(len(_A ) ) ) ) snake_case_ = {} snake_case_ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions snake_case_ = re.compile(r"\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def snake_case__ ( self : Any ): """simple docstring""" return len(self.encoder ) def snake_case__ ( self : List[Any] ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def snake_case__ ( self : Optional[int] , __lowercase : List[str] ): """simple docstring""" if token in self.cache: return self.cache[token] snake_case_ = tuple(_A ) snake_case_ = get_pairs(_A ) if not pairs: return token while True: snake_case_ = min(_A , key=lambda __lowercase : self.bpe_ranks.get(_A , float("inf" ) ) ) if bigram not in self.bpe_ranks: break snake_case_ = bigram snake_case_ = [] snake_case_ = 0 while i < len(_A ): try: snake_case_ = word.index(_A , _A ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) snake_case_ = j if word[i] == first and i < len(_A ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 snake_case_ = tuple(_A ) snake_case_ = new_word if len(_A ) == 1: break else: snake_case_ = get_pairs(_A ) snake_case_ = ' '.join(_A ) snake_case_ = word return word def snake_case__ ( self : Union[str, Any] , __lowercase : List[str] ): """simple docstring""" snake_case_ = [] for token in re.findall(self.pat , _A ): snake_case_ = ''.join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_A ).split(" " ) ) return bpe_tokens def snake_case__ ( self : Optional[int] , __lowercase : Dict ): """simple docstring""" return self.encoder.get(_A , self.encoder.get(self.unk_token ) ) def snake_case__ ( self : Dict , __lowercase : Dict ): """simple docstring""" return self.decoder.get(_A ) def snake_case__ ( self : Optional[Any] , __lowercase : Optional[Any] ): """simple docstring""" snake_case_ = ''.join(_A ) snake_case_ = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def snake_case__ ( self : int , __lowercase : str , __lowercase : Optional[str] = None ): """simple docstring""" if not os.path.isdir(_A ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return snake_case_ = os.path.join( _A , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) snake_case_ = os.path.join( _A , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(_A , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_A , ensure_ascii=_A ) + "\n" ) snake_case_ = 0 with open(_A , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __lowercase : kv[1] ): if index != token_index: logger.warning( f"Saving vocabulary to {merge_file}: BPE merge indices are not consecutive." " Please check that the tokenizer is not corrupted!" ) snake_case_ = token_index writer.write(" ".join(_A ) + "\n" ) index += 1 return vocab_file, merge_file def snake_case__ ( self : Union[str, Any] , __lowercase : List[int] , __lowercase : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] snake_case_ = [self.cls_token_id] snake_case_ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def snake_case__ ( self : Optional[Any] , __lowercase : List[int] , __lowercase : Optional[List[int]] = None , __lowercase : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_A , token_ids_a=_A , already_has_special_tokens=_A ) if token_ids_a is None: return [1] + ([0] * len(_A )) + [1] return [1] + ([0] * len(_A )) + [1, 1] + ([0] * len(_A )) + [1] def snake_case__ ( self : str , __lowercase : List[int] , __lowercase : Optional[List[int]] = None ): """simple docstring""" snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case__ ( self : int , __lowercase : Any , __lowercase : Any=False , **__lowercase : List[str] ): """simple docstring""" snake_case_ = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(_A ) > 0 and not text[0].isspace()): snake_case_ = ' ' + text return (text, kwargs) def snake_case__ ( self : Any , __lowercase : Union[Dict[str, EncodedInput], BatchEncoding] , __lowercase : Optional[int] = None , __lowercase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , __lowercase : Optional[int] = None , __lowercase : Optional[bool] = None , ): """simple docstring""" snake_case_ = super()._pad( encoded_inputs=_A , max_length=_A , padding_strategy=_A , pad_to_multiple_of=_A , return_attention_mask=_A , ) # Load from model defaults if return_attention_mask is None: snake_case_ = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: snake_case_ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. snake_case_ = len(encoded_inputs["global_attention_mask"] ) != len(_A ) if needs_to_be_padded: snake_case_ = len(_A ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` snake_case_ = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": snake_case_ = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
187
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax lowerCAmelCase :Any = logging.get_logger(__name__) @add_end_docstrings(lowercase__ ) class _lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self : Optional[Any] , **_A : Union[str, Any] ) -> Tuple: super().__init__(**_A ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING ) def __call__( self : Optional[int] , _A : Union[str, List[str], "Image", List["Image"]] , **_A : Dict ) -> Dict: return super().__call__(_A , **_A ) def __lowerCAmelCase ( self : Any , **_A : Dict ) -> Optional[int]: __magic_name__ : str = {} if "candidate_labels" in kwargs: __magic_name__ : str = kwargs['candidate_labels'] if "hypothesis_template" in kwargs: __magic_name__ : Tuple = kwargs['hypothesis_template'] return preprocess_params, {}, {} def __lowerCAmelCase ( self : str , _A : Dict , _A : Optional[Any]=None , _A : int="This is a photo of {}." ) -> int: __magic_name__ : Dict = load_image(_A ) __magic_name__ : List[str] = self.image_processor(images=[image] , return_tensors=self.framework ) __magic_name__ : Optional[Any] = candidate_labels __magic_name__ : List[Any] = [hypothesis_template.format(_A ) for x in candidate_labels] __magic_name__ : str = self.tokenizer(_A , return_tensors=self.framework , padding=_A ) __magic_name__ : Optional[Any] = [text_inputs] return inputs def __lowerCAmelCase ( self : Union[str, Any] , _A : Tuple ) -> str: __magic_name__ : str = model_inputs.pop('candidate_labels' ) __magic_name__ : str = model_inputs.pop('text_inputs' ) if isinstance(text_inputs[0] , _A ): __magic_name__ : Dict = text_inputs[0] else: # Batching case. __magic_name__ : Optional[Any] = text_inputs[0][0] __magic_name__ : List[Any] = self.model(**_A , **_A ) __magic_name__ : str = { 'candidate_labels': candidate_labels, 'logits': outputs.logits_per_image, } return model_outputs def __lowerCAmelCase ( self : Optional[int] , _A : Optional[Any] ) -> Optional[int]: __magic_name__ : Tuple = model_outputs.pop('candidate_labels' ) __magic_name__ : Union[str, Any] = model_outputs['logits'][0] if self.framework == "pt": __magic_name__ : Tuple = logits.softmax(dim=-1 ).squeeze(-1 ) __magic_name__ : Tuple = probs.tolist() if not isinstance(_A , _A ): __magic_name__ : Any = [scores] elif self.framework == "tf": __magic_name__ : Any = stable_softmax(_A , axis=-1 ) __magic_name__ : Dict = probs.numpy().tolist() else: raise ValueError(F'Unsupported framework: {self.framework}' ) __magic_name__ : Union[str, Any] = [ {'score': score, 'label': candidate_label} for score, candidate_label in sorted(zip(_A , _A ) , key=lambda _A : -x[0] ) ] return result
331
0
'''simple docstring''' import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class UpperCamelCase_ (unittest.TestCase ): """simple docstring""" def _a ( self : List[str] , _lowerCamelCase : Any ): """simple docstring""" for model_result in results.values(): for batch_size, sequence_length in zip(model_result['''bs'''] , model_result['''ss'''] ): A_ : Optional[int] = model_result['''result'''][batch_size][sequence_length] self.assertIsNotNone(_lowerCamelCase ) def _a ( self : Dict ): """simple docstring""" A_ : Dict = '''sshleifer/tiny-gpt2''' A_ : Optional[int] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , ) A_ : Union[str, Any] = PyTorchBenchmark(_lowerCamelCase ) A_ : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _a ( self : Optional[Any] ): """simple docstring""" A_ : Union[str, Any] = '''sgugger/tiny-distilbert-classification''' A_ : int = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , only_pretrain_model=_lowerCamelCase , ) A_ : List[str] = PyTorchBenchmark(_lowerCamelCase ) A_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _a ( self : List[Any] ): """simple docstring""" A_ : Union[str, Any] = '''sshleifer/tiny-gpt2''' A_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , torchscript=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , ) A_ : List[Any] = PyTorchBenchmark(_lowerCamelCase ) A_ : Union[str, Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def _a ( self : Union[str, Any] ): """simple docstring""" A_ : List[str] = '''sshleifer/tiny-gpt2''' A_ : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , fpaa=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , ) A_ : int = PyTorchBenchmark(_lowerCamelCase ) A_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _a ( self : Tuple ): """simple docstring""" A_ : int = '''sshleifer/tiny-gpt2''' A_ : Dict = AutoConfig.from_pretrained(_lowerCamelCase ) # set architectures equal to `None` A_ : List[str] = None A_ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , ) A_ : Optional[int] = PyTorchBenchmark(_lowerCamelCase , configs=[config] ) A_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _a ( self : Union[str, Any] ): """simple docstring""" A_ : Union[str, Any] = '''sshleifer/tiny-gpt2''' A_ : List[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , ) A_ : List[Any] = PyTorchBenchmark(_lowerCamelCase ) A_ : Any = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == '''cpu''' , '''Can\'t do half precision''' ) def _a ( self : Optional[int] ): """simple docstring""" A_ : Optional[int] = '''sshleifer/tiny-gpt2''' A_ : Union[str, Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , fpaa=_lowerCamelCase , multi_process=_lowerCamelCase , ) A_ : int = PyTorchBenchmark(_lowerCamelCase ) A_ : str = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _a ( self : Tuple ): """simple docstring""" A_ : Optional[Any] = '''sshleifer/tiny-gpt2''' A_ : int = AutoConfig.from_pretrained(_lowerCamelCase ) A_ : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , ) A_ : str = PyTorchBenchmark(_lowerCamelCase , configs=[config] ) A_ : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _a ( self : Union[str, Any] ): """simple docstring""" A_ : int = '''sshleifer/tinier_bart''' A_ : str = AutoConfig.from_pretrained(_lowerCamelCase ) A_ : Union[str, Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , ) A_ : Any = PyTorchBenchmark(_lowerCamelCase , configs=[config] ) A_ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _a ( self : Dict ): """simple docstring""" A_ : Union[str, Any] = '''sshleifer/tiny-gpt2''' A_ : Optional[int] = AutoConfig.from_pretrained(_lowerCamelCase ) A_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , ) A_ : List[str] = PyTorchBenchmark(_lowerCamelCase , configs=[config] ) A_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _a ( self : Tuple ): """simple docstring""" A_ : Optional[Any] = '''sshleifer/tinier_bart''' A_ : List[Any] = AutoConfig.from_pretrained(_lowerCamelCase ) A_ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowerCamelCase , ) A_ : List[Any] = PyTorchBenchmark(_lowerCamelCase , configs=[config] ) A_ : int = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _a ( self : int ): """simple docstring""" A_ : str = '''sshleifer/tiny-gpt2''' with tempfile.TemporaryDirectory() as tmp_dir: A_ : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , save_to_csv=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(_lowerCamelCase , '''inf_time.csv''' ) , train_memory_csv_file=os.path.join(_lowerCamelCase , '''train_mem.csv''' ) , inference_memory_csv_file=os.path.join(_lowerCamelCase , '''inf_mem.csv''' ) , train_time_csv_file=os.path.join(_lowerCamelCase , '''train_time.csv''' ) , env_info_csv_file=os.path.join(_lowerCamelCase , '''env.csv''' ) , multi_process=_lowerCamelCase , ) A_ : int = PyTorchBenchmark(_lowerCamelCase ) benchmark.run() self.assertTrue(Path(os.path.join(_lowerCamelCase , '''inf_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowerCamelCase , '''train_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowerCamelCase , '''inf_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowerCamelCase , '''train_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowerCamelCase , '''env.csv''' ) ).exists() ) def _a ( self : Optional[int] ): """simple docstring""" A_ : Any = '''sshleifer/tiny-gpt2''' def _check_summary_is_not_empty(_lowerCamelCase : Dict ): self.assertTrue(hasattr(_lowerCamelCase , '''sequential''' ) ) self.assertTrue(hasattr(_lowerCamelCase , '''cumulative''' ) ) self.assertTrue(hasattr(_lowerCamelCase , '''current''' ) ) self.assertTrue(hasattr(_lowerCamelCase , '''total''' ) ) with tempfile.TemporaryDirectory() as tmp_dir: A_ : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowerCamelCase , inference=_lowerCamelCase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(_lowerCamelCase , '''log.txt''' ) , log_print=_lowerCamelCase , trace_memory_line_by_line=_lowerCamelCase , multi_process=_lowerCamelCase , ) A_ : str = PyTorchBenchmark(_lowerCamelCase ) A_ : Any = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(_lowerCamelCase , '''log.txt''' ) ).exists() )
362
'''simple docstring''' from collections.abc import Sequence def snake_case__ ( lowerCamelCase__ : Sequence[float] , lowerCamelCase__ : bool = False ) -> float: if not arr: return 0 A_ : Union[str, Any] = 0 if allow_empty_subarrays else float('''-inf''' ) A_ : str = 0.0 for num in arr: A_ : Any = max(0 if allow_empty_subarrays else num , curr_sum + num ) A_ : Tuple = max(lowerCamelCase__ , lowerCamelCase__ ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() snake_case__ = [-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F'{max_subarray_sum(nums) = }')
4
0
"""simple docstring""" import math def A_ ( ): """simple docstring""" _a = input('''Enter message: ''' ) _a = int(input(f'Enter key [2-{len(_lowerCAmelCase ) - 1}]: ' ) ) _a = input('''Encryption/Decryption [e/d]: ''' ) if mode.lower().startswith('''e''' ): _a = encrypt_message(_lowerCAmelCase, _lowerCAmelCase ) elif mode.lower().startswith('''d''' ): _a = decrypt_message(_lowerCAmelCase, _lowerCAmelCase ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(f'Output:\n{text + "|"}' ) def A_ ( _lowerCAmelCase : int, _lowerCAmelCase : str ): """simple docstring""" _a = [''''''] * key for col in range(_lowerCAmelCase ): _a = col while pointer < len(_lowerCAmelCase ): cipher_text[col] += message[pointer] pointer += key return "".join(_lowerCAmelCase ) def A_ ( _lowerCAmelCase : int, _lowerCAmelCase : str ): """simple docstring""" _a = math.ceil(len(_lowerCAmelCase ) / key ) _a = key _a = (num_cols * num_rows) - len(_lowerCAmelCase ) _a = [''''''] * num_cols _a = 0 _a = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): _a = 0 row += 1 return "".join(_lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
320
"""simple docstring""" import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging __snake_case = logging.get_logger(__name__) __snake_case = { '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/config.json''', # See all BART models at https://huggingface.co/models?filter=bart } class __lowerCamelCase ( a__ ): '''simple docstring''' A_ : List[str] = 'bart' A_ : Optional[Any] = ['past_key_values'] A_ : Union[str, Any] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , __UpperCAmelCase=50265 , __UpperCAmelCase=1024 , __UpperCAmelCase=12 , __UpperCAmelCase=4096 , __UpperCAmelCase=16 , __UpperCAmelCase=12 , __UpperCAmelCase=4096 , __UpperCAmelCase=16 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , __UpperCAmelCase="gelu" , __UpperCAmelCase=1024 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.02 , __UpperCAmelCase=0.0 , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=3 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=2 , __UpperCAmelCase=True , __UpperCAmelCase=2 , __UpperCAmelCase=2 , **__UpperCAmelCase , ) -> Tuple: _a = vocab_size _a = max_position_embeddings _a = d_model _a = encoder_ffn_dim _a = encoder_layers _a = encoder_attention_heads _a = decoder_ffn_dim _a = decoder_layers _a = decoder_attention_heads _a = dropout _a = attention_dropout _a = activation_dropout _a = activation_function _a = init_std _a = encoder_layerdrop _a = decoder_layerdrop _a = classifier_dropout _a = use_cache _a = encoder_layers _a = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=__UpperCAmelCase , pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , is_encoder_decoder=__UpperCAmelCase , decoder_start_token_id=__UpperCAmelCase , forced_eos_token_id=__UpperCAmelCase , **__UpperCAmelCase , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' , __UpperCAmelCase ): _a = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' '''The config can simply be saved and uploaded again to be fixed.''' ) class __lowerCamelCase ( a__ ): '''simple docstring''' @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: _a = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: _a = {0: '''batch'''} _a = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: _a = {0: '''batch''', 1: '''decoder_sequence'''} _a = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(__UpperCAmelCase , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. _a = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: _a , _a = self.num_layers for i in range(__UpperCAmelCase ): _a = {0: '''batch''', 2: '''past_sequence + sequence'''} _a = {0: '''batch''', 2: '''past_sequence + sequence'''} else: _a = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: _a = super().outputs else: _a = super(__UpperCAmelCase , self ).outputs if self.use_past: _a , _a = self.num_layers for i in range(__UpperCAmelCase ): _a = {0: '''batch''', 2: '''past_sequence + sequence'''} _a = {0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = -1 , __UpperCAmelCase = -1 , __UpperCAmelCase = False , __UpperCAmelCase = None , ) -> Mapping[str, Any]: _a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # Generate decoder inputs _a = seq_length if not self.use_past else 1 _a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) _a = {F'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} _a = dict(**__UpperCAmelCase , **__UpperCAmelCase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch _a , _a = common_inputs['''input_ids'''].shape _a = common_inputs['''decoder_input_ids'''].shape[1] _a , _a = self.num_attention_heads _a = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) _a = decoder_seq_length + 3 _a = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) _a = torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(__UpperCAmelCase , __UpperCAmelCase )] , dim=1 ) _a = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered _a , _a = self.num_layers _a = min(__UpperCAmelCase , __UpperCAmelCase ) _a = max(__UpperCAmelCase , __UpperCAmelCase ) - min_num_layers _a = '''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(__UpperCAmelCase ): common_inputs["past_key_values"].append( ( torch.zeros(__UpperCAmelCase ), torch.zeros(__UpperCAmelCase ), torch.zeros(__UpperCAmelCase ), torch.zeros(__UpperCAmelCase ), ) ) # TODO: test this. _a = encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(__UpperCAmelCase , __UpperCAmelCase ): common_inputs["past_key_values"].append((torch.zeros(__UpperCAmelCase ), torch.zeros(__UpperCAmelCase )) ) return common_inputs def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = -1 , __UpperCAmelCase = -1 , __UpperCAmelCase = False , __UpperCAmelCase = None , ) -> Mapping[str, Any]: _a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch _a , _a = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values _a = seqlen + 2 _a , _a = self.num_layers _a , _a = self.num_attention_heads _a = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) _a = common_inputs['''attention_mask'''].dtype _a = torch.cat( [common_inputs['''attention_mask'''], torch.ones(__UpperCAmelCase , __UpperCAmelCase , dtype=__UpperCAmelCase )] , dim=1 ) _a = [ (torch.zeros(__UpperCAmelCase ), torch.zeros(__UpperCAmelCase )) for _ in range(__UpperCAmelCase ) ] return common_inputs def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = -1 , __UpperCAmelCase = -1 , __UpperCAmelCase = False , __UpperCAmelCase = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _a = compute_effective_axis_dimension( __UpperCAmelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX _a = tokenizer.num_special_tokens_to_add(__UpperCAmelCase ) _a = compute_effective_axis_dimension( __UpperCAmelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__UpperCAmelCase ) # Generate dummy inputs according to compute batch and sequence _a = [''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size _a = dict(tokenizer(__UpperCAmelCase , return_tensors=__UpperCAmelCase ) ) return common_inputs def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = -1 , __UpperCAmelCase = -1 , __UpperCAmelCase = False , __UpperCAmelCase = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: _a = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __UpperCAmelCase , batch_size=__UpperCAmelCase , seq_length=__UpperCAmelCase , is_pair=__UpperCAmelCase , framework=__UpperCAmelCase ) elif self.task == "causal-lm": _a = self._generate_dummy_inputs_for_causal_lm( __UpperCAmelCase , batch_size=__UpperCAmelCase , seq_length=__UpperCAmelCase , is_pair=__UpperCAmelCase , framework=__UpperCAmelCase ) else: _a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __UpperCAmelCase , batch_size=__UpperCAmelCase , seq_length=__UpperCAmelCase , is_pair=__UpperCAmelCase , framework=__UpperCAmelCase ) return common_inputs def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: if self.task in ["default", "seq2seq-lm"]: _a = super()._flatten_past_key_values_(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) else: _a = super(__UpperCAmelCase , self )._flatten_past_key_values_( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase )
320
1
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class SCREAMING_SNAKE_CASE : def __init__( self , _UpperCAmelCase , _UpperCAmelCase=13 , _UpperCAmelCase=7 , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=99 , _UpperCAmelCase=32 , _UpperCAmelCase=2 , _UpperCAmelCase=4 , _UpperCAmelCase=37 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=512 , _UpperCAmelCase=16 , _UpperCAmelCase=2 , _UpperCAmelCase=0.02 , _UpperCAmelCase=3 , _UpperCAmelCase=4 , _UpperCAmelCase=None , _UpperCAmelCase=1000 , ): '''simple docstring''' __A : str = parent __A : List[Any] = batch_size __A : str = seq_length __A : int = is_training __A : int = use_input_mask __A : int = use_token_type_ids __A : List[Any] = use_labels __A : List[str] = vocab_size __A : Any = hidden_size __A : Dict = num_hidden_layers __A : Union[str, Any] = num_attention_heads __A : Dict = intermediate_size __A : List[Any] = hidden_act __A : List[Any] = hidden_dropout_prob __A : Optional[int] = attention_probs_dropout_prob __A : Any = max_position_embeddings __A : int = type_vocab_size __A : List[Any] = type_sequence_label_size __A : List[Any] = initializer_range __A : Any = num_labels __A : Any = num_choices __A : Any = scope __A : Optional[Any] = range_bbox def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) # convert bbox to numpy since TF does not support item assignment __A : Dict = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0]): for j in range(bbox.shape[1]): if bbox[i, j, 3] < bbox[i, j, 1]: __A : Optional[Any] = bbox[i, j, 3] __A : str = bbox[i, j, 1] __A : Dict = t if bbox[i, j, 2] < bbox[i, j, 0]: __A : Tuple = bbox[i, j, 2] __A : int = bbox[i, j, 0] __A : int = t __A : str = tf.convert_to_tensor(_UpperCAmelCase) __A : Optional[Any] = None if self.use_input_mask: __A : Tuple = random_attention_mask([self.batch_size, self.seq_length]) __A : Union[str, Any] = None if self.use_token_type_ids: __A : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) __A : Union[str, Any] = None __A : str = None __A : Any = None if self.use_labels: __A : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size) __A : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __A : Dict = ids_tensor([self.batch_size] , self.num_choices) __A : List[Any] = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' __A : Dict = TFLayoutLMModel(config=_UpperCAmelCase) __A : Union[str, Any] = model(_UpperCAmelCase , _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase) __A : Tuple = model(_UpperCAmelCase , _UpperCAmelCase , token_type_ids=_UpperCAmelCase) __A : Union[str, Any] = model(_UpperCAmelCase , _UpperCAmelCase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' __A : Dict = TFLayoutLMForMaskedLM(config=_UpperCAmelCase) __A : str = model(_UpperCAmelCase , _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' __A : str = self.num_labels __A : Optional[int] = TFLayoutLMForSequenceClassification(config=_UpperCAmelCase) __A : str = model(_UpperCAmelCase , _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' __A : Optional[int] = self.num_labels __A : int = TFLayoutLMForTokenClassification(config=_UpperCAmelCase) __A : Tuple = model(_UpperCAmelCase , _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' __A : Dict = TFLayoutLMForQuestionAnswering(config=_UpperCAmelCase) __A : Dict = model(_UpperCAmelCase , _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : int = self.prepare_config_and_inputs() ( ( __A ) ,( __A ) ,( __A ) ,( __A ) ,( __A ) ,( __A ) ,( __A ) ,( __A ) , ) : List[Any] = config_and_inputs __A : Optional[int] = { 'input_ids': input_ids, 'bbox': bbox, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE (a__ , a__ , unittest.TestCase ): lowerCAmelCase = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) lowerCAmelCase = ( { '''feature-extraction''': TFLayoutLMModel, '''fill-mask''': TFLayoutLMForMaskedLM, '''text-classification''': TFLayoutLMForSequenceClassification, '''token-classification''': TFLayoutLMForTokenClassification, '''zero-shot''': TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase = False lowerCAmelCase = True lowerCAmelCase = 10 def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : int = TFLayoutLMModelTester(self) __A : Any = ConfigTester(self , config_class=_UpperCAmelCase , hidden_size=37) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_UpperCAmelCase) @slow def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __A : int = TFLayoutLMModel.from_pretrained(_UpperCAmelCase) self.assertIsNotNone(_UpperCAmelCase) @unittest.skip('Onnx compliancy broke with TF 2.10') def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' pass def _lowerCAmelCase ( ) -> int: # Here we prepare a batch of 2 sequences to test a LayoutLM forward pass on: # fmt: off __A : List[str] = tf.convert_to_tensor([[1_01,10_19,10_14,10_16,10_37,1_28_49,47_47,10_04,1_42_46,22_78,54_39,45_24,50_02,29_30,21_93,29_30,43_41,32_08,10_05,10_55,21_71,28_48,1_13_00,35_31,1_02],[1_01,40_70,40_34,70_20,10_24,30_58,10_15,10_13,28_61,10_13,60_70,1_92_74,27_72,62_05,2_78_14,1_61_47,1_61_47,43_43,20_47,1_02_83,1_09_69,1_43_89,10_12,23_38,1_02]] ) # noqa: E231 __A : List[str] = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 __A : str = tf.convert_to_tensor([[[0,0,0,0],[4_23,2_37,4_40,2_51],[4_27,2_72,4_41,2_87],[4_19,1_15,4_37,1_29],[9_61,8_85,9_92,9_12],[2_56,38,3_30,58],[2_56,38,3_30,58],[3_36,42,3_53,57],[3_60,39,4_01,56],[3_60,39,4_01,56],[4_11,39,4_71,59],[4_79,41,5_28,59],[5_33,39,6_30,60],[67,1_13,1_34,1_31],[1_41,1_15,2_09,1_32],[68,1_49,1_33,1_66],[1_41,1_49,1_87,1_64],[1_95,1_48,2_87,1_65],[1_95,1_48,2_87,1_65],[1_95,1_48,2_87,1_65],[2_95,1_48,3_49,1_65],[4_41,1_49,4_92,1_66],[4_97,1_49,5_46,1_64],[64,2_01,1_25,2_18],[10_00,10_00,10_00,10_00]],[[0,0,0,0],[6_62,1_50,7_54,1_66],[6_65,1_99,7_42,2_11],[5_19,2_13,5_54,2_28],[5_19,2_13,5_54,2_28],[1_34,4_33,1_87,4_54],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[3_14,4_69,3_76,4_82],[5_04,6_84,5_82,7_06],[9_41,8_25,9_73,9_00],[9_41,8_25,9_73,9_00],[9_41,8_25,9_73,9_00],[9_41,8_25,9_73,9_00],[6_10,7_49,6_52,7_65],[1_30,6_59,1_68,6_72],[1_76,6_57,2_37,6_72],[2_38,6_57,3_12,6_72],[4_43,6_53,6_28,6_72],[4_43,6_53,6_28,6_72],[7_16,3_01,8_25,3_17],[10_00,10_00,10_00,10_00]]] ) # noqa: E231 __A : Optional[Any] = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) __A : str = tf.convert_to_tensor([[-1_00,10,10,10,9,1,-1_00,7,7,-1_00,7,7,4,2,5,2,8,8,-1_00,-1_00,5,0,3,2,-1_00],[-1_00,12,12,12,-1_00,12,10,-1_00,-1_00,-1_00,-1_00,10,12,9,-1_00,-1_00,-1_00,10,10,10,9,12,-1_00,10,-1_00]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class SCREAMING_SNAKE_CASE (unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Any = TFLayoutLMModel.from_pretrained('microsoft/layoutlm-base-uncased') __A ,__A ,__A ,__A ,__A : int = prepare_layoutlm_batch_inputs() # forward pass __A : List[Any] = model(input_ids=_UpperCAmelCase , bbox=_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase) # test the sequence output on [0, :3, :3] __A : Dict = tf.convert_to_tensor( [[0.1785, -0.1947, -0.0425], [-0.3254, -0.2807, 0.2553], [-0.5391, -0.3322, 0.3364]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , _UpperCAmelCase , atol=1e-3)) # test the pooled output on [1, :3] __A : Union[str, Any] = tf.convert_to_tensor([-0.6580, -0.0214, 0.8552]) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , _UpperCAmelCase , atol=1e-3)) @slow def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Union[str, Any] = TFLayoutLMForSequenceClassification.from_pretrained('microsoft/layoutlm-base-uncased' , num_labels=2) __A ,__A ,__A ,__A ,__A : Tuple = prepare_layoutlm_batch_inputs() # forward pass __A : str = model( input_ids=_UpperCAmelCase , bbox=_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=tf.convert_to_tensor([1, 1]) , ) # test whether we get a loss as a scalar __A : Union[str, Any] = outputs.loss __A : Any = (2,) self.assertEqual(loss.shape , _UpperCAmelCase) # test the shape of the logits __A : Tuple = outputs.logits __A : List[str] = (2, 2) self.assertEqual(logits.shape , _UpperCAmelCase) @slow def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = TFLayoutLMForTokenClassification.from_pretrained('microsoft/layoutlm-base-uncased' , num_labels=13) __A ,__A ,__A ,__A ,__A : int = prepare_layoutlm_batch_inputs() # forward pass __A : Dict = model( input_ids=_UpperCAmelCase , bbox=_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase) # test the shape of the logits __A : Any = outputs.logits __A : Optional[Any] = tf.convert_to_tensor((2, 25, 13)) self.assertEqual(logits.shape , _UpperCAmelCase) @slow def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[str] = TFLayoutLMForQuestionAnswering.from_pretrained('microsoft/layoutlm-base-uncased') __A ,__A ,__A ,__A ,__A : Optional[Any] = prepare_layoutlm_batch_inputs() # forward pass __A : Union[str, Any] = model(input_ids=_UpperCAmelCase , bbox=_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase) # test the shape of the logits __A : Dict = tf.convert_to_tensor((2, 25)) self.assertEqual(outputs.start_logits.shape , _UpperCAmelCase) self.assertEqual(outputs.end_logits.shape , _UpperCAmelCase)
190
'''simple docstring''' import argparse lowercase__ : Any = '''docs/source/_static/js/custom.js''' def _lowerCAmelCase ( __snake_case : Union[str, Any] ) -> str: with open(__snake_case , encoding='utf-8' , newline='\n' ) as f: __A : Optional[Any] = f.readlines() __A : List[str] = 0 # First let's put the right version while not lines[index].startswith('const stableVersion =' ): index += 1 __A : Tuple = f'const stableVersion = "v{version}"\n' # Then update the dictionary while not lines[index].startswith('const versionMapping = {' ): index += 1 # We go until the end while not lines[index].startswith('}' ): index += 1 # We add the new version at the end lines[index - 1] += f' "v{version}": "v{version}",\n' with open(__snake_case , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(__snake_case ) if __name__ == "__main__": lowercase__ : Optional[Any] = argparse.ArgumentParser() parser.add_argument('''--version''', help='''Release version.''') lowercase__ : List[str] = parser.parse_args() update_custom_js(args.version)
190
1
"""simple docstring""" from typing import Dict, Optional import numpy as np import datasets __A = "\nIoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union\nbetween the predicted segmentation and the ground truth. For binary (two classes) or multi-class segmentation,\nthe mean IoU of the image is calculated by taking the IoU of each class and averaging them.\n" __A = "\nArgs:\n predictions (`List[ndarray]`):\n List of predicted segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n references (`List[ndarray]`):\n List of ground truth segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n num_labels (`int`):\n Number of classes (categories).\n ignore_index (`int`):\n Index that will be ignored during evaluation.\n nan_to_num (`int`, *optional*):\n If specified, NaN values will be replaced by the number defined by the user.\n label_map (`dict`, *optional*):\n If specified, dictionary mapping old label indices to new label indices.\n reduce_labels (`bool`, *optional*, defaults to `False`):\n Whether or not to reduce all label values of segmentation maps by 1. Usually used for datasets where 0 is used for background,\n and background itself is not included in all classes of a dataset (e.g. ADE20k). The background label will be replaced by 255.\n\nReturns:\n `Dict[str, float | ndarray]` comprising various elements:\n - *mean_iou* (`float`):\n Mean Intersection-over-Union (IoU averaged over all categories).\n - *mean_accuracy* (`float`):\n Mean accuracy (averaged over all categories).\n - *overall_accuracy* (`float`):\n Overall accuracy on all images.\n - *per_category_accuracy* (`ndarray` of shape `(num_labels,)`):\n Per category accuracy.\n - *per_category_iou* (`ndarray` of shape `(num_labels,)`):\n Per category IoU.\n\nExamples:\n\n >>> import numpy as np\n\n >>> mean_iou = datasets.load_metric(\"mean_iou\")\n\n >>> # suppose one has 3 different segmentation maps predicted\n >>> predicted_1 = np.array([[1, 2], [3, 4], [5, 255]])\n >>> actual_1 = np.array([[0, 3], [5, 4], [6, 255]])\n\n >>> predicted_2 = np.array([[2, 7], [9, 2], [3, 6]])\n >>> actual_2 = np.array([[1, 7], [9, 2], [3, 6]])\n\n >>> predicted_3 = np.array([[2, 2, 3], [8, 2, 4], [3, 255, 2]])\n >>> actual_3 = np.array([[1, 2, 2], [8, 2, 1], [3, 255, 1]])\n\n >>> predicted = [predicted_1, predicted_2, predicted_3]\n >>> ground_truth = [actual_1, actual_2, actual_3]\n\n >>> results = mean_iou.compute(predictions=predicted, references=ground_truth, num_labels=10, ignore_index=255, reduce_labels=False)\n >>> print(results) # doctest: +NORMALIZE_WHITESPACE\n {\'mean_iou\': 0.47750000000000004, \'mean_accuracy\': 0.5916666666666666, \'overall_accuracy\': 0.5263157894736842, \'per_category_iou\': array([0. , 0. , 0.375, 0.4 , 0.5 , 0. , 0.5 , 1. , 1. , 1. ]), \'per_category_accuracy\': array([0. , 0. , 0.75 , 0.66666667, 1. , 0. , 0.5 , 1. , 1. , 1. ])}\n" __A = "\\n@software{MMSegmentation_Contributors_OpenMMLab_Semantic_Segmentation_2020,\nauthor = {{MMSegmentation Contributors}},\nlicense = {Apache-2.0},\nmonth = {7},\ntitle = {{OpenMMLab Semantic Segmentation Toolbox and Benchmark}},\nurl = {https://github.com/open-mmlab/mmsegmentation},\nyear = {2020}\n}" def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False , ) -> int: if label_map is not None: for old_id, new_id in label_map.items(): lowercase__: Union[str, Any] = new_id # turn into Numpy arrays lowercase__: str = np.array(UpperCamelCase__ ) lowercase__: List[Any] = np.array(UpperCamelCase__ ) if reduce_labels: lowercase__: str = 2_5_5 lowercase__: Union[str, Any] = label - 1 lowercase__: Optional[int] = 2_5_5 lowercase__: List[Any] = label != ignore_index lowercase__: Any = np.not_equal(UpperCamelCase__ , UpperCamelCase__ ) lowercase__: str = pred_label[mask] lowercase__: Any = np.array(UpperCamelCase__ )[mask] lowercase__: Tuple = pred_label[pred_label == label] lowercase__: int = np.histogram(UpperCamelCase__ , bins=UpperCamelCase__ , range=(0, num_labels - 1) )[0] lowercase__: List[Any] = np.histogram(UpperCamelCase__ , bins=UpperCamelCase__ , range=(0, num_labels - 1) )[0] lowercase__: Optional[int] = np.histogram(UpperCamelCase__ , bins=UpperCamelCase__ , range=(0, num_labels - 1) )[0] lowercase__: Tuple = area_pred_label + area_label - area_intersect return area_intersect, area_union, area_pred_label, area_label def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False , ) -> Union[str, Any]: lowercase__: List[Any] = np.zeros((num_labels,) , dtype=np.floataa ) lowercase__: int = np.zeros((num_labels,) , dtype=np.floataa ) lowercase__: Union[str, Any] = np.zeros((num_labels,) , dtype=np.floataa ) lowercase__: int = np.zeros((num_labels,) , dtype=np.floataa ) for result, gt_seg_map in zip(UpperCamelCase__ , UpperCamelCase__ ): lowercase__: str = intersect_and_union( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) total_area_intersect += area_intersect total_area_union += area_union total_area_pred_label += area_pred_label total_area_label += area_label return total_area_intersect, total_area_union, total_area_pred_label, total_area_label def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , ) -> List[str]: lowercase__: Optional[int] = total_intersect_and_union( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # compute metrics lowercase__: List[str] = {} lowercase__: Any = total_area_intersect.sum() / total_area_label.sum() lowercase__: Union[str, Any] = total_area_intersect / total_area_union lowercase__: Union[str, Any] = total_area_intersect / total_area_label lowercase__: Union[str, Any] = np.nanmean(UpperCamelCase__ ) lowercase__: List[Any] = np.nanmean(UpperCamelCase__ ) lowercase__: List[str] = all_acc lowercase__: Tuple = iou lowercase__: Tuple = acc if nan_to_num is not None: lowercase__: List[Any] = {metric: np.nan_to_num(UpperCamelCase__ , nan=UpperCamelCase__ ) for metric, metric_value in metrics.items()} return metrics @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class UpperCAmelCase (datasets.Metric ): """simple docstring""" def _snake_case ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( # 1st Seq - height dim, 2nd - width dim { '''predictions''': datasets.Sequence(datasets.Sequence(datasets.Value('''uint16''' ) ) ), '''references''': datasets.Sequence(datasets.Sequence(datasets.Value('''uint16''' ) ) ), } ) , reference_urls=[ '''https://github.com/open-mmlab/mmsegmentation/blob/71c201b1813267d78764f306a297ca717827c4bf/mmseg/core/evaluation/metrics.py''' ] , ) def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = False , ): lowercase__: int = mean_iou( results=__lowerCamelCase , gt_seg_maps=__lowerCamelCase , num_labels=__lowerCamelCase , ignore_index=__lowerCamelCase , nan_to_num=__lowerCamelCase , label_map=__lowerCamelCase , reduce_labels=__lowerCamelCase , ) return iou_result
177
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: lowerCAmelCase__ = None lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} lowerCAmelCase__ = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'facebook/mbart-large-en-ro': 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json', 'facebook/mbart-large-cc25': 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json', }, } lowerCAmelCase__ = { 'facebook/mbart-large-en-ro': 10_24, 'facebook/mbart-large-cc25': 10_24, } # fmt: off lowerCAmelCase__ = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class lowerCAmelCase__ ( a): '''simple docstring''' __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = ["input_ids", "attention_mask"] __SCREAMING_SNAKE_CASE = MBartTokenizer __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = [] def __init__( self , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase="<s>" , __lowerCamelCase="</s>" , __lowerCamelCase="</s>" , __lowerCamelCase="<s>" , __lowerCamelCase="<unk>" , __lowerCamelCase="<pad>" , __lowerCamelCase="<mask>" , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , **__lowerCamelCase , ) -> Optional[int]: # Mask token behave like a normal word, i.e. include the space before it _A : List[str] = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase) if isinstance(__lowerCamelCase , __lowerCamelCase) else mask_token super().__init__( vocab_file=__lowerCamelCase , tokenizer_file=__lowerCamelCase , bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , sep_token=__lowerCamelCase , cls_token=__lowerCamelCase , unk_token=__lowerCamelCase , pad_token=__lowerCamelCase , mask_token=__lowerCamelCase , src_lang=__lowerCamelCase , tgt_lang=__lowerCamelCase , additional_special_tokens=__lowerCamelCase , **__lowerCamelCase , ) _A : Union[str, Any] = vocab_file _A : int = False if not self.vocab_file else True _A : Optional[int] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens}) _A : Union[str, Any] = { lang_code: self.convert_tokens_to_ids(__lowerCamelCase) for lang_code in FAIRSEQ_LANGUAGE_CODES } _A : Optional[int] = src_lang if src_lang is not None else "en_XX" _A : Union[str, Any] = self.convert_tokens_to_ids(self._src_lang) _A : int = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def _lowerCamelCase ( self) -> str: return self._src_lang @src_lang.setter def _lowerCamelCase ( self , __lowerCamelCase) -> None: _A : Dict = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = None) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = None) -> List[int]: _A : List[str] = [self.sep_token_id] _A : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , **__lowerCamelCase) -> Dict: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model") _A : str = src_lang _A : Any = self(__lowerCamelCase , add_special_tokens=__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase) _A : Tuple = self.convert_tokens_to_ids(__lowerCamelCase) _A : Dict = tgt_lang_id return inputs def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = "en_XX" , __lowerCamelCase = None , __lowerCamelCase = "ro_RO" , **__lowerCamelCase , ) -> BatchEncoding: _A : Any = src_lang _A : int = tgt_lang return super().prepare_seqaseq_batch(__lowerCamelCase , __lowerCamelCase , **__lowerCamelCase) def _lowerCamelCase ( self) -> List[str]: return self.set_src_lang_special_tokens(self.src_lang) def _lowerCamelCase ( self) -> List[Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang) def _lowerCamelCase ( self , __lowerCamelCase) -> None: _A : int = self.convert_tokens_to_ids(__lowerCamelCase) _A : int = [] _A : List[str] = [self.eos_token_id, self.cur_lang_code] _A : Union[str, Any] = self.convert_ids_to_tokens(self.prefix_tokens) _A : str = self.convert_ids_to_tokens(self.suffix_tokens) _A : List[Any] = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def _lowerCamelCase ( self , __lowerCamelCase) -> None: _A : Optional[int] = self.convert_tokens_to_ids(__lowerCamelCase) _A : List[Any] = [] _A : str = [self.eos_token_id, self.cur_lang_code] _A : Optional[int] = self.convert_ids_to_tokens(self.prefix_tokens) _A : int = self.convert_ids_to_tokens(self.suffix_tokens) _A : str = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase = None) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer.") if not os.path.isdir(__lowerCamelCase): logger.error(F"Vocabulary path ({save_directory}) should be a directory.") return _A : int = os.path.join( __lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(__lowerCamelCase): copyfile(self.vocab_file , __lowerCamelCase) return (out_vocab_file,)
11
0
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): snake_case_ = '''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() snake_case_ = dict(zip(__A , range(len(__A ) ) ) ) snake_case_ = { '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } snake_case_ = { '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 1_60_00, '''return_attention_mask''': False, '''do_normalize''': True, } snake_case_ = tempfile.mkdtemp() snake_case_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) snake_case_ = os.path.join(self.tmpdirname , __A ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__A ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__A ) + '''\n''' ) # load decoder from hub snake_case_ = '''hf-internal-testing/ngram-beam-search-decoder''' def UpperCamelCase__ ( self , **_UpperCAmelCase ): snake_case_ = self.add_kwargs_tokens_map.copy() kwargs.update(__A ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **__A ) def UpperCamelCase__ ( self , **_UpperCAmelCase ): return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **__A ) def UpperCamelCase__ ( self , **_UpperCAmelCase ): return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **__A ) def UpperCamelCase__ ( self ): shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): snake_case_ = self.get_tokenizer() snake_case_ = self.get_feature_extractor() snake_case_ = self.get_decoder() snake_case_ = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A ) processor.save_pretrained(self.tmpdirname ) snake_case_ = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , __A ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , __A ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , __A ) def UpperCamelCase__ ( self ): snake_case_ = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match snake_case_ = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def UpperCamelCase__ ( self ): snake_case_ = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(__A , '''include''' ): WavaVecaProcessorWithLM( tokenizer=__A , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def UpperCamelCase__ ( self ): snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = self.get_decoder() snake_case_ = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A ) snake_case_ = floats_list((3, 10_00) ) snake_case_ = feature_extractor(__A , return_tensors='''np''' ) snake_case_ = processor(__A , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCamelCase__ ( self ): snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = self.get_decoder() snake_case_ = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A ) snake_case_ = '''This is a test string''' snake_case_ = processor(text=__A ) snake_case_ = tokenizer(__A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase__ ( self , _UpperCAmelCase=(2, 10, 16) , _UpperCAmelCase=77 ): np.random.seed(__A ) return np.random.rand(*__A ) def UpperCamelCase__ ( self ): snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = self.get_decoder() snake_case_ = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A ) snake_case_ = self._get_dummy_logits(shape=(10, 16) , seed=13 ) snake_case_ = processor.decode(__A ) snake_case_ = decoder.decode_beams(__A )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def UpperCamelCase__ ( self , _UpperCAmelCase ): snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = self.get_decoder() snake_case_ = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A ) snake_case_ = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: snake_case_ = processor.batch_decode(__A ) else: with get_context(__A ).Pool() as pool: snake_case_ = processor.batch_decode(__A , __A ) snake_case_ = list(__A ) with get_context('''fork''' ).Pool() as p: snake_case_ = decoder.decode_beams_batch(__A , __A ) snake_case_ , snake_case_ , snake_case_ = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(__A , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(__A , decoded_processor.logit_score ) self.assertListEqual(__A , decoded_processor.lm_score ) def UpperCamelCase__ ( self ): snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = self.get_decoder() snake_case_ = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A ) snake_case_ = self._get_dummy_logits() snake_case_ = 15 snake_case_ = -20.0 snake_case_ = -4.0 snake_case_ = processor.batch_decode( __A , beam_width=__A , beam_prune_logp=__A , token_min_logp=__A , ) snake_case_ = decoded_processor_out.text snake_case_ = list(__A ) with get_context('''fork''' ).Pool() as pool: snake_case_ = decoder.decode_beams_batch( __A , __A , beam_width=__A , beam_prune_logp=__A , token_min_logp=__A , ) snake_case_ = [d[0][0] for d in decoded_decoder_out] snake_case_ = [d[0][2] for d in decoded_decoder_out] snake_case_ = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(__A , __A ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , __A ) self.assertTrue(np.array_equal(__A , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , __A , atol=1E-3 ) ) self.assertTrue(np.array_equal(__A , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9_474] , __A , atol=1E-3 ) ) def UpperCamelCase__ ( self ): snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = self.get_decoder() snake_case_ = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A ) snake_case_ = self._get_dummy_logits() snake_case_ = 2.0 snake_case_ = 5.0 snake_case_ = -20.0 snake_case_ = True snake_case_ = processor.batch_decode( __A , alpha=__A , beta=__A , unk_score_offset=__A , lm_score_boundary=__A , ) snake_case_ = decoded_processor_out.text snake_case_ = list(__A ) decoder.reset_params( alpha=__A , beta=__A , unk_score_offset=__A , lm_score_boundary=__A , ) with get_context('''fork''' ).Pool() as pool: snake_case_ = decoder.decode_beams_batch( __A , __A , ) snake_case_ = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(__A , __A ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , __A ) snake_case_ = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , __A ) def UpperCamelCase__ ( self ): snake_case_ = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) snake_case_ = processor.decoder.model_container[processor.decoder._model_key] snake_case_ = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() snake_case_ = os.listdir(__A ) snake_case_ = ['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(__A , __A ) def UpperCamelCase__ ( self ): snake_case_ = snapshot_download('''hf-internal-testing/processor_with_lm''' ) snake_case_ = WavaVecaProcessorWithLM.from_pretrained(__A ) snake_case_ = processor.decoder.model_container[processor.decoder._model_key] snake_case_ = Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() snake_case_ = os.listdir(__A ) snake_case_ = os.listdir(__A ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(__A , __A ) def UpperCamelCase__ ( self ): snake_case_ = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) snake_case_ = AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) snake_case_ = floats_list((3, 10_00) ) snake_case_ = processor_wavaveca(__A , return_tensors='''np''' ) snake_case_ = processor_auto(__A , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) snake_case_ = self._get_dummy_logits() snake_case_ = processor_wavaveca.batch_decode(__A ) snake_case_ = processor_auto.batch_decode(__A ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def UpperCamelCase__ ( self ): snake_case_ = self.get_feature_extractor() snake_case_ = self.get_tokenizer() snake_case_ = self.get_decoder() snake_case_ = WavaVecaProcessorWithLM(tokenizer=__A , feature_extractor=__A , decoder=__A ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def UpperCamelCase__ ( _UpperCAmelCase , _UpperCAmelCase ): snake_case_ = [d[key] for d in offsets] return retrieved_list def UpperCamelCase__ ( self ): snake_case_ = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) snake_case_ = self._get_dummy_logits()[0] snake_case_ = processor.decode(__A , output_word_offsets=__A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(__A , __A ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def UpperCamelCase__ ( self ): snake_case_ = WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) snake_case_ = self._get_dummy_logits() snake_case_ = processor.batch_decode(__A , output_word_offsets=__A ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(__A , __A ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(__A , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def UpperCamelCase__ ( self ): import torch snake_case_ = load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=__A ) snake_case_ = ds.cast_column('''audio''' , datasets.Audio(sampling_rate=1_60_00 ) ) snake_case_ = iter(__A ) snake_case_ = next(__A ) snake_case_ = AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) snake_case_ = WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train snake_case_ = processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): snake_case_ = model(__A ).logits.cpu().numpy() snake_case_ = processor.decode(logits[0] , output_word_offsets=__A ) snake_case_ = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate snake_case_ = [ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] snake_case_ = '''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(__A , '''word''' ) ) , __A ) self.assertEqual(''' '''.join(self.get_from_offsets(__A , '''word''' ) ) , output.text ) # output times snake_case_ = torch.tensor(self.get_from_offsets(__A , '''start_time''' ) ) snake_case_ = torch.tensor(self.get_from_offsets(__A , '''end_time''' ) ) # fmt: off snake_case_ = torch.tensor([1.4_199, 1.6_599, 2.2_599, 3.0, 3.24, 3.5_999, 3.7_999, 4.0_999, 4.26, 4.94, 5.28, 5.6_599, 5.78, 5.94, 6.32, 6.5_399, 6.6_599] ) snake_case_ = torch.tensor([1.5_399, 1.8_999, 2.9, 3.16, 3.5_399, 3.72, 4.0_199, 4.1_799, 4.76, 5.1_599, 5.5_599, 5.6_999, 5.86, 6.1_999, 6.38, 6.6_199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(__A , __A , atol=0.01 ) ) self.assertTrue(torch.allclose(__A , __A , atol=0.01 ) )
359
import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class lowerCAmelCase_ ( lowerCamelCase__ ): '''simple docstring''' def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): super().__init__() snake_case_ = value_function snake_case_ = unet snake_case_ = scheduler snake_case_ = env snake_case_ = env.get_dataset() snake_case_ = {} for key in self.data.keys(): try: snake_case_ = self.data[key].mean() except: # noqa: E722 pass snake_case_ = {} for key in self.data.keys(): try: snake_case_ = self.data[key].std() except: # noqa: E722 pass snake_case_ = env.observation_space.shape[0] snake_case_ = env.action_space.shape[0] def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase ): return (x_in - self.means[key]) / self.stds[key] def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase ): return x_in * self.stds[key] + self.means[key] def UpperCamelCase__ ( self , _UpperCAmelCase ): if type(_UpperCAmelCase ) is dict: return {k: self.to_torch(_UpperCAmelCase ) for k, v in x_in.items()} elif torch.is_tensor(_UpperCAmelCase ): return x_in.to(self.unet.device ) return torch.tensor(_UpperCAmelCase , device=self.unet.device ) def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): for key, val in cond.items(): snake_case_ = val.clone() return x_in def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): snake_case_ = x.shape[0] snake_case_ = None for i in tqdm.tqdm(self.scheduler.timesteps ): # create batch of timesteps to pass into model snake_case_ = torch.full((batch_size,) , _UpperCAmelCase , device=self.unet.device , dtype=torch.long ) for _ in range(_UpperCAmelCase ): with torch.enable_grad(): x.requires_grad_() # permute to match dimension for pre-trained models snake_case_ = self.value_function(x.permute(0 , 2 , 1 ) , _UpperCAmelCase ).sample snake_case_ = torch.autograd.grad([y.sum()] , [x] )[0] snake_case_ = self.scheduler._get_variance(_UpperCAmelCase ) snake_case_ = torch.exp(0.5 * posterior_variance ) snake_case_ = model_std * grad snake_case_ = 0 snake_case_ = x.detach() snake_case_ = x + scale * grad snake_case_ = self.reset_xa(_UpperCAmelCase , _UpperCAmelCase , self.action_dim ) snake_case_ = self.unet(x.permute(0 , 2 , 1 ) , _UpperCAmelCase ).sample.permute(0 , 2 , 1 ) # TODO: verify deprecation of this kwarg snake_case_ = self.scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , predict_epsilon=_UpperCAmelCase )['''prev_sample'''] # apply conditions to the trajectory (set the initial state) snake_case_ = self.reset_xa(_UpperCAmelCase , _UpperCAmelCase , self.action_dim ) snake_case_ = self.to_torch(_UpperCAmelCase ) return x, y def __call__( self , _UpperCAmelCase , _UpperCAmelCase=64 , _UpperCAmelCase=32 , _UpperCAmelCase=2 , _UpperCAmelCase=0.1 ): # normalize the observations and create batch dimension snake_case_ = self.normalize(_UpperCAmelCase , '''observations''' ) snake_case_ = obs[None].repeat(_UpperCAmelCase , axis=0 ) snake_case_ = {0: self.to_torch(_UpperCAmelCase )} snake_case_ = (batch_size, planning_horizon, self.state_dim + self.action_dim) # generate initial noise and apply our conditions (to make the trajectories start at current state) snake_case_ = randn_tensor(_UpperCAmelCase , device=self.unet.device ) snake_case_ = self.reset_xa(_UpperCAmelCase , _UpperCAmelCase , self.action_dim ) snake_case_ = self.to_torch(_UpperCAmelCase ) # run the diffusion process snake_case_ , snake_case_ = self.run_diffusion(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # sort output trajectories by value snake_case_ = y.argsort(0 , descending=_UpperCAmelCase ).squeeze() snake_case_ = x[sorted_idx] snake_case_ = sorted_values[:, :, : self.action_dim] snake_case_ = actions.detach().cpu().numpy() snake_case_ = self.de_normalize(_UpperCAmelCase , key='''actions''' ) # select the action with the highest value if y is not None: snake_case_ = 0 else: # if we didn't run value guiding, select a random action snake_case_ = np.random.randint(0 , _UpperCAmelCase ) snake_case_ = denorm_actions[selected_index, 0] return denorm_actions
267
0
import sys from collections import defaultdict class _snake_case : def __init__( self ): __magic_name__ : Union[str, Any] = [] def SCREAMING_SNAKE_CASE ( self , _a ): return self.node_position[vertex] def SCREAMING_SNAKE_CASE ( self , _a , _a ): __magic_name__ : str = pos def SCREAMING_SNAKE_CASE ( self , _a , _a , _a , _a ): if start > size // 2 - 1: return else: if 2 * start + 2 >= size: __magic_name__ : Any = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: __magic_name__ : Optional[Any] = 2 * start + 1 else: __magic_name__ : Dict = 2 * start + 2 if heap[smallest_child] < heap[start]: __magic_name__ , __magic_name__ : Any = heap[smallest_child], positions[smallest_child] __magic_name__ , __magic_name__ : Dict = ( heap[start], positions[start], ) __magic_name__ , __magic_name__ : Union[str, Any] = temp, tempa __magic_name__ : Optional[Any] = self.get_position(positions[smallest_child] ) self.set_position( positions[smallest_child] , self.get_position(positions[start] ) ) self.set_position(positions[start] , _a ) self.top_to_bottom(_a , _a , _a , _a ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a , _a ): __magic_name__ : int = position[index] while index != 0: __magic_name__ : Tuple = int((index - 2) / 2 ) if index % 2 == 0 else int((index - 1) / 2 ) if val < heap[parent]: __magic_name__ : Optional[int] = heap[parent] __magic_name__ : List[str] = position[parent] self.set_position(position[parent] , _a ) else: __magic_name__ : Dict = val __magic_name__ : Optional[Any] = temp self.set_position(_a , _a ) break __magic_name__ : List[Any] = parent else: __magic_name__ : Union[str, Any] = val __magic_name__ : Optional[int] = temp self.set_position(_a , 0 ) def SCREAMING_SNAKE_CASE ( self , _a , _a ): __magic_name__ : Tuple = len(_a ) // 2 - 1 for i in range(_a , -1 , -1 ): self.top_to_bottom(_a , _a , len(_a ) , _a ) def SCREAMING_SNAKE_CASE ( self , _a , _a ): __magic_name__ : List[Any] = positions[0] __magic_name__ : List[str] = sys.maxsize self.top_to_bottom(_a , 0 , len(_a ) , _a ) return temp def lowerCAmelCase_ ( _snake_case : Optional[Any] ) -> Tuple: '''simple docstring''' __magic_name__ : Union[str, Any] = Heap() __magic_name__ : str = [0] * len(_snake_case ) __magic_name__ : Optional[Any] = [-1] * len(_snake_case ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph __magic_name__ : Optional[Any] = [] # Heap of Distance of vertices from their neighboring vertex __magic_name__ : List[Any] = [] for vertex in range(len(_snake_case ) ): distance_tv.append(sys.maxsize ) positions.append(_snake_case ) heap.node_position.append(_snake_case ) __magic_name__ : Any = [] __magic_name__ : int = 1 __magic_name__ : Tuple = sys.maxsize for neighbor, distance in adjacency_list[0]: __magic_name__ : List[str] = 0 __magic_name__ : Union[str, Any] = distance heap.heapify(_snake_case , _snake_case ) for _ in range(1 , len(_snake_case ) ): __magic_name__ : str = heap.delete_minimum(_snake_case , _snake_case ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) __magic_name__ : int = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(_snake_case )] ): __magic_name__ : str = distance heap.bottom_to_top( _snake_case , heap.get_position(_snake_case ) , _snake_case , _snake_case ) __magic_name__ : List[Any] = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > snake_case : Any = int(input("Enter number of edges: ").strip()) snake_case : List[Any] = defaultdict(list) for _ in range(edges_number): snake_case : List[str] = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
281
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
281
1
'''simple docstring''' def __UpperCamelCase ( UpperCAmelCase ): lowercase__ : str = 1 lowercase__ : Any = 2 while i * i <= n: lowercase__ : Optional[Any] = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def __UpperCamelCase ( ): lowercase__ : Dict = 1 lowercase__ : Union[str, Any] = 1 while True: i += 1 t_num += i if count_divisors(UpperCAmelCase ) > 500: break return t_num if __name__ == "__main__": print(solution())
214
'''simple docstring''' import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __a: Union[str, Any] = logging.get_logger(__name__) __a: Tuple = {"""tokenizer_file""": """tokenizer.json"""} __a: Union[str, Any] = { """tokenizer_file""": { """bigscience/tokenizer""": """https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json""", """bigscience/bloom-560m""": """https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json""", """bigscience/bloom-1b1""": """https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json""", """bigscience/bloom-1b7""": """https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json""", """bigscience/bloom-3b""": """https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json""", """bigscience/bloom-7b1""": """https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json""", """bigscience/bloom""": """https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json""", }, } class UpperCAmelCase ( a__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE = None def __init__( self , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase="<unk>" , __lowerCAmelCase="<s>" , __lowerCAmelCase="</s>" , __lowerCAmelCase="<pad>" , __lowerCAmelCase=False , __lowerCAmelCase=False , **__lowerCAmelCase , ) -> Union[str, Any]: super().__init__( __lowerCAmelCase , __lowerCAmelCase , tokenizer_file=__lowerCAmelCase , unk_token=__lowerCAmelCase , bos_token=__lowerCAmelCase , eos_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , add_prefix_space=__lowerCAmelCase , clean_up_tokenization_spaces=__lowerCAmelCase , **__lowerCAmelCase , ) lowercase__ : int = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , __lowerCAmelCase ) != add_prefix_space: lowercase__ : int = getattr(__lowerCAmelCase , pre_tok_state.pop('''type''' ) ) lowercase__ : Tuple = add_prefix_space lowercase__ : List[str] = pre_tok_class(**__lowerCAmelCase ) lowercase__ : Union[str, Any] = add_prefix_space def _lowerCAmelCase( self , *__lowerCAmelCase , **__lowerCAmelCase ) -> BatchEncoding: lowercase__ : Dict = kwargs.get('''is_split_into_words''' , __lowerCAmelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''' ) return super()._batch_encode_plus(*__lowerCAmelCase , **__lowerCAmelCase ) def _lowerCAmelCase( self , *__lowerCAmelCase , **__lowerCAmelCase ) -> BatchEncoding: lowercase__ : str = kwargs.get('''is_split_into_words''' , __lowerCAmelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''' ) return super()._encode_plus(*__lowerCAmelCase , **__lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase = None ) -> Tuple[str]: lowercase__ : List[Any] = self._tokenizer.model.save(__lowerCAmelCase , name=__lowerCAmelCase ) return tuple(__lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> List[int]: lowercase__ : Dict = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase ) + [self.eos_token_id] ) if len(__lowerCAmelCase ) > self.model_max_length: lowercase__ : Optional[Any] = input_ids[-self.model_max_length :] return input_ids
214
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCAmelCase__ : int = {'configuration_glpn': ['GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GLPNConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : List[str] = ['GLPNFeatureExtractor'] lowerCAmelCase__ : Any = ['GLPNImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[int] = [ 'GLPN_PRETRAINED_MODEL_ARCHIVE_LIST', 'GLPNForDepthEstimation', 'GLPNLayer', 'GLPNModel', 'GLPNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys lowerCAmelCase__ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
98
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[int] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Optional[int]=7 , lowerCAmelCase__ : List[Any]=3 , lowerCAmelCase__ : Optional[Any]=18 , lowerCAmelCase__ : Union[str, Any]=30 , lowerCAmelCase__ : Any=400 , lowerCAmelCase__ : Any=True , lowerCAmelCase__ : Tuple=None , lowerCAmelCase__ : str=True , lowerCAmelCase__ : List[str]=None , lowerCAmelCase__ : Optional[Any]=True , lowerCAmelCase__ : str=[0.5, 0.5, 0.5] , lowerCAmelCase__ : int=[0.5, 0.5, 0.5] , ) -> List[str]: '''simple docstring''' _UpperCamelCase = size if size is not None else {'''shortest_edge''': 18} _UpperCamelCase = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = num_channels _UpperCamelCase = image_size _UpperCamelCase = min_resolution _UpperCamelCase = max_resolution _UpperCamelCase = do_resize _UpperCamelCase = size _UpperCamelCase = do_center_crop _UpperCamelCase = crop_size _UpperCamelCase = do_normalize _UpperCamelCase = image_mean _UpperCamelCase = image_std def snake_case__ ( self : Union[str, Any] ) -> List[Any]: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __lowerCAmelCase ( __magic_name__ , unittest.TestCase ): """simple docstring""" _snake_case : Tuple = LevitImageProcessor if is_vision_available() else None def snake_case__ ( self : int ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = LevitImageProcessingTester(self ) @property def snake_case__ ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self : Tuple ) -> List[Any]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase__ , '''image_mean''' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , '''image_std''' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , '''do_normalize''' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , '''do_center_crop''' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , '''size''' ) ) def snake_case__ ( self : str ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) _UpperCamelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) def snake_case__ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' pass def snake_case__ ( self : Dict ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(lowerCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def snake_case__ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , numpify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , np.ndarray ) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(lowerCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def snake_case__ ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _UpperCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , torchify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , torch.Tensor ) # Test not batched input _UpperCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _UpperCamelCase = image_processing(lowerCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
324
0
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available if is_sentencepiece_available(): from transformers.models.marian.tokenization_marian import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin __UpperCamelCase : Optional[int] = get_tests_dir("fixtures/test_sentencepiece.model") __UpperCamelCase : Any = {"target_lang": "fi", "source_lang": "en"} __UpperCamelCase : Optional[int] = ">>zh<<" __UpperCamelCase : List[Any] = "Helsinki-NLP/" if is_torch_available(): __UpperCamelCase : str = "pt" elif is_tf_available(): __UpperCamelCase : Union[str, Any] = "tf" else: __UpperCamelCase : int = "jax" @require_sentencepiece class __lowerCAmelCase ( __magic_name__ , unittest.TestCase ): UpperCamelCase__ = MarianTokenizer UpperCamelCase__ = False UpperCamelCase__ = True def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' super().setUp() a = ["""</s>""", """<unk>""", """▁This""", """▁is""", """▁a""", """▁t""", """est""", """\u0120""", """<pad>"""] a = dict(zip(__magic_name__ , range(len(__magic_name__ ) ) ) ) a = Path(self.tmpdirname ) save_json(__magic_name__ , save_dir / VOCAB_FILES_NAMES["""vocab"""] ) save_json(__magic_name__ , save_dir / VOCAB_FILES_NAMES["""tokenizer_config_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["source_spm"]).exists(): copyfile(__magic_name__ , save_dir / VOCAB_FILES_NAMES["""source_spm"""] ) copyfile(__magic_name__ , save_dir / VOCAB_FILES_NAMES["""target_spm"""] ) a = MarianTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase__ ( self :Any , **__magic_name__ :str ): '''simple docstring''' return MarianTokenizer.from_pretrained(self.tmpdirname , **__magic_name__ ) def lowerCamelCase__ ( self :Dict , __magic_name__ :List[Any] ): '''simple docstring''' return ( "This is a test", "This is a test", ) def lowerCamelCase__ ( self :int ): '''simple docstring''' a = """</s>""" a = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__magic_name__ ) , __magic_name__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__magic_name__ ) , __magic_name__ ) def lowerCamelCase__ ( self :Dict ): '''simple docstring''' a = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """</s>""" ) self.assertEqual(vocab_keys[1] , """<unk>""" ) self.assertEqual(vocab_keys[-1] , """<pad>""" ) self.assertEqual(len(__magic_name__ ) , 9 ) def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 9 ) def lowerCamelCase__ ( self :int ): '''simple docstring''' a = MarianTokenizer.from_pretrained(F'{ORG_NAME}opus-mt-en-de' ) a = en_de_tokenizer(["""I am a small frog"""] , return_tensors=__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) a = [38, 121, 14, 697, 3_8848, 0] self.assertListEqual(__magic_name__ , batch.input_ids[0] ) a = tempfile.mkdtemp() en_de_tokenizer.save_pretrained(__magic_name__ ) a = [x.name for x in Path(__magic_name__ ).glob("""*""" )] self.assertIn("""source.spm""" , __magic_name__ ) MarianTokenizer.from_pretrained(__magic_name__ ) def lowerCamelCase__ ( self :Optional[int] ): '''simple docstring''' a = self.get_tokenizer() a = tok( ["""I am a small frog""" * 1000, """I am a small frog"""] , padding=__magic_name__ , truncation=__magic_name__ , return_tensors=__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) self.assertEqual(batch.input_ids.shape , (2, 512) ) def lowerCamelCase__ ( self :Dict ): '''simple docstring''' a = self.get_tokenizer() a = tok(["""I am a tiny frog""", """I am a small frog"""] , padding=__magic_name__ , return_tensors=__magic_name__ ) self.assertIsInstance(__magic_name__ , __magic_name__ ) self.assertEqual(batch_smaller.input_ids.shape , (2, 10) ) @slow def lowerCamelCase__ ( self :Optional[int] ): '''simple docstring''' a = {"""input_ids""": [[4_3495, 462, 20, 4_2164, 1369, 52, 464, 132, 1703, 492, 13, 7491, 3_8999, 6, 8, 464, 132, 1703, 492, 13, 4669, 3_7867, 13, 7525, 27, 1593, 988, 13, 3_3972, 7029, 6, 20, 8251, 383, 2, 270, 5866, 3788, 2, 2353, 8251, 1_2338, 2, 1_3958, 387, 2, 3629, 6953, 188, 2900, 2, 1_3958, 8011, 1_1501, 23, 8460, 4073, 3_4009, 20, 435, 1_1439, 27, 8, 8460, 4073, 6004, 20, 9988, 375, 27, 33, 266, 1945, 1076, 1350, 3_7867, 3288, 5, 577, 1076, 4374, 8, 5082, 5, 2_6453, 257, 556, 403, 2, 242, 132, 383, 316, 492, 8, 1_0767, 6, 316, 304, 4239, 3, 0], [148, 1_5722, 19, 1839, 12, 1350, 13, 2_2327, 5082, 5418, 4_7567, 3_5938, 59, 318, 1_9552, 108, 2183, 54, 1_4976, 4835, 32, 547, 1114, 8, 315, 2417, 5, 92, 1_9088, 3, 0, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100], [36, 6395, 1_2570, 3_9147, 1_1597, 6, 266, 4, 4_5405, 7296, 3, 0, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100, 5_8100]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__magic_name__ , model_name="""Helsinki-NLP/opus-mt-en-de""" , revision="""1a8c2263da11e68e50938f97e10cd57820bd504c""" , decode_kwargs={"""use_source_tokenizer""": True} , ) def lowerCamelCase__ ( self :Any ): '''simple docstring''' a = MarianTokenizer.from_pretrained("""hf-internal-testing/test-marian-two-vocabs""" ) a = """Tämä on testi""" a = """This is a test""" a = [76, 7, 2047, 2] a = [69, 12, 11, 940, 2] a = tokenizer(__magic_name__ ).input_ids self.assertListEqual(__magic_name__ , __magic_name__ ) a = tokenizer(text_target=__magic_name__ ).input_ids self.assertListEqual(__magic_name__ , __magic_name__ ) a = tokenizer.decode(__magic_name__ , skip_special_tokens=__magic_name__ ) self.assertEqual(__magic_name__ , __magic_name__ )
361
import json import os import shutil import tempfile import unittest from transformers import BatchEncoding, CanineTokenizer from transformers.testing_utils import require_tokenizers, require_torch from transformers.tokenization_utils import AddedToken from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin class __lowerCAmelCase ( __magic_name__ , unittest.TestCase ): UpperCamelCase__ = CanineTokenizer UpperCamelCase__ = False def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' super().setUp() a = CanineTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCamelCase__ ( self :Dict ): '''simple docstring''' return CanineTokenizer.from_pretrained("""google/canine-s""" ) def lowerCamelCase__ ( self :Tuple , **__magic_name__ :Dict ): '''simple docstring''' a = self.tokenizer_class.from_pretrained(self.tmpdirname , **__magic_name__ ) a = 1024 return tokenizer @require_torch def lowerCamelCase__ ( self :int ): '''simple docstring''' a = self.canine_tokenizer a = ["""Life is like a box of chocolates.""", """You never know what you're gonna get."""] # fmt: off a = [5_7344, 76, 105, 102, 101, 32, 105, 115, 32, 108, 105, 107, 101, 32, 97, 32, 98, 111, 120, 32, 111, 102, 32, 99, 104, 111, 99, 111, 108, 97, 116, 101, 115, 46, 5_7345, 0, 0, 0, 0] # fmt: on a = tokenizer(__magic_name__ , padding=__magic_name__ , return_tensors="""pt""" ) self.assertIsInstance(__magic_name__ , __magic_name__ ) a = list(batch.input_ids.numpy()[0] ) self.assertListEqual(__magic_name__ , __magic_name__ ) self.assertEqual((2, 39) , batch.input_ids.shape ) self.assertEqual((2, 39) , batch.attention_mask.shape ) @require_torch def lowerCamelCase__ ( self :Any ): '''simple docstring''' a = self.canine_tokenizer a = ["""Once there was a man.""", """He wrote a test in HuggingFace Tranformers."""] a = tokenizer(__magic_name__ , padding=__magic_name__ , return_tensors="""pt""" ) # check if input_ids, attention_mask and token_type_ids are returned self.assertIn("""input_ids""" , __magic_name__ ) self.assertIn("""attention_mask""" , __magic_name__ ) self.assertIn("""token_type_ids""" , __magic_name__ ) @require_torch def lowerCamelCase__ ( self :Any ): '''simple docstring''' a = self.canine_tokenizer a = [ """What's the weater?""", """It's about 25 degrees.""", ] a = tokenizer( text_target=__magic_name__ , max_length=32 , padding="""max_length""" , truncation=__magic_name__ , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' a = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test a = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): # Isolate this from the other tests because we save additional tokens/etc a = tempfile.mkdtemp() a = """ He is very happy, UNwant\u00E9d,running""" a = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) tokenizer.save_pretrained(__magic_name__ ) a = tokenizer.__class__.from_pretrained(__magic_name__ ) a = after_tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) shutil.rmtree(__magic_name__ ) a = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): # Isolate this from the other tests because we save additional tokens/etc a = tempfile.mkdtemp() a = """ He is very happy, UNwant\u00E9d,running""" a = tokenizer.additional_special_tokens # We can add a new special token for Canine as follows: a = chr(0Xe_0_0_7 ) additional_special_tokens.append(__magic_name__ ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) a = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) tokenizer.save_pretrained(__magic_name__ ) a = tokenizer.__class__.from_pretrained(__magic_name__ ) a = after_tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) self.assertIn(__magic_name__ , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) a = tokenizer.__class__.from_pretrained(__magic_name__ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(__magic_name__ ) def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' a = self.get_tokenizers(do_lower_case=__magic_name__ ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): a , a = self.get_clean_sequence(__magic_name__ ) # a special token for Canine can be defined as follows: a = 0Xe_0_0_5 a = chr(__magic_name__ ) tokenizer.add_special_tokens({"""cls_token""": special_token} ) a = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) self.assertEqual(len(__magic_name__ ) , 1 ) a = tokenizer.decode(ids + encoded_special_token , clean_up_tokenization_spaces=__magic_name__ ) a = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) a = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) a = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) self.assertEqual(__magic_name__ , input_encoded + special_token_id ) a = tokenizer.decode(__magic_name__ , skip_special_tokens=__magic_name__ ) self.assertTrue(special_token not in decoded ) def lowerCamelCase__ ( self :int ): '''simple docstring''' a = self.get_tokenizers(do_lower_case=__magic_name__ ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): a = chr(0Xe_0_0_5 ) a = chr(0Xe_0_0_6 ) # `add_tokens` method stores special tokens only in `tokenizer.unique_no_split_tokens`. (in tokenization_utils.py) tokenizer.add_tokens([SPECIAL_TOKEN_1] , special_tokens=__magic_name__ ) # `add_special_tokens` method stores special tokens in `tokenizer.additional_special_tokens`, # which also occur in `tokenizer.all_special_tokens`. (in tokenization_utils_base.py) tokenizer.add_special_tokens({"""additional_special_tokens""": [SPECIAL_TOKEN_2]} ) a = tokenizer.tokenize(__magic_name__ ) a = tokenizer.tokenize(__magic_name__ ) self.assertEqual(len(__magic_name__ ) , 1 ) self.assertEqual(len(__magic_name__ ) , 1 ) self.assertEqual(token_a[0] , __magic_name__ ) self.assertEqual(token_a[0] , __magic_name__ ) @require_tokenizers def lowerCamelCase__ ( self :Dict ): '''simple docstring''' a = self.get_tokenizers(do_lower_case=__magic_name__ ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): # a special token for Canine can be defined as follows: a = 0Xe_0_0_6 a = chr(__magic_name__ ) a = AddedToken(__magic_name__ , lstrip=__magic_name__ ) tokenizer.add_special_tokens({"""additional_special_tokens""": [new_token]} ) with tempfile.TemporaryDirectory() as tmp_dir_name: tokenizer.save_pretrained(__magic_name__ ) tokenizer.from_pretrained(__magic_name__ ) def lowerCamelCase__ ( self :int ): '''simple docstring''' a = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__magic_name__ ) with open(os.path.join(__magic_name__ , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: a = json.load(__magic_name__ ) with open(os.path.join(__magic_name__ , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: a = json.load(__magic_name__ ) # a special token for Canine can be defined as follows: a = 0Xe_0_0_6 a = chr(__magic_name__ ) a = [new_token_a] a = [new_token_a] with open(os.path.join(__magic_name__ , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(__magic_name__ , __magic_name__ ) with open(os.path.join(__magic_name__ , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(__magic_name__ , __magic_name__ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files a = tokenizer_class.from_pretrained(__magic_name__ , extra_ids=0 ) self.assertIn(__magic_name__ , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids([new_token_a] ) ) , ) a = 0Xe_0_0_7 a = chr(__magic_name__ ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained a = [AddedToken(__magic_name__ , lstrip=__magic_name__ )] a = tokenizer_class.from_pretrained( __magic_name__ , additional_special_tokens=__magic_name__ , extra_ids=0 ) self.assertIn(__magic_name__ , tokenizer.additional_special_tokens ) # self.assertIn(new_token_2,tokenizer.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer.convert_ids_to_tokens(tokenizer.convert_tokens_to_ids([new_token_a] ) ) ) @require_tokenizers def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' a = self.get_tokenizers(do_lower_case=__magic_name__ ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): a = """hello world""" if self.space_between_special_tokens: a = """[CLS] hello world [SEP]""" else: a = input a = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) a = tokenizer.decode(__magic_name__ , spaces_between_special_tokens=self.space_between_special_tokens ) self.assertIn(__magic_name__ , [output, output.lower()] ) def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' a = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): a = [ """bos_token""", """eos_token""", """unk_token""", """sep_token""", """pad_token""", """cls_token""", """mask_token""", ] a = """a""" a = ord(__magic_name__ ) for attr in attributes_list: setattr(__magic_name__ , attr + """_id""" , __magic_name__ ) self.assertEqual(getattr(__magic_name__ , __magic_name__ ) , __magic_name__ ) self.assertEqual(getattr(__magic_name__ , attr + """_id""" ) , __magic_name__ ) setattr(__magic_name__ , attr + """_id""" , __magic_name__ ) self.assertEqual(getattr(__magic_name__ , __magic_name__ ) , __magic_name__ ) self.assertEqual(getattr(__magic_name__ , attr + """_id""" ) , __magic_name__ ) setattr(__magic_name__ , """additional_special_tokens_ids""" , [] ) self.assertListEqual(getattr(__magic_name__ , """additional_special_tokens""" ) , [] ) self.assertListEqual(getattr(__magic_name__ , """additional_special_tokens_ids""" ) , [] ) a = 0Xe_0_0_6 a = chr(__magic_name__ ) setattr(__magic_name__ , """additional_special_tokens_ids""" , [additional_special_token_id] ) self.assertListEqual(getattr(__magic_name__ , """additional_special_tokens""" ) , [additional_special_token] ) self.assertListEqual(getattr(__magic_name__ , """additional_special_tokens_ids""" ) , [additional_special_token_id] ) def lowerCamelCase__ ( self :Any ): '''simple docstring''' pass def lowerCamelCase__ ( self :str ): '''simple docstring''' pass def lowerCamelCase__ ( self :Optional[int] ): '''simple docstring''' pass def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' pass def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' pass def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' pass def lowerCamelCase__ ( self :Any ): '''simple docstring''' pass def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' pass
347
0
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class a ( _lowerCamelCase , unittest.TestCase ): snake_case_ = BarthezTokenizer snake_case_ = BarthezTokenizerFast snake_case_ = True snake_case_ = True def A_ ( self : List[str] ): super().setUp() snake_case_ = BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=lowercase_ ) snake_case_ = tokenizer def A_ ( self : Optional[Any] ): snake_case_ = '''<pad>''' snake_case_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase_ ) , lowercase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase_ ) , lowercase_ ) def A_ ( self : List[Any] ): snake_case_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(lowercase_ ) , 10_1122 ) def A_ ( self : int ): self.assertEqual(self.get_tokenizer().vocab_size , 10_1122 ) @require_torch def A_ ( self : str ): snake_case_ = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] snake_case_ = [0, 57, 3018, 7_0307, 91, 2] snake_case_ = self.tokenizer( lowercase_ , max_length=len(lowercase_ ) , padding=lowercase_ , truncation=lowercase_ , return_tensors='''pt''' ) self.assertIsInstance(lowercase_ , lowercase_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) snake_case_ = batch.input_ids.tolist()[0] self.assertListEqual(lowercase_ , lowercase_ ) def A_ ( self : Tuple ): if not self.test_rust_tokenizer: return snake_case_ = self.get_tokenizer() snake_case_ = self.get_rust_tokenizer() snake_case_ = '''I was born in 92000, and this is falsé.''' snake_case_ = tokenizer.tokenize(lowercase_ ) snake_case_ = rust_tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) snake_case_ = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) snake_case_ = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) snake_case_ = self.get_rust_tokenizer() snake_case_ = tokenizer.encode(lowercase_ ) snake_case_ = rust_tokenizer.encode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) @slow def A_ ( self : List[Any] ): # fmt: off snake_case_ = {'''input_ids''': [[0, 490, 1_4328, 4507, 354, 47, 4_3669, 95, 25, 7_8117, 2_0215, 1_9779, 190, 22, 400, 4, 3_5343, 8_0310, 603, 86, 2_4937, 105, 3_3438, 9_4762, 196, 3_9642, 7, 15, 1_5933, 173, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0534, 87, 25, 66, 3358, 196, 5_5289, 8, 8_2961, 81, 2204, 7_5203, 7, 15, 763, 1_2956, 216, 178, 1_4328, 9595, 1377, 6_9693, 7, 448, 7_1021, 196, 1_8106, 1437, 1_3974, 108, 9083, 4, 4_9315, 7, 39, 86, 1326, 2793, 4_6333, 4, 448, 196, 7_4588, 7, 4_9315, 7, 39, 21, 822, 3_8470, 74, 21, 6_6723, 6_2480, 8, 2_2050, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. snake_case_ = [ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=lowercase_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=lowercase_ , )
56
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE :List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE :List[str] = {'vocab_file': 'spiece.model'} SCREAMING_SNAKE_CASE :Dict = { 'vocab_file': { 'AI-Sweden/gpt-sw3-126m': 'https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model', 'AI-Sweden/gpt-sw3-350m': 'https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model', 'AI-Sweden/gpt-sw3-1.6b': 'https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model', 'AI-Sweden/gpt-sw3-6.7b': 'https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model', 'AI-Sweden/gpt-sw3-20b': 'https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model', } } SCREAMING_SNAKE_CASE :Optional[Any] = { 'AI-Sweden/gpt-sw3-126m': 2048, 'AI-Sweden/gpt-sw3-350m': 2048, 'AI-Sweden/gpt-sw3-1.6b': 2048, 'AI-Sweden/gpt-sw3-6.7b': 2048, 'AI-Sweden/gpt-sw3-20b': 2048, } class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = ["input_ids", "attention_mask"] def __init__( self : Optional[int] ,A : Optional[Any] ,A : Optional[int]=False ,A : int=False ,A : Union[str, Any]=False ,A : int=None ,A : Optional[Any]=None ,A : Union[str, Any]=None ,A : Optional[Any]=None ,A : Optional[Dict[str, Any]] = None ,**A : Tuple ,): __A = {} if sp_model_kwargs is None else sp_model_kwargs __A = kwargs.get("name_or_path" ) if name_or_path is None: logger.warning( "name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b," " you are testing the model, this can safely be ignored" ) __A = "None" # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __A = "<|endoftext|>" if eos_token is None else eos_token __A = "<unk>" if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __A = unk_token if pad_token is None else pad_token __A = eos_token if bos_token is None else bos_token else: __A = "<pad>" if pad_token is None else pad_token __A = "<s>" if bos_token is None else bos_token super().__init__( do_lower_case=A ,remove_space=A ,keep_accents=A ,bos_token=A ,eos_token=A ,unk_token=A ,pad_token=A ,sp_model_kwargs=self.sp_model_kwargs ,**A ,) __A = do_lower_case __A = remove_space __A = keep_accents __A = vocab_file __A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A ) # Used for whitespace normalization in input texts # fmt : off __A = {" ", " ", " ", " ", " ", " ", " ", " ", " ", " ", "", "„"} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __A = re.compile( f'''[{''.join(map(A ,list(range(0 ,9 ) ) + list(range(11 ,32 ) ) + list(range(1_27 ,1_60 ) ) + [1_60, 1_73, 82_03] ) )}]''' ) def __getstate__( self : Optional[int] ): __A = self.__dict__.copy() __A = None return state def __setstate__( self : Optional[Any] ,A : Union[str, Any] ): __A = d # for backward compatibility if not hasattr(self ,"sp_model_kwargs" ): __A = {} __A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def UpperCamelCase_ ( self : List[str] ): return len(self.sp_model ) def UpperCamelCase_ ( self : int ,A : str ): __A = self.non_printing_characters_re.sub("" ,A ) # Normalize whitespaces __A = "".join([char if char not in self.whitespaces else " " for char in text] ) # NFC Unicode normalization __A = unicodedata.normalize("NFC" ,A ) return text def UpperCamelCase_ ( self : Union[str, Any] ,A : str ,**A : Optional[int] ): __A = self.preprocess_text(A ) return self.sp_model.encode(A ,out_type=A ) def UpperCamelCase_ ( self : Any ,A : str ): return self.sp_model.PieceToId(A ) def UpperCamelCase_ ( self : Dict ,A : int ): return self.sp_model.IdToPiece(A ) @staticmethod def UpperCamelCase_ ( A : str ): return out_string def UpperCamelCase_ ( self : str ,A : List[str] ): __A = [] __A = "" __A = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(A ) + token __A = True __A = [] else: current_sub_tokens.append(A ) __A = False out_string += self.sp_model.decode(A ) return out_string def UpperCamelCase_ ( self : str ): __A = {self.convert_ids_to_tokens(A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCamelCase_ ( self : List[str] ,A : str ,A : Optional[str] = None ): if not os.path.isdir(A ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __A = os.path.join( A ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,A ) elif not os.path.isfile(self.vocab_file ): with open(A ,"wb" ) as fi: __A = self.sp_model.serialized_model_proto() fi.write(A ) return (out_vocab_file,) def UpperCamelCase_ ( self : Union[str, Any] ,A : Union[str, List[str]] ,A : Union[str, bool] = False ): if isinstance(A ,A ): __A = self.preprocess_text(A ) __A = self.sp_model.encode(A ) else: __A = [self.preprocess_text(A ) for t in text] __A = self.sp_model.encode(A ) if return_tensors is True or return_tensors == "pt": __A = torch.tensor(A ) return token_ids def UpperCamelCase_ ( self : List[Any] ,A : Union[int, List[int]] ): return self.sp_model.decode(A ) def UpperCamelCase_ ( self : List[str] ,A : "Conversation" ): __A = [f'''User: {text}''' if is_user else f'''Bot: {text}''' for is_user, text in conversation.iter_texts()] __A = ( f'''{self.eos_token}{self.bos_token}''' + f'''{self.bos_token}'''.join(A ) + f'''{self.bos_token}Bot:''' ) return self.encode(text=A )
15
0
import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def snake_case ( snake_case__ :Tuple , snake_case__ :Optional[int] , snake_case__ :List[Any]) -> Tuple: _A = WavaVecaForSequenceClassification.from_pretrained(snake_case__ , config=snake_case__) _A = downstream_dict["""projector.weight"""] _A = downstream_dict["""projector.bias"""] _A = downstream_dict["""model.post_net.linear.weight"""] _A = downstream_dict["""model.post_net.linear.bias"""] return model def snake_case ( snake_case__ :List[Any] , snake_case__ :int , snake_case__ :Any) -> Union[str, Any]: _A = WavaVecaForAudioFrameClassification.from_pretrained(snake_case__ , config=snake_case__) _A = downstream_dict["""model.linear.weight"""] _A = downstream_dict["""model.linear.bias"""] return model def snake_case ( snake_case__ :Optional[int] , snake_case__ :List[Any] , snake_case__ :List[Any]) -> Any: _A = WavaVecaForXVector.from_pretrained(snake_case__ , config=snake_case__) _A = downstream_dict["""connector.weight"""] _A = downstream_dict["""connector.bias"""] for i, kernel_size in enumerate(hf_config.tdnn_kernel): _A = downstream_dict[ F'''model.framelevel_feature_extractor.module.{i}.kernel.weight''' ] _A = downstream_dict[F'''model.framelevel_feature_extractor.module.{i}.kernel.bias'''] _A = downstream_dict["""model.utterancelevel_feature_extractor.linear1.weight"""] _A = downstream_dict["""model.utterancelevel_feature_extractor.linear1.bias"""] _A = downstream_dict["""model.utterancelevel_feature_extractor.linear2.weight"""] _A = downstream_dict["""model.utterancelevel_feature_extractor.linear2.bias"""] _A = downstream_dict["""objective.W"""] return model @torch.no_grad() def snake_case ( snake_case__ :Any , snake_case__ :Any , snake_case__ :Tuple , snake_case__ :Union[str, Any]) -> str: _A = torch.load(snake_case__ , map_location="""cpu""") _A = checkpoint["""Downstream"""] _A = WavaVecaConfig.from_pretrained(snake_case__) _A = WavaVecaFeatureExtractor.from_pretrained( snake_case__ , return_attention_mask=snake_case__ , do_normalize=snake_case__) _A = hf_config.architectures[0] if arch.endswith("""ForSequenceClassification"""): _A = convert_classification(snake_case__ , snake_case__ , snake_case__) elif arch.endswith("""ForAudioFrameClassification"""): _A = convert_diarization(snake_case__ , snake_case__ , snake_case__) elif arch.endswith("""ForXVector"""): _A = convert_xvector(snake_case__ , snake_case__ , snake_case__) else: raise NotImplementedError(F'''S3PRL weights conversion is not supported for {arch}''') if hf_config.use_weighted_layer_sum: _A = checkpoint["""Featurizer"""]["""weights"""] hf_feature_extractor.save_pretrained(snake_case__) hf_model.save_pretrained(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument( '--base_model_name', default=None, type=str, help='Name of the huggingface pretrained base model.' ) parser.add_argument('--config_path', default=None, type=str, help='Path to the huggingface classifier config.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to the s3prl checkpoint.') parser.add_argument('--model_dump_path', default=None, type=str, help='Path to the final converted model.') _SCREAMING_SNAKE_CASE = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
81
import cva import numpy as np class a : """simple docstring""" def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: if k in (0.04, 0.06): _A = k _A = window_size else: raise ValueError("""invalid k value""" ) def __str__( self ) -> str: return str(self.k ) def UpperCAmelCase ( self , lowerCAmelCase_ ) -> tuple[cva.Mat, list[list[int]]]: _A = cva.imread(lowerCAmelCase_ , 0 ) _A , _A = img.shape _A = [] _A = img.copy() _A = cva.cvtColor(lowerCAmelCase_ , cva.COLOR_GRAY2RGB ) _A , _A = np.gradient(lowerCAmelCase_ ) _A = dx**2 _A = dy**2 _A = dx * dy _A = 0.04 _A = self.window_size // 2 for y in range(lowerCAmelCase_ , h - offset ): for x in range(lowerCAmelCase_ , w - offset ): _A = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _A = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _A = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _A = (wxx * wyy) - (wxy**2) _A = wxx + wyy _A = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 2_55 ) return color_img, corner_list if __name__ == "__main__": _SCREAMING_SNAKE_CASE = HarrisCorner(0.04, 3) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
81
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _SCREAMING_SNAKE_CASE : Optional[Any] = {"configuration_xglm": ["XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XGLMConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : List[Any] = ["XGLMTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : str = ["XGLMTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : List[str] = [ "XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XGLMForCausalLM", "XGLMModel", "XGLMPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : Optional[Any] = [ "FlaxXGLMForCausalLM", "FlaxXGLMModel", "FlaxXGLMPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : Union[str, Any] = [ "TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXGLMForCausalLM", "TFXGLMModel", "TFXGLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure)
85
'''simple docstring''' import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class lowerCamelCase_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" a_ =["""image_processor""", """tokenizer"""] a_ ="""OwlViTImageProcessor""" a_ =("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self : Any , _a : str=None , _a : Optional[Any]=None , **_a : int ) -> List[str]: __lowerCamelCase : List[Any] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , _a , ) __lowerCamelCase : Tuple = kwargs.pop('feature_extractor' ) __lowerCamelCase : int = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(_a , _a ) def __call__( self : Tuple , _a : Tuple=None , _a : int=None , _a : List[Any]=None , _a : List[Any]="max_length" , _a : Union[str, Any]="np" , **_a : Union[str, Any] ) -> List[str]: if text is None and query_images is None and images is None: raise ValueError( 'You have to specify at least one text or query image or image. All three cannot be none.' ) if text is not None: if isinstance(_a , _a ) or (isinstance(_a , _a ) and not isinstance(text[0] , _a )): __lowerCamelCase : Any = [self.tokenizer(_a , padding=_a , return_tensors=_a , **_a )] elif isinstance(_a , _a ) and isinstance(text[0] , _a ): __lowerCamelCase : List[Any] = [] # Maximum number of queries across batch __lowerCamelCase : str = max([len(_a ) for t in text] ) # Pad all batch samples to max number of text queries for t in text: if len(_a ) != max_num_queries: __lowerCamelCase : List[Any] = t + [' '] * (max_num_queries - len(_a )) __lowerCamelCase : Dict = self.tokenizer(_a , padding=_a , return_tensors=_a , **_a ) encodings.append(_a ) else: raise TypeError('Input text should be a string, a list of strings or a nested list of strings' ) if return_tensors == "np": __lowerCamelCase : Optional[Any] = np.concatenate([encoding['input_ids'] for encoding in encodings] , axis=0 ) __lowerCamelCase : str = np.concatenate([encoding['attention_mask'] for encoding in encodings] , axis=0 ) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp __lowerCamelCase : int = jnp.concatenate([encoding['input_ids'] for encoding in encodings] , axis=0 ) __lowerCamelCase : List[str] = jnp.concatenate([encoding['attention_mask'] for encoding in encodings] , axis=0 ) elif return_tensors == "pt" and is_torch_available(): import torch __lowerCamelCase : str = torch.cat([encoding['input_ids'] for encoding in encodings] , dim=0 ) __lowerCamelCase : Optional[Any] = torch.cat([encoding['attention_mask'] for encoding in encodings] , dim=0 ) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf __lowerCamelCase : List[Any] = tf.stack([encoding['input_ids'] for encoding in encodings] , axis=0 ) __lowerCamelCase : int = tf.stack([encoding['attention_mask'] for encoding in encodings] , axis=0 ) else: raise ValueError('Target return tensor type could not be returned' ) __lowerCamelCase : Any = BatchEncoding() __lowerCamelCase : Dict = input_ids __lowerCamelCase : str = attention_mask if query_images is not None: __lowerCamelCase : Optional[int] = BatchEncoding() __lowerCamelCase : List[Any] = self.image_processor( _a , return_tensors=_a , **_a ).pixel_values __lowerCamelCase : str = query_pixel_values if images is not None: __lowerCamelCase : Union[str, Any] = self.image_processor(_a , return_tensors=_a , **_a ) if text is not None and images is not None: __lowerCamelCase : Tuple = image_features.pixel_values return encoding elif query_images is not None and images is not None: __lowerCamelCase : Tuple = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**_a ) , tensor_type=_a ) def _lowercase ( self : Optional[Any] , *_a : List[str] , **_a : Dict ) -> int: return self.image_processor.post_process(*_a , **_a ) def _lowercase ( self : str , *_a : str , **_a : List[str] ) -> int: return self.image_processor.post_process_object_detection(*_a , **_a ) def _lowercase ( self : int , *_a : List[Any] , **_a : Optional[int] ) -> str: return self.image_processor.post_process_image_guided_detection(*_a , **_a ) def _lowercase ( self : Tuple , *_a : Optional[Any] , **_a : List[Any] ) -> Tuple: return self.tokenizer.batch_decode(*_a , **_a ) def _lowercase ( self : str , *_a : Optional[Any] , **_a : str ) -> Union[str, Any]: return self.tokenizer.decode(*_a , **_a ) @property def _lowercase ( self : Any ) -> Any: warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , _a , ) return self.image_processor_class @property def _lowercase ( self : Union[str, Any] ) -> Any: warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , _a , ) return self.image_processor
208
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : List[Any] = logging.get_logger(__name__) lowerCAmelCase : Any = { """studio-ousia/luke-base""": """https://huggingface.co/studio-ousia/luke-base/resolve/main/config.json""", """studio-ousia/luke-large""": """https://huggingface.co/studio-ousia/luke-large/resolve/main/config.json""", } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __magic_name__ = "luke" def __init__( self , snake_case__=5_0267 , snake_case__=50_0000 , snake_case__=768 , snake_case__=256 , snake_case__=12 , snake_case__=12 , snake_case__=3072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=2 , snake_case__=0.02 , snake_case__=1E-12 , snake_case__=True , snake_case__=None , snake_case__=1 , snake_case__=0 , snake_case__=2 , **snake_case__ , ): '''simple docstring''' super().__init__(pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ , **snake_case__ ) _lowerCAmelCase : Union[str, Any] = vocab_size _lowerCAmelCase : Any = entity_vocab_size _lowerCAmelCase : Optional[Any] = hidden_size _lowerCAmelCase : Union[str, Any] = entity_emb_size _lowerCAmelCase : int = num_hidden_layers _lowerCAmelCase : Union[str, Any] = num_attention_heads _lowerCAmelCase : str = hidden_act _lowerCAmelCase : Optional[int] = intermediate_size _lowerCAmelCase : Any = hidden_dropout_prob _lowerCAmelCase : str = attention_probs_dropout_prob _lowerCAmelCase : Tuple = max_position_embeddings _lowerCAmelCase : int = type_vocab_size _lowerCAmelCase : List[str] = initializer_range _lowerCAmelCase : Union[str, Any] = layer_norm_eps _lowerCAmelCase : Any = use_entity_aware_attention _lowerCAmelCase : Union[str, Any] = classifier_dropout
25
'''simple docstring''' lowerCAmelCase : Union[str, Any] = 0 # The first color of the flag. lowerCAmelCase : Optional[int] = 1 # The second color of the flag. lowerCAmelCase : int = 2 # The third color of the flag. lowerCAmelCase : Any = (red, white, blue) def lowercase (_A ): """simple docstring""" if not sequence: return [] if len(_A ) == 1: return list(_A ) _lowerCAmelCase : Optional[int] = 0 _lowerCAmelCase : List[str] = len(_A ) - 1 _lowerCAmelCase : Optional[Any] = 0 while mid <= high: if sequence[mid] == colors[0]: _lowerCAmelCase , _lowerCAmelCase : Tuple = sequence[mid], sequence[low] low += 1 mid += 1 elif sequence[mid] == colors[1]: mid += 1 elif sequence[mid] == colors[2]: _lowerCAmelCase , _lowerCAmelCase : Tuple = sequence[high], sequence[mid] high -= 1 else: _lowerCAmelCase : Optional[int] = f'The elements inside the sequence must contains only {colors} values' raise ValueError(_A ) return sequence if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase : str = input("""Enter numbers separated by commas:\n""").strip() lowerCAmelCase : Dict = [int(item.strip()) for item in user_input.split(""",""")] print(F'''{dutch_national_flag_sort(unsorted)}''')
25
1
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Sequence, Value from .base import TaskTemplate @dataclass(frozen=__SCREAMING_SNAKE_CASE ) class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case_ = field(default="question-answering-extractive" , metadata={"include_in_asdict_even_if_is_default": True} ) snake_case_ = Features({"question": Value("string" ), "context": Value("string" )} ) snake_case_ = Features( { "answers": Sequence( { "text": Value("string" ), "answer_start": Value("int32" ), } ) } ) snake_case_ = "question" snake_case_ = "context" snake_case_ = "answers" @property def UpperCamelCase_ ( self : Dict ): return {self.question_column: "question", self.context_column: "context", self.answers_column: "answers"}
15
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar a = TypeVar('''T''') class lowercase_ ( Generic[T] ): '''simple docstring''' def __init__( self : Any , _UpperCAmelCase : T ): _A = data _A = None def __str__( self : str ): return F'''{self.data}''' class lowercase_ ( Generic[T] ): '''simple docstring''' def __init__( self : Tuple ): _A = None def __iter__( self : List[Any] ): _A = self.top while node: yield node.data _A = node.next def __str__( self : Union[str, Any] ): return "->".join([str(_UpperCAmelCase ) for item in self] ) def __len__( self : List[Any] ): return len(tuple(iter(self ) ) ) def lowerCAmelCase_ ( self : str ): return self.top is None def lowerCAmelCase_ ( self : Union[str, Any] , _UpperCAmelCase : T ): _A = Node(_UpperCAmelCase ) if not self.is_empty(): _A = self.top _A = node def lowerCAmelCase_ ( self : Dict ): if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top , _UpperCAmelCase ) _A = self.top _A = self.top.next return pop_node.data def lowerCAmelCase_ ( self : Tuple ): if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def lowerCAmelCase_ ( self : Optional[Any] ): _A = None if __name__ == "__main__": from doctest import testmod testmod()
315
0
"""simple docstring""" from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. __UpperCamelCase : Union[str, Any] = 1_0 def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ ): for i in range(lowerCAmelCase__ , lowerCAmelCase__ ): if array[i] == target: return i return -1 def __SCREAMING_SNAKE_CASE ( A_ , A_ ): lowerCAmelCase__ : Optional[int] = 0 lowerCAmelCase__ : Any = len(lowerCAmelCase__ ) while left <= right: if right - left < precision: return lin_search(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) lowerCAmelCase__ : List[Any] = (left + right) // 3 + 1 lowerCAmelCase__ : List[str] = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: lowerCAmelCase__ : int = one_third - 1 elif array[two_third] < target: lowerCAmelCase__ : str = two_third + 1 else: lowerCAmelCase__ : int = one_third + 1 lowerCAmelCase__ : int = two_third - 1 else: return -1 def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ ): if left < right: if right - left < precision: return lin_search(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) lowerCAmelCase__ : Tuple = (left + right) // 3 + 1 lowerCAmelCase__ : List[Any] = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(lowerCAmelCase__ , one_third - 1 , lowerCAmelCase__ , lowerCAmelCase__ ) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) else: return rec_ternary_search(one_third + 1 , two_third - 1 , lowerCAmelCase__ , lowerCAmelCase__ ) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() __UpperCamelCase : Dict = input('''Enter numbers separated by comma:\n''').strip() __UpperCamelCase : Dict = [int(item.strip()) for item in user_input.split(''',''')] assert collection == sorted(collection), F"List must be ordered.\n{collection}." __UpperCamelCase : List[Any] = int(input('''Enter the number to be found in the list:\n''').strip()) __UpperCamelCase : Optional[int] = ite_ternary_search(collection, target) __UpperCamelCase : Union[str, Any] = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(F'''Iterative search: {target} found at positions: {resulta}''') print(F'''Recursive search: {target} found at positions: {resulta}''') else: print('''Not found''')
361
"""simple docstring""" import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax from transformers.models.vit.modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __init__( self : List[Any] ,lowercase_ : List[str] ,lowercase_ : int=1_3 ,lowercase_ : Optional[int]=3_0 ,lowercase_ : int=2 ,lowercase_ : List[Any]=3 ,lowercase_ : str=True ,lowercase_ : int=True ,lowercase_ : str=3_2 ,lowercase_ : Optional[int]=5 ,lowercase_ : Optional[Any]=4 ,lowercase_ : Any=3_7 ,lowercase_ : str="gelu" ,lowercase_ : Any=0.1 ,lowercase_ : List[Any]=0.1 ,lowercase_ : int=1_0 ,lowercase_ : str=0.02 ,): lowerCAmelCase__ : Optional[int] = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : str = image_size lowerCAmelCase__ : Dict = patch_size lowerCAmelCase__ : Dict = num_channels lowerCAmelCase__ : Union[str, Any] = is_training lowerCAmelCase__ : Optional[int] = use_labels lowerCAmelCase__ : List[Any] = hidden_size lowerCAmelCase__ : Dict = num_hidden_layers lowerCAmelCase__ : int = num_attention_heads lowerCAmelCase__ : Any = intermediate_size lowerCAmelCase__ : List[Any] = hidden_act lowerCAmelCase__ : Optional[int] = hidden_dropout_prob lowerCAmelCase__ : List[str] = attention_probs_dropout_prob lowerCAmelCase__ : Any = type_sequence_label_size lowerCAmelCase__ : Optional[int] = initializer_range # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCAmelCase__ : int = (image_size // patch_size) ** 2 lowerCAmelCase__ : Dict = num_patches + 1 def __lowerCAmelCase ( self : List[str] ): lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : List[Any] = ViTConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=lowercase_ ,initializer_range=self.initializer_range ,) return config, pixel_values def __lowerCAmelCase ( self : Tuple ,lowercase_ : List[Any] ,lowercase_ : Optional[int] ): lowerCAmelCase__ : Optional[Any] = FlaxViTModel(config=lowercase_ ) lowerCAmelCase__ : Dict = model(lowercase_ ) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) lowerCAmelCase__ : int = (self.image_size, self.image_size) lowerCAmelCase__ : int = (self.patch_size, self.patch_size) lowerCAmelCase__ : Union[str, Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, num_patches + 1, self.hidden_size) ) def __lowerCAmelCase ( self : int ,lowercase_ : List[Any] ,lowercase_ : List[str] ): lowerCAmelCase__ : Optional[int] = self.type_sequence_label_size lowerCAmelCase__ : Any = FlaxViTForImageClassification(config=lowercase_ ) lowerCAmelCase__ : Any = model(lowercase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCAmelCase__ : List[Any] = 1 lowerCAmelCase__ : Tuple = FlaxViTForImageClassification(lowercase_ ) lowerCAmelCase__ : Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : str = model(lowercase_ ) def __lowerCAmelCase ( self : Union[str, Any] ): lowerCAmelCase__ : Any = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) ,( lowerCAmelCase__ ) , ) : Any = config_and_inputs lowerCAmelCase__ : List[str] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class SCREAMING_SNAKE_CASE ( a_ , unittest.TestCase ): """simple docstring""" lowercase__ = (FlaxViTModel, FlaxViTForImageClassification) if is_flax_available() else () def __lowerCAmelCase ( self : Optional[Any] ): lowerCAmelCase__ : Tuple = FlaxViTModelTester(self ) lowerCAmelCase__ : List[str] = ConfigTester(self ,config_class=lowercase_ ,has_text_modality=lowercase_ ,hidden_size=3_7 ) def __lowerCAmelCase ( self : Dict ): self.config_tester.run_common_tests() def __lowerCAmelCase ( self : List[Any] ): lowerCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def __lowerCAmelCase ( self : Tuple ): lowerCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) def __lowerCAmelCase ( self : Optional[int] ): lowerCAmelCase__ ,lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Union[str, Any] = model_class(lowercase_ ) lowerCAmelCase__ : List[str] = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : List[Any] = [*signature.parameters.keys()] lowerCAmelCase__ : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase_ ) def __lowerCAmelCase ( self : str ): lowerCAmelCase__ ,lowerCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): lowerCAmelCase__ : Dict = self._prepare_for_class(lowercase_ ,lowercase_ ) lowerCAmelCase__ : Tuple = model_class(lowercase_ ) @jax.jit def model_jitted(lowercase_ : List[Any] ,**lowercase_ : Optional[int] ): return model(pixel_values=lowercase_ ,**lowercase_ ) with self.subTest('''JIT Enabled''' ): lowerCAmelCase__ : Optional[Any] = model_jitted(**lowercase_ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): lowerCAmelCase__ : Optional[int] = model_jitted(**lowercase_ ).to_tuple() self.assertEqual(len(lowercase_ ) ,len(lowercase_ ) ) for jitted_output, output in zip(lowercase_ ,lowercase_ ): self.assertEqual(jitted_output.shape ,output.shape ) @slow def __lowerCAmelCase ( self : List[Any] ): for model_class_name in self.all_model_classes: lowerCAmelCase__ : List[Any] = model_class_name.from_pretrained('''google/vit-base-patch16-224''' ) lowerCAmelCase__ : Optional[int] = model(np.ones((1, 3, 2_2_4, 2_2_4) ) ) self.assertIsNotNone(lowercase_ )
74
0
"""simple docstring""" from PIL import Image def _snake_case ( UpperCAmelCase_ : Image , UpperCAmelCase_ : float ): def brightness(UpperCAmelCase_ : int ) -> float: return 128 + level + (c - 128) if not -2_55.0 <= level <= 2_55.0: raise ValueError("""level must be between -255.0 (black) and 255.0 (white)""" ) return img.point(UpperCAmelCase_ ) if __name__ == "__main__": # Load image with Image.open('image_data/lena.jpg') as img: # Change brightness to 100 SCREAMING_SNAKE_CASE_ : Optional[int] = change_brightness(img, 1_0_0) brigt_img.save('image_data/lena_brightness.png', format='png')
335
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig SCREAMING_SNAKE_CASE_ : Union[str, Any] = logging.get_logger(__name__) # General docstring SCREAMING_SNAKE_CASE_ : Union[str, Any] = 'MobileNetV1Config' # Base docstring SCREAMING_SNAKE_CASE_ : str = 'google/mobilenet_v1_1.0_224' SCREAMING_SNAKE_CASE_ : List[str] = [1, 1_0_2_4, 7, 7] # Image classification docstring SCREAMING_SNAKE_CASE_ : Optional[Any] = 'google/mobilenet_v1_1.0_224' SCREAMING_SNAKE_CASE_ : Tuple = 'tabby, tabby cat' SCREAMING_SNAKE_CASE_ : Union[str, Any] = [ 'google/mobilenet_v1_1.0_224', 'google/mobilenet_v1_0.75_192', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def _snake_case ( UpperCAmelCase_ : int , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Dict=None ): A__ = {} if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): A__ = model.mobilenet_va else: A__ = model A__ = """MobilenetV1/Conv2d_0/""" A__ = backbone.conv_stem.convolution.weight A__ = backbone.conv_stem.normalization.bias A__ = backbone.conv_stem.normalization.weight A__ = backbone.conv_stem.normalization.running_mean A__ = backbone.conv_stem.normalization.running_var for i in range(13 ): A__ = i + 1 A__ = i * 2 A__ = backbone.layer[pt_index] A__ = F"""MobilenetV1/Conv2d_{tf_index}_depthwise/""" A__ = pointer.convolution.weight A__ = pointer.normalization.bias A__ = pointer.normalization.weight A__ = pointer.normalization.running_mean A__ = pointer.normalization.running_var A__ = backbone.layer[pt_index + 1] A__ = F"""MobilenetV1/Conv2d_{tf_index}_pointwise/""" A__ = pointer.convolution.weight A__ = pointer.normalization.bias A__ = pointer.normalization.weight A__ = pointer.normalization.running_mean A__ = pointer.normalization.running_var if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): A__ = """MobilenetV1/Logits/Conv2d_1c_1x1/""" A__ = model.classifier.weight A__ = model.classifier.bias return tf_to_pt_map def _snake_case ( UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : List[Any] ): try: import numpy as np import tensorflow as tf except ImportError: logger.error( """Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see """ """https://www.tensorflow.org/install/ for installation instructions.""" ) raise # Load weights from TF model A__ = tf.train.list_variables(UpperCAmelCase_ ) A__ = {} for name, shape in init_vars: logger.info(F"""Loading TF weight {name} with shape {shape}""" ) A__ = tf.train.load_variable(UpperCAmelCase_ , UpperCAmelCase_ ) A__ = array # Build TF to PyTorch weights loading map A__ = _build_tf_to_pytorch_map(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) for name, pointer in tf_to_pt_map.items(): logger.info(F"""Importing {name}""" ) if name not in tf_weights: logger.info(F"""{name} not in tf pre-trained weights, skipping""" ) continue A__ = tf_weights[name] if "depthwise_weights" in name: logger.info("""Transposing depthwise""" ) A__ = np.transpose(UpperCAmelCase_ , (2, 3, 0, 1) ) elif "weights" in name: logger.info("""Transposing""" ) if len(pointer.shape ) == 2: # copying into linear layer A__ = array.squeeze().transpose() else: A__ = np.transpose(UpperCAmelCase_ , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F"""Pointer shape {pointer.shape} and array shape {array.shape} mismatched""" ) logger.info(F"""Initialize PyTorch weight {name} {array.shape}""" ) A__ = torch.from_numpy(UpperCAmelCase_ ) tf_weights.pop(UpperCAmelCase_ , UpperCAmelCase_ ) tf_weights.pop(name + """/RMSProp""" , UpperCAmelCase_ ) tf_weights.pop(name + """/RMSProp_1""" , UpperCAmelCase_ ) tf_weights.pop(name + """/ExponentialMovingAverage""" , UpperCAmelCase_ ) logger.info(F"""Weights not copied to PyTorch model: {', '.join(tf_weights.keys() )}""" ) return model def _snake_case ( UpperCAmelCase_ : torch.Tensor , UpperCAmelCase_ : nn.Convad ): A__ , A__ = features.shape[-2:] A__ , A__ = conv_layer.stride A__ , A__ = conv_layer.kernel_size if in_height % stride_height == 0: A__ = max(kernel_height - stride_height , 0 ) else: A__ = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: A__ = max(kernel_width - stride_width , 0 ) else: A__ = max(kernel_width - (in_width % stride_width) , 0 ) A__ = pad_along_width // 2 A__ = pad_along_width - pad_left A__ = pad_along_height // 2 A__ = pad_along_height - pad_top A__ = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(UpperCAmelCase_ , UpperCAmelCase_ , """constant""" , 0.0 ) class a ( nn.Module ): """simple docstring""" def __init__( self: Union[str, Any] , UpperCamelCase: MobileNetVaConfig , UpperCamelCase: int , UpperCamelCase: int , UpperCamelCase: int , UpperCamelCase: Optional[int] = 1 , UpperCamelCase: Optional[int] = 1 , UpperCamelCase: bool = False , UpperCamelCase: Optional[bool] = True , UpperCamelCase: Optional[bool or str] = True , ): """simple docstring""" super().__init__() A__ = config if in_channels % groups != 0: raise ValueError(f"""Input channels ({in_channels}) are not divisible by {groups} groups.""" ) if out_channels % groups != 0: raise ValueError(f"""Output channels ({out_channels}) are not divisible by {groups} groups.""" ) A__ = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) A__ = nn.Convad( in_channels=UpperCamelCase , out_channels=UpperCamelCase , kernel_size=UpperCamelCase , stride=UpperCamelCase , padding=UpperCamelCase , groups=UpperCamelCase , bias=UpperCamelCase , padding_mode="""zeros""" , ) if use_normalization: A__ = nn.BatchNormad( num_features=UpperCamelCase , eps=config.layer_norm_eps , momentum=0.9_997 , affine=UpperCamelCase , track_running_stats=UpperCamelCase , ) else: A__ = None if use_activation: if isinstance(UpperCamelCase , UpperCamelCase ): A__ = ACTaFN[use_activation] elif isinstance(config.hidden_act , UpperCamelCase ): A__ = ACTaFN[config.hidden_act] else: A__ = config.hidden_act else: A__ = None def UpperCamelCase ( self: List[Any] , UpperCamelCase: torch.Tensor ): """simple docstring""" if self.config.tf_padding: A__ = apply_tf_padding(UpperCamelCase , self.convolution ) A__ = self.convolution(UpperCamelCase ) if self.normalization is not None: A__ = self.normalization(UpperCamelCase ) if self.activation is not None: A__ = self.activation(UpperCamelCase ) return features class a ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = MobileNetVaConfig UpperCAmelCase = load_tf_weights_in_mobilenet_va UpperCAmelCase = "mobilenet_v1" UpperCAmelCase = "pixel_values" UpperCAmelCase = False def UpperCamelCase ( self: Any , UpperCamelCase: Union[nn.Linear, nn.Convad] ): """simple docstring""" if isinstance(UpperCamelCase , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCamelCase , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) SCREAMING_SNAKE_CASE_ : Optional[Any] = r'\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' SCREAMING_SNAKE_CASE_ : Optional[Any] = r'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`MobileNetV1ImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top.", _lowerCamelCase, ) class a ( _lowerCamelCase ): """simple docstring""" def __init__( self: Any , UpperCamelCase: MobileNetVaConfig , UpperCamelCase: bool = True ): """simple docstring""" super().__init__(UpperCamelCase ) A__ = config A__ = 32 A__ = max(int(depth * config.depth_multiplier ) , config.min_depth ) A__ = MobileNetVaConvLayer( UpperCamelCase , in_channels=config.num_channels , out_channels=UpperCamelCase , kernel_size=3 , stride=2 , ) A__ = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] A__ = nn.ModuleList() for i in range(13 ): A__ = out_channels if strides[i] == 2 or i == 0: depth *= 2 A__ = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( UpperCamelCase , in_channels=UpperCamelCase , out_channels=UpperCamelCase , kernel_size=3 , stride=strides[i] , groups=UpperCamelCase , ) ) self.layer.append( MobileNetVaConvLayer( UpperCamelCase , in_channels=UpperCamelCase , out_channels=UpperCamelCase , kernel_size=1 , ) ) A__ = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def UpperCamelCase ( self: Dict , UpperCamelCase: Optional[Any] ): """simple docstring""" raise NotImplementedError @add_start_docstrings_to_model_forward(UpperCamelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCamelCase , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCamelCase ( self: Tuple , UpperCamelCase: Optional[torch.Tensor] = None , UpperCamelCase: Optional[bool] = None , UpperCamelCase: Optional[bool] = None , ): """simple docstring""" A__ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) A__ = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("""You have to specify pixel_values""" ) A__ = self.conv_stem(UpperCamelCase ) A__ = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): A__ = layer_module(UpperCamelCase ) if output_hidden_states: A__ = all_hidden_states + (hidden_states,) A__ = hidden_states if self.pooler is not None: A__ = torch.flatten(self.pooler(UpperCamelCase ) , start_dim=1 ) else: A__ = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCamelCase , pooler_output=UpperCamelCase , hidden_states=UpperCamelCase , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n ", _lowerCamelCase, ) class a ( _lowerCamelCase ): """simple docstring""" def __init__( self: Union[str, Any] , UpperCamelCase: MobileNetVaConfig ): """simple docstring""" super().__init__(UpperCamelCase ) A__ = config.num_labels A__ = MobileNetVaModel(UpperCamelCase ) A__ = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head A__ = nn.Dropout(config.classifier_dropout_prob , inplace=UpperCamelCase ) A__ = nn.Linear(UpperCamelCase , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCamelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCamelCase ( self: Union[str, Any] , UpperCamelCase: Optional[torch.Tensor] = None , UpperCamelCase: Optional[bool] = None , UpperCamelCase: Optional[torch.Tensor] = None , UpperCamelCase: Optional[bool] = None , ): """simple docstring""" A__ = return_dict if return_dict is not None else self.config.use_return_dict A__ = self.mobilenet_va(UpperCamelCase , output_hidden_states=UpperCamelCase , return_dict=UpperCamelCase ) A__ = outputs.pooler_output if return_dict else outputs[1] A__ = self.classifier(self.dropout(UpperCamelCase ) ) A__ = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: A__ = """regression""" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): A__ = """single_label_classification""" else: A__ = """multi_label_classification""" if self.config.problem_type == "regression": A__ = MSELoss() if self.num_labels == 1: A__ = loss_fct(logits.squeeze() , labels.squeeze() ) else: A__ = loss_fct(UpperCamelCase , UpperCamelCase ) elif self.config.problem_type == "single_label_classification": A__ = CrossEntropyLoss() A__ = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": A__ = BCEWithLogitsLoss() A__ = loss_fct(UpperCamelCase , UpperCamelCase ) if not return_dict: A__ = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=UpperCamelCase , logits=UpperCamelCase , hidden_states=outputs.hidden_states , )
335
1
"""simple docstring""" from __future__ import annotations def lowercase__( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : list[str] | None = None , __SCREAMING_SNAKE_CASE : dict[str, float] | None = None , __SCREAMING_SNAKE_CASE : bool = False , ): """simple docstring""" lowercase_ : Tuple = cipher_alphabet or [chr(lowerCAmelCase__ ) for i in range(97 , 1_23 )] # If the argument is None or the user provided an empty dictionary if not frequencies_dict: # Frequencies of letters in the english language (how much they show up) lowercase_ : int = { '''a''': 0.0_8497, '''b''': 0.0_1492, '''c''': 0.0_2202, '''d''': 0.0_4253, '''e''': 0.1_1162, '''f''': 0.0_2228, '''g''': 0.0_2015, '''h''': 0.0_6094, '''i''': 0.0_7546, '''j''': 0.0_0153, '''k''': 0.0_1292, '''l''': 0.0_4025, '''m''': 0.0_2406, '''n''': 0.0_6749, '''o''': 0.0_7507, '''p''': 0.0_1929, '''q''': 0.0_0095, '''r''': 0.0_7587, '''s''': 0.0_6327, '''t''': 0.0_9356, '''u''': 0.0_2758, '''v''': 0.0_0978, '''w''': 0.0_2560, '''x''': 0.0_0150, '''y''': 0.0_1994, '''z''': 0.0_0077, } else: # Custom frequencies dictionary lowercase_ : Union[str, Any] = frequencies_dict if not case_sensitive: lowercase_ : Any = ciphertext.lower() # Chi squared statistic values lowercase_ : dict[int, tuple[float, str]] = {} # cycle through all of the shifts for shift in range(len(lowerCAmelCase__ ) ): lowercase_ : Tuple = '''''' # decrypt the message with the shift for letter in ciphertext: try: # Try to index the letter in the alphabet lowercase_ : Optional[int] = (alphabet_letters.index(letter.lower() ) - shift) % len( lowerCAmelCase__ ) decrypted_with_shift += ( alphabet_letters[new_key].upper() if case_sensitive and letter.isupper() else alphabet_letters[new_key] ) except ValueError: # Append the character if it isn't in the alphabet decrypted_with_shift += letter lowercase_ : str = 0.0 # Loop through each letter in the decoded message with the shift for letter in decrypted_with_shift: if case_sensitive: lowercase_ : Optional[int] = letter.lower() if letter in frequencies: # Get the amount of times the letter occurs in the message lowercase_ : Tuple = decrypted_with_shift.lower().count(lowerCAmelCase__ ) # Get the excepcted amount of times the letter should appear based # on letter frequencies lowercase_ : Any = frequencies[letter] * occurrences # Complete the chi squared statistic formula lowercase_ : Optional[int] = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value else: if letter.lower() in frequencies: # Get the amount of times the letter occurs in the message lowercase_ : Optional[Any] = decrypted_with_shift.count(lowerCAmelCase__ ) # Get the excepcted amount of times the letter should appear based # on letter frequencies lowercase_ : Dict = frequencies[letter] * occurrences # Complete the chi squared statistic formula lowercase_ : int = ((occurrences - expected) ** 2) / expected # Add the margin of error to the total chi squared statistic chi_squared_statistic += chi_letter_value # Add the data to the chi_squared_statistic_values dictionary lowercase_ : Optional[int] = ( chi_squared_statistic, decrypted_with_shift, ) # Get the most likely cipher by finding the cipher with the smallest chi squared # statistic def chi_squared_statistic_values_sorting_key(__SCREAMING_SNAKE_CASE : int ) -> tuple[float, str]: return chi_squared_statistic_values[key] lowercase_ : int = min( lowerCAmelCase__ , key=lowerCAmelCase__ , ) # Get all the data from the most likely cipher (key, decoded message) ( lowercase_ ) : int = chi_squared_statistic_values[most_likely_cipher] # Return the data on the most likely shift return ( most_likely_cipher, most_likely_cipher_chi_squared_value, decoded_most_likely_cipher, )
352
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) class UpperCamelCase ( lowercase_ ): lowercase = ['input_values', 'padding_mask'] def __init__( self ,__UpperCamelCase = 1 ,__UpperCamelCase = 2_4000 ,__UpperCamelCase = 0.0 ,__UpperCamelCase = None ,__UpperCamelCase = None ,**__UpperCamelCase ,) -> Any: '''simple docstring''' super().__init__(feature_size=__UpperCamelCase ,sampling_rate=__UpperCamelCase ,padding_value=__UpperCamelCase ,**__UpperCamelCase ) lowercase_ : List[str] = chunk_length_s lowercase_ : Tuple = overlap @property def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 ,int((1.0 - self.overlap) * self.chunk_length ) ) def __call__( self ,__UpperCamelCase ,__UpperCamelCase = None ,__UpperCamelCase = False ,__UpperCamelCase = None ,__UpperCamelCase = None ,__UpperCamelCase = None ,) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided audio input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) if padding and truncation: raise ValueError('Both padding and truncation were set. Make sure you only set one.' ) elif padding is None: # by default let's pad the inputs lowercase_ : Optional[int] = True lowercase_ : Optional[int] = bool( isinstance(__UpperCamelCase ,(list, tuple) ) and (isinstance(raw_audio[0] ,(np.ndarray, tuple, list) )) ) if is_batched: lowercase_ : int = [np.asarray(__UpperCamelCase ,dtype=np.floataa ).T for audio in raw_audio] elif not is_batched and not isinstance(__UpperCamelCase ,np.ndarray ): lowercase_ : Any = np.asarray(__UpperCamelCase ,dtype=np.floataa ) elif isinstance(__UpperCamelCase ,np.ndarray ) and raw_audio.dtype is np.dtype(np.floataa ): lowercase_ : List[str] = raw_audio.astype(np.floataa ) # always return batch if not is_batched: lowercase_ : Dict = [np.asarray(__UpperCamelCase ).T] # verify inputs are valid for idx, example in enumerate(__UpperCamelCase ): if example.ndim > 2: raise ValueError(f'''Expected input shape (channels, length) but got shape {example.shape}''' ) if self.feature_size == 1 and example.ndim != 1: raise ValueError(f'''Expected mono audio but example has {example.shape[-1]} channels''' ) if self.feature_size == 2 and example.shape[-1] != 2: raise ValueError(f'''Expected stereo audio but example has {example.shape[-1]} channels''' ) lowercase_ : Optional[int] = None lowercase_ : List[Any] = BatchFeature({'input_values': raw_audio} ) if self.chunk_stride is not None and self.chunk_length is not None and max_length is None: if truncation: lowercase_ : List[Any] = min(array.shape[0] for array in raw_audio ) lowercase_ : int = int(np.floor(max_length / self.chunk_stride ) ) lowercase_ : Dict = (nb_step - 1) * self.chunk_stride + self.chunk_length elif padding: lowercase_ : List[Any] = max(array.shape[0] for array in raw_audio ) lowercase_ : Tuple = int(np.ceil(max_length / self.chunk_stride ) ) lowercase_ : List[str] = (nb_step - 1) * self.chunk_stride + self.chunk_length lowercase_ : Union[str, Any] = 'max_length' else: lowercase_ : int = input_values # normal padding on batch if padded_inputs is None: lowercase_ : int = self.pad( __UpperCamelCase ,max_length=__UpperCamelCase ,truncation=__UpperCamelCase ,padding=__UpperCamelCase ,return_attention_mask=__UpperCamelCase ,) if padding: lowercase_ : Optional[int] = padded_inputs.pop('attention_mask' ) lowercase_ : Dict = [] for example in padded_inputs.pop('input_values' ): if self.feature_size == 1: lowercase_ : Optional[int] = example[..., None] input_values.append(example.T ) lowercase_ : str = input_values if return_tensors is not None: lowercase_ : List[Any] = padded_inputs.convert_to_tensors(__UpperCamelCase ) return padded_inputs
321
0
'''simple docstring''' from __future__ import annotations lowerCAmelCase: Any = [True] * 1_0_0_0_0_0_1 lowerCAmelCase: List[str] = 2 while i * i <= 1_0_0_0_0_0_0: if seive[i]: for j in range(i * i, 1_0_0_0_0_0_1, i): lowerCAmelCase: List[str] = False i += 1 def lowerCamelCase__ ( _A ): return seive[n] def lowerCamelCase__ ( _A ): return any(digit in '02468' for digit in str(_A ) ) def lowerCamelCase__ ( _A = 100_0000 ): a : List[str] = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(_A ) and not contains_an_even_digit(_A ): a : Any = str(_A ) a : List[Any] = [int(str_num[j:] + str_num[:j] ) for j in range(len(_A ) )] if all(is_prime(_A ) for i in list_nums ): result.append(_A ) return result def lowerCamelCase__ ( ): return len(find_circular_primes() ) if __name__ == "__main__": print(F"{len(find_circular_primes()) = }")
297
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) lowerCAmelCase: Union[str, Any] = { 'configuration_speecht5': [ 'SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP', 'SpeechT5Config', 'SpeechT5HifiGanConfig', ], 'feature_extraction_speecht5': ['SpeechT5FeatureExtractor'], 'processing_speecht5': ['SpeechT5Processor'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase: List[Any] = ['SpeechT5Tokenizer'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase: Any = [ 'SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST', 'SpeechT5ForSpeechToText', 'SpeechT5ForSpeechToSpeech', 'SpeechT5ForTextToSpeech', 'SpeechT5Model', 'SpeechT5PreTrainedModel', 'SpeechT5HifiGan', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys lowerCAmelCase: Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
297
1
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE_ : List[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ : Any = { 'facebook/wav2vec2-base-960h': 'https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json', # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class a ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = "wav2vec2" def __init__( self: List[str] , UpperCamelCase: List[str]=32 , UpperCamelCase: Union[str, Any]=7_68 , UpperCamelCase: List[str]=12 , UpperCamelCase: Tuple=12 , UpperCamelCase: Union[str, Any]=30_72 , UpperCamelCase: List[Any]="gelu" , UpperCamelCase: Union[str, Any]=0.1 , UpperCamelCase: Optional[int]=0.1 , UpperCamelCase: List[Any]=0.1 , UpperCamelCase: int=0.0 , UpperCamelCase: Dict=0.0 , UpperCamelCase: Optional[int]=0.1 , UpperCamelCase: Dict=0.1 , UpperCamelCase: str=0.02 , UpperCamelCase: Any=1e-5 , UpperCamelCase: List[str]="group" , UpperCamelCase: Optional[Any]="gelu" , UpperCamelCase: int=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , UpperCamelCase: str=(5, 2, 2, 2, 2, 2, 2) , UpperCamelCase: Union[str, Any]=(10, 3, 3, 3, 3, 2, 2) , UpperCamelCase: Any=False , UpperCamelCase: Tuple=1_28 , UpperCamelCase: Union[str, Any]=16 , UpperCamelCase: List[str]=False , UpperCamelCase: str=True , UpperCamelCase: int=0.05 , UpperCamelCase: List[str]=10 , UpperCamelCase: Any=2 , UpperCamelCase: int=0.0 , UpperCamelCase: List[str]=10 , UpperCamelCase: List[str]=0 , UpperCamelCase: Optional[int]=3_20 , UpperCamelCase: Any=2 , UpperCamelCase: List[Any]=0.1 , UpperCamelCase: int=1_00 , UpperCamelCase: Any=2_56 , UpperCamelCase: List[Any]=2_56 , UpperCamelCase: List[str]=0.1 , UpperCamelCase: Union[str, Any]="sum" , UpperCamelCase: int=False , UpperCamelCase: str=False , UpperCamelCase: List[Any]=2_56 , UpperCamelCase: Optional[Any]=(5_12, 5_12, 5_12, 5_12, 15_00) , UpperCamelCase: int=(5, 3, 3, 1, 1) , UpperCamelCase: Union[str, Any]=(1, 2, 3, 1, 1) , UpperCamelCase: List[str]=5_12 , UpperCamelCase: Tuple=0 , UpperCamelCase: int=1 , UpperCamelCase: Dict=2 , UpperCamelCase: Optional[int]=False , UpperCamelCase: Tuple=3 , UpperCamelCase: Any=2 , UpperCamelCase: List[Any]=3 , UpperCamelCase: int=None , UpperCamelCase: List[str]=None , **UpperCamelCase: List[Any] , ): """simple docstring""" super().__init__(**UpperCamelCase , pad_token_id=UpperCamelCase , bos_token_id=UpperCamelCase , eos_token_id=UpperCamelCase ) A__ = hidden_size A__ = feat_extract_norm A__ = feat_extract_activation A__ = list(UpperCamelCase ) A__ = list(UpperCamelCase ) A__ = list(UpperCamelCase ) A__ = conv_bias A__ = num_conv_pos_embeddings A__ = num_conv_pos_embedding_groups A__ = len(self.conv_dim ) A__ = num_hidden_layers A__ = intermediate_size A__ = hidden_act A__ = num_attention_heads A__ = hidden_dropout A__ = attention_dropout A__ = activation_dropout A__ = feat_proj_dropout A__ = final_dropout A__ = layerdrop A__ = layer_norm_eps A__ = initializer_range A__ = vocab_size A__ = do_stable_layer_norm A__ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" f""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 A__ = apply_spec_augment A__ = mask_time_prob A__ = mask_time_length A__ = mask_time_min_masks A__ = mask_feature_prob A__ = mask_feature_length A__ = mask_feature_min_masks # parameters for pretraining with codevector quantized representations A__ = num_codevectors_per_group A__ = num_codevector_groups A__ = contrastive_logits_temperature A__ = feat_quantizer_dropout A__ = num_negatives A__ = codevector_dim A__ = proj_codevector_dim A__ = diversity_loss_weight # ctc loss A__ = ctc_loss_reduction A__ = ctc_zero_infinity # adapter A__ = add_adapter A__ = adapter_kernel_size A__ = adapter_stride A__ = num_adapter_layers A__ = output_hidden_size or hidden_size A__ = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. A__ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. A__ = list(UpperCamelCase ) A__ = list(UpperCamelCase ) A__ = list(UpperCamelCase ) A__ = xvector_output_dim @property def UpperCamelCase ( self: Optional[Any] ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
69
"""simple docstring""" from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class a ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = 42 UpperCAmelCase = 42 if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
69
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig SCREAMING_SNAKE_CASE_: Union[str, Any] ={ 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/config.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/config.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/config.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/config.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/config.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/config.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/config.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/config.json', } class __A ( UpperCamelCase__ ): a__ : Optional[int] = """albert""" def __init__(self : Optional[int] , __a : Optional[int]=30000 , __a : Optional[Any]=128 , __a : Any=4096 , __a : Tuple=12 , __a : int=1 , __a : str=64 , __a : Optional[int]=16384 , __a : Optional[Any]=1 , __a : Dict="gelu_new" , __a : Dict=0 , __a : Dict=0 , __a : int=512 , __a : Optional[Any]=2 , __a : Tuple=0.02 , __a : Any=1E-12 , __a : List[Any]=0.1 , __a : Optional[Any]="absolute" , __a : Union[str, Any]=0 , __a : Union[str, Any]=2 , __a : str=3 , **__a : Union[str, Any] , ): super().__init__(pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , **__a ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = embedding_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_hidden_groups UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = inner_group_num UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = classifier_dropout_prob UpperCAmelCase_ = position_embedding_type class __A ( UpperCamelCase__ ): @property def _lowercase (self : Tuple ): if self.task == "multiple-choice": UpperCAmelCase_ = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
1
'''simple docstring''' from math import log from scipy.constants import Boltzmann, physical_constants SCREAMING_SNAKE_CASE_: Optional[int] =3_00 # TEMPERATURE (unit = K) def lowerCAmelCase_ ( snake_case_ : float , snake_case_ : float , snake_case_ : float , ) -> float: '''simple docstring''' if donor_conc <= 0: raise ValueError("Donor concentration should be positive" ) elif acceptor_conc <= 0: raise ValueError("Acceptor concentration should be positive" ) elif intrinsic_conc <= 0: raise ValueError("Intrinsic concentration should be positive" ) elif donor_conc <= intrinsic_conc: raise ValueError( "Donor concentration should be greater than intrinsic concentration" ) elif acceptor_conc <= intrinsic_conc: raise ValueError( "Acceptor concentration should be greater than intrinsic concentration" ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
1
1
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class A__ : def __init__( self , A_ , A_=99 , A_=13 , A_=7 , A_=9 , A_=True , A_=True , A_=False , A_=32 , A_=5 , A_=4 , A_=37 , A_=8 , A_=0.1 , A_=0.0_02 , A_=1 , A_=0 , A_=0 , A_=None , A_=None , ): '''simple docstring''' UpperCamelCase : Dict = parent UpperCamelCase : List[Any] = batch_size UpperCamelCase : List[Any] = encoder_seq_length UpperCamelCase : Tuple = decoder_seq_length # For common tests UpperCamelCase : Tuple = self.decoder_seq_length UpperCamelCase : Any = is_training UpperCamelCase : int = use_attention_mask UpperCamelCase : Optional[int] = use_labels UpperCamelCase : List[Any] = vocab_size UpperCamelCase : Union[str, Any] = hidden_size UpperCamelCase : List[str] = num_hidden_layers UpperCamelCase : List[Any] = num_attention_heads UpperCamelCase : Any = d_ff UpperCamelCase : List[Any] = relative_attention_num_buckets UpperCamelCase : List[Any] = dropout_rate UpperCamelCase : Union[str, Any] = initializer_factor UpperCamelCase : List[str] = eos_token_id UpperCamelCase : Optional[Any] = pad_token_id UpperCamelCase : Optional[int] = decoder_start_token_id UpperCamelCase : Optional[int] = None UpperCamelCase : str = decoder_layers def __UpperCamelCase( self ): '''simple docstring''' return TaConfig.from_pretrained("google/umt5-base" ) def __UpperCamelCase( self , A_ , A_ , A_ , A_=None , A_=None , A_=None , A_=None , A_=None , ): '''simple docstring''' if attention_mask is None: UpperCamelCase : int = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: UpperCamelCase : List[Any] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: UpperCamelCase : Any = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=A_ ) if decoder_head_mask is None: UpperCamelCase : Dict = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=A_ ) if cross_attn_head_mask is None: UpperCamelCase : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=A_ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) UpperCamelCase : Dict = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input UpperCamelCase : int = input_ids.clamp(self.pad_token_id + 1 ) UpperCamelCase : Dict = decoder_input_ids.clamp(self.pad_token_id + 1 ) UpperCamelCase : Dict = self.get_config() UpperCamelCase : Any = config.num_attention_heads UpperCamelCase : Optional[int] = self.prepare_inputs_dict(A_ , A_ , A_ ) return config, input_dict def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.prepare_config_and_inputs() return config, inputs_dict def __UpperCamelCase( self ): '''simple docstring''' return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def __UpperCamelCase( self ): '''simple docstring''' return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , ): '''simple docstring''' UpperCamelCase : Dict = UMTaModel(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : int = model( input_ids=A_ , decoder_input_ids=A_ , attention_mask=A_ , decoder_attention_mask=A_ , ) UpperCamelCase : str = model(input_ids=A_ , decoder_input_ids=A_ ) UpperCamelCase : Dict = result.last_hidden_state UpperCamelCase : Tuple = result.past_key_values UpperCamelCase : List[str] = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(A_ ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , ): '''simple docstring''' UpperCamelCase : Dict = UMTaModel(config=A_ ).get_decoder().to(A_ ).eval() # first forward pass UpperCamelCase : Optional[Any] = model(A_ , use_cache=A_ ) UpperCamelCase : List[Any] = model(A_ ) UpperCamelCase : List[Any] = model(A_ , use_cache=A_ ) self.parent.assertTrue(len(A_ ) == len(A_ ) ) self.parent.assertTrue(len(A_ ) == len(A_ ) + 1 ) UpperCamelCase : int = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCamelCase : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and UpperCamelCase : Any = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCamelCase : Union[str, Any] = model(A_ )["last_hidden_state"] UpperCamelCase : List[str] = model(A_ , past_key_values=A_ )["last_hidden_state"] # select random slice UpperCamelCase : Optional[int] = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCamelCase : List[str] = output_from_no_past[:, -1, random_slice_idx].detach() UpperCamelCase : Dict = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A_ , A_ , atol=1e-3 ) ) def __UpperCamelCase( self , A_ , A_ , ): '''simple docstring''' UpperCamelCase : Any = UMTaModel(config=A_ ).to(A_ ).half().eval() UpperCamelCase : int = model(**A_ )["last_hidden_state"] self.parent.assertFalse(torch.isnan(A_ ).any().item() ) @require_torch class A__ ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :str = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) _UpperCAmelCase :Union[str, Any] = (UMTaForConditionalGeneration,) if is_torch_available() else () _UpperCAmelCase :Dict = ( { 'conversational': UMTaForConditionalGeneration, 'feature-extraction': UMTaModel, 'summarization': UMTaForConditionalGeneration, 'text2text-generation': UMTaForConditionalGeneration, 'translation': UMTaForConditionalGeneration, 'question-answering': UMTaForQuestionAnswering, } if is_torch_available() else {} ) _UpperCAmelCase :Dict = True _UpperCAmelCase :Any = False _UpperCAmelCase :Optional[int] = False _UpperCAmelCase :List[Any] = True _UpperCAmelCase :Any = True # The small UMT5 model needs higher percentages for CPU/MP tests _UpperCAmelCase :Optional[int] = [0.8, 0.9] def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = UMTaModelTester(self ) @unittest.skip("Test has a segmentation fault on torch 1.8.0" ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() UpperCamelCase : Optional[Any] = UMTaModel(config_and_inputs[0] ).to(A_ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( A_ , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"""{tmpdirname}/t5_test.onnx""" , export_params=A_ , opset_version=9 , input_names=["input_ids", "decoder_input_ids"] , ) @unittest.skipIf(torch_device == "cpu" , "Cant do half precision" ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = ["encoder_attentions", "decoder_attentions", "cross_attentions"] UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() UpperCamelCase : Tuple = config_and_inputs[0] UpperCamelCase : Tuple = UMTaForConditionalGeneration(A_ ).eval() model.to(A_ ) UpperCamelCase : Any = { "head_mask": torch.zeros(config.num_layers , config.num_heads , device=A_ ), "decoder_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=A_ ), "cross_attn_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=A_ ), } for attn_name, (name, mask) in zip(A_ , head_masking.items() ): UpperCamelCase : Optional[int] = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": UpperCamelCase : List[str] = torch.ones( config.num_decoder_layers , config.num_heads , device=A_ ) UpperCamelCase : List[str] = model.generate( config_and_inputs[1]["input_ids"] , num_beams=1 , max_length=3 , output_attentions=A_ , return_dict_in_generate=A_ , **A_ , ) # We check the state of decoder_attentions and cross_attentions just from the last step UpperCamelCase : Any = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip("Does not work on the tiny model as we keep hitting edge cases." ) def __UpperCamelCase( self ): '''simple docstring''' pass @require_torch @require_sentencepiece @require_tokenizers class A__ ( unittest.TestCase ): @slow @unittest.skip( "Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged" ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = UMTaForConditionalGeneration.from_pretrained("google/umt5-small" , return_dict=A_ ).to(A_ ) UpperCamelCase : List[str] = AutoTokenizer.from_pretrained("google/umt5-small" , use_fast=A_ , legacy=A_ ) UpperCamelCase : Dict = [ "Bonjour monsieur <extra_id_0> bien <extra_id_1>.", "No se como puedo <extra_id_0>.", "This is the reason why we <extra_id_0> them.", "The <extra_id_0> walks in <extra_id_1>, seats", "A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.", ] UpperCamelCase : Optional[Any] = tokenizer(A_ , return_tensors="pt" , padding=A_ ).input_ids # fmt: off UpperCamelCase : int = torch.tensor( [ [ 3_8530, 21_0703, 25_6299, 1410, 25_6298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 2_5922, 25_6299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 1_9014, 1_0620, 758, 25_6299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 25_6299, 1_4869, 281, 301, 25_6298, 275, 11_9983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 25_6299, 1_4869, 281, 2234, 289, 2275, 333,6_1391, 289, 25_6298, 543, 25_6297, 16_8714, 329, 25_6296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(A_ , A_ ) UpperCamelCase : str = model.generate(input_ids.to(A_ ) ) UpperCamelCase : str = [ "<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>", "<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", ] UpperCamelCase : List[Any] = tokenizer.batch_decode(A_ ) self.assertEqual(A_ , A_ )
365
from scipy.stats import spearmanr import datasets __lowerCamelCase : List[str] = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. """ __lowerCamelCase : List[Any] = """ Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {'spearmanr': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results['spearmanr']) -0.7 >>> print(round(results['spearmanr_pvalue'], 2)) 0.19 """ __lowerCamelCase : Optional[int] = r"""\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): def __UpperCamelCase( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def __UpperCamelCase( self , A_ , A_ , A_=False ): '''simple docstring''' UpperCamelCase : Tuple = spearmanr(A_ , A_ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
140
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __A : List[str] = { "configuration_mobilebert": [ "MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MobileBertConfig", "MobileBertOnnxConfig", ], "tokenization_mobilebert": ["MobileBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[str] = ["MobileBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Union[str, Any] = [ "MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "MobileBertForMaskedLM", "MobileBertForMultipleChoice", "MobileBertForNextSentencePrediction", "MobileBertForPreTraining", "MobileBertForQuestionAnswering", "MobileBertForSequenceClassification", "MobileBertForTokenClassification", "MobileBertLayer", "MobileBertModel", "MobileBertPreTrainedModel", "load_tf_weights_in_mobilebert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[str] = [ "TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFMobileBertForMaskedLM", "TFMobileBertForMultipleChoice", "TFMobileBertForNextSentencePrediction", "TFMobileBertForPreTraining", "TFMobileBertForQuestionAnswering", "TFMobileBertForSequenceClassification", "TFMobileBertForTokenClassification", "TFMobileBertMainLayer", "TFMobileBertModel", "TFMobileBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys __A : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
120
'''simple docstring''' from __future__ import annotations from typing import Any class __snake_case ( _SCREAMING_SNAKE_CASE): """simple docstring""" pass class __snake_case : """simple docstring""" def __init__( self : Union[str, Any] , lowerCamelCase : Any ) -> None: lowerCAmelCase_ : Any = data lowerCAmelCase_ : Node | None = None def __iter__( self : Union[str, Any] ) -> Optional[Any]: lowerCAmelCase_ : Union[str, Any] = self lowerCAmelCase_ : Any = [] while node: if node in visited: raise ContainsLoopError visited.append(lowerCamelCase ) yield node.data lowerCAmelCase_ : int = node.next_node @property def __lowercase ( self : str ) -> bool: try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": __A : Dict = Node(1) __A : Optional[Any] = Node(2) __A : int = Node(3) __A : Optional[Any] = Node(4) print(root_node.has_loop) # False __A : Any = root_node.next_node print(root_node.has_loop) # True __A : List[Any] = Node(5) __A : Dict = Node(6) __A : str = Node(5) __A : Dict = Node(6) print(root_node.has_loop) # False __A : Optional[int] = Node(1) print(root_node.has_loop) # False
120
1
from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowercase_ ( __snake_case ): _lowerCamelCase = 'new-model' if is_tf_available(): class lowercase_ ( __snake_case ): _lowerCamelCase = NewModelConfig @require_tf class lowercase_ ( unittest.TestCase ): @slow def UpperCamelCase ( self ): _snake_case : Dict = "bert-base-cased" _snake_case : str = AutoConfig.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : Optional[int] = TFAutoModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) @slow def UpperCamelCase ( self ): _snake_case : Optional[int] = "bert-base-cased" _snake_case : List[Any] = AutoConfig.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : List[str] = TFAutoModelForPreTraining.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) @slow def UpperCamelCase ( self ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case : List[Any] = AutoConfig.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : int = TFAutoModelForCausalLM.from_pretrained(lowercase_ ) _snake_case ,_snake_case : int = TFAutoModelForCausalLM.from_pretrained(lowercase_ , output_loading_info=lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) @slow def UpperCamelCase ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case : Optional[Any] = AutoConfig.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : Dict = TFAutoModelWithLMHead.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) @slow def UpperCamelCase ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case : Optional[Any] = AutoConfig.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : Dict = TFAutoModelForMaskedLM.from_pretrained(lowercase_ ) _snake_case ,_snake_case : str = TFAutoModelForMaskedLM.from_pretrained(lowercase_ , output_loading_info=lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) @slow def UpperCamelCase ( self ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case : int = AutoConfig.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase_ ) _snake_case ,_snake_case : Dict = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase_ , output_loading_info=lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) @slow def UpperCamelCase ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _snake_case : List[Any] = AutoConfig.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : Any = TFAutoModelForSequenceClassification.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) @slow def UpperCamelCase ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _snake_case : int = AutoConfig.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : List[Any] = TFAutoModelForQuestionAnswering.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) @slow @require_tensorflow_probability def UpperCamelCase ( self ): for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: _snake_case : Optional[int] = AutoConfig.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : Any = TFAutoModelForTableQuestionAnswering.from_pretrained(lowercase_ ) _snake_case ,_snake_case : List[Any] = TFAutoModelForTableQuestionAnswering.from_pretrained( lowercase_ , output_loading_info=lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) def UpperCamelCase ( self ): _snake_case : Optional[Any] = TFAutoModelWithLMHead.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase_ ) , 14_410 ) def UpperCamelCase ( self ): _snake_case : Dict = TFAutoModelWithLMHead.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase_ ) , 14_410 ) def UpperCamelCase ( self ): # For the auto model mapping, FunnelConfig has two models: FunnelModel and FunnelBaseModel _snake_case : List[Any] = TFAutoModel.from_pretrained("sgugger/funnel-random-tiny" ) self.assertIsInstance(lowercase_ , lowercase_ ) _snake_case : Optional[int] = copy.deepcopy(model.config ) _snake_case : str = ["FunnelBaseModel"] _snake_case : Any = TFAutoModel.from_config(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase_ ) _snake_case : Optional[Any] = TFAutoModel.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) def UpperCamelCase ( self ): try: AutoConfig.register("new-model" , lowercase_ ) _snake_case : Optional[Any] = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowercase_ ): auto_class.register(lowercase_ , lowercase_ ) auto_class.register(lowercase_ , lowercase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase_ ): auto_class.register(lowercase_ , lowercase_ ) # Now that the config is registered, it can be used as any other config with the auto-API _snake_case : str = BertModelTester(self ).get_config() _snake_case : int = NewModelConfig(**tiny_config.to_dict() ) _snake_case : str = auto_class.from_config(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase_ ) _snake_case : Optional[Any] = auto_class.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def UpperCamelCase ( self ): with self.assertRaisesRegex( lowercase_ , "bert-base is not a local folder and is not a valid model identifier" ): _snake_case : Tuple = TFAutoModel.from_pretrained("bert-base" ) def UpperCamelCase ( self ): with self.assertRaisesRegex( lowercase_ , R"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): _snake_case : List[Any] = TFAutoModel.from_pretrained(lowercase_ , revision="aaaaaa" ) def UpperCamelCase ( self ): with self.assertRaisesRegex( lowercase_ , "hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin" , ): _snake_case : Union[str, Any] = TFAutoModel.from_pretrained("hf-internal-testing/config-no-model" ) def UpperCamelCase ( self ): with self.assertRaisesRegex(lowercase_ , "Use `from_pt=True` to load this model" ): _snake_case : Union[str, Any] = TFAutoModel.from_pretrained("hf-internal-testing/tiny-bert-pt-only" ) def UpperCamelCase ( self ): # Make sure we have cached the model. _snake_case : Union[str, Any] = TFAutoModel.from_pretrained("hf-internal-testing/tiny-random-bert" ) with RequestCounter() as counter: _snake_case : Optional[Any] = TFAutoModel.from_pretrained("hf-internal-testing/tiny-random-bert" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint _snake_case : Dict = TFAutoModel.from_pretrained("ArthurZ/tiny-random-bert-sharded" ) with RequestCounter() as counter: _snake_case : Tuple = TFAutoModel.from_pretrained("ArthurZ/tiny-random-bert-sharded" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
284
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = ['pixel_values'] def __init__( self , lowercase_ = True , lowercase_ = None , lowercase_ = PILImageResampling.BILINEAR , lowercase_ = True , lowercase_ = 1 / 255 , lowercase_ = True , lowercase_ = None , lowercase_ = True , **lowercase_ , ): super().__init__(**lowercase_ ) _snake_case : Dict = size if size is not None else {"shortest_edge": 224} _snake_case : Optional[int] = get_size_dict(lowercase_ , default_to_square=lowercase_ ) _snake_case : Any = crop_size if crop_size is not None else {"height": 256, "width": 256} _snake_case : str = get_size_dict(lowercase_ , param_name="crop_size" ) _snake_case : List[Any] = do_resize _snake_case : Tuple = size _snake_case : Union[str, Any] = resample _snake_case : str = do_rescale _snake_case : Dict = rescale_factor _snake_case : int = do_center_crop _snake_case : int = crop_size _snake_case : List[Any] = do_flip_channel_order def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ = PIL.Image.BILINEAR , lowercase_ = None , **lowercase_ , ): _snake_case : Optional[Any] = get_size_dict(lowercase_ , default_to_square=lowercase_ ) if "shortest_edge" not in size: raise ValueError(f"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) _snake_case : int = get_resize_output_image_size(lowercase_ , size=size["shortest_edge"] , default_to_square=lowercase_ ) return resize(lowercase_ , size=lowercase_ , resample=lowercase_ , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ = None , **lowercase_ , ): _snake_case : List[str] = get_size_dict(lowercase_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(lowercase_ , size=(size["height"], size["width"]) , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ = None , **lowercase_ , ): return rescale(lowercase_ , scale=lowercase_ , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ = None ): return flip_channel_order(lowercase_ , data_format=lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = ChannelDimension.FIRST , **lowercase_ , ): _snake_case : Dict = do_resize if do_resize is not None else self.do_resize _snake_case : Union[str, Any] = resample if resample is not None else self.resample _snake_case : Dict = do_rescale if do_rescale is not None else self.do_rescale _snake_case : str = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case : List[str] = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) _snake_case : Optional[int] = size if size is not None else self.size _snake_case : Optional[int] = get_size_dict(lowercase_ , default_to_square=lowercase_ ) _snake_case : Optional[int] = crop_size if crop_size is not None else self.crop_size _snake_case : Union[str, Any] = get_size_dict(lowercase_ , param_name="crop_size" ) _snake_case : Tuple = make_list_of_images(lowercase_ ) if not valid_images(lowercase_ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) # All transformations expect numpy arrays. _snake_case : Tuple = [to_numpy_array(lowercase_ ) for image in images] if do_resize: _snake_case : int = [self.resize(image=lowercase_ , size=lowercase_ , resample=lowercase_ ) for image in images] if do_center_crop: _snake_case : List[Any] = [self.center_crop(image=lowercase_ , size=lowercase_ ) for image in images] if do_rescale: _snake_case : Optional[int] = [self.rescale(image=lowercase_ , scale=lowercase_ ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: _snake_case : int = [self.flip_channel_order(image=lowercase_ ) for image in images] _snake_case : str = [to_channel_dimension_format(lowercase_ , lowercase_ ) for image in images] _snake_case : List[Any] = {"pixel_values": images} return BatchFeature(data=lowercase_ , tensor_type=lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ = None ): _snake_case : int = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(lowercase_ ) != len(lowercase_ ): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits" ) if is_torch_tensor(lowercase_ ): _snake_case : Union[str, Any] = target_sizes.numpy() _snake_case : int = [] for idx in range(len(lowercase_ ) ): _snake_case : Optional[Any] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="bilinear" , align_corners=lowercase_ ) _snake_case : List[str] = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(lowercase_ ) else: _snake_case : List[Any] = logits.argmax(dim=1 ) _snake_case : Optional[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
284
1
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import BnbQuantizationConfig from .modeling import ( find_tied_parameters, get_balanced_memory, infer_auto_device_map, load_checkpoint_in_model, offload_weight, set_module_tensor_to_device, ) if is_bnb_available(): import bitsandbytes as bnb from copy import deepcopy UpperCAmelCase_ : Any = logging.getLogger(__name__) def SCREAMING_SNAKE_CASE_ ( __A : torch.nn.Module , __A : BnbQuantizationConfig , __A : Union[str, os.PathLike] = None , __A : Optional[Dict[str, Union[int, str, torch.device]]] = None , __A : Optional[List[str]] = None , __A : Optional[Dict[Union[int, str], Union[int, str]]] = None , __A : Optional[Union[str, os.PathLike]] = None , __A : bool = False , ) -> Dict: """simple docstring""" a_ : Optional[int] = bnb_quantization_config.load_in_abit a_ : Any = bnb_quantization_config.load_in_abit if load_in_abit and not is_abit_bnb_available(): raise ImportError( 'You have a version of `bitsandbytes` that is not compatible with 8bit quantization,' ' make sure you have the latest version of `bitsandbytes` installed.' ) if load_in_abit and not is_abit_bnb_available(): raise ValueError( 'You have a version of `bitsandbytes` that is not compatible with 4bit quantization,' 'make sure you have the latest version of `bitsandbytes` installed.' ) a_ : List[Any] = [] # custom device map if isinstance(__A , __A ) and len(device_map.keys() ) > 1: a_ : Optional[int] = [key for key, value in device_map.items() if value in ['disk', 'cpu']] # We keep some modules such as the lm_head in their original dtype for numerical stability reasons if bnb_quantization_config.skip_modules is None: a_ : List[str] = get_keys_to_not_convert(__A ) # add cpu modules to skip modules only for 4-bit modules if load_in_abit: bnb_quantization_config.skip_modules.extend(__A ) a_ : Union[str, Any] = bnb_quantization_config.skip_modules # We add the modules we want to keep in full precision if bnb_quantization_config.keep_in_fpaa_modules is None: a_ : List[Any] = [] a_ : str = bnb_quantization_config.keep_in_fpaa_modules modules_to_not_convert.extend(__A ) # compatibility with peft a_ : Any = load_in_abit a_ : List[Any] = load_in_abit a_ : List[Any] = get_parameter_device(__A ) if model_device.type != "meta": # quantization of an already loaded model logger.warning( 'It is not recommended to quantize a loaded model. ' 'The model should be instantiated under the `init_empty_weights` context manager.' ) a_ : Any = replace_with_bnb_layers(__A , __A , modules_to_not_convert=__A ) # convert param to the right dtype a_ : Union[str, Any] = bnb_quantization_config.torch_dtype for name, param in model.state_dict().items(): if any(module_to_keep_in_fpaa in name for module_to_keep_in_fpaa in keep_in_fpaa_modules ): param.to(torch.floataa ) if param.dtype != torch.floataa: a_ : int = name.replace('.weight' , '' ).replace('.bias' , '' ) a_ : int = getattr(__A , __A , __A ) if param is not None: param.to(torch.floataa ) elif torch.is_floating_point(__A ): param.to(__A ) if model_device.type == "cuda": # move everything to cpu in the first place because we can't do quantization if the weights are already on cuda model.cuda(torch.cuda.current_device() ) torch.cuda.empty_cache() elif torch.cuda.is_available(): model.to(torch.cuda.current_device() ) else: raise RuntimeError('No GPU found. A GPU is needed for quantization.' ) logger.info( F"""The model device type is {model_device.type}. However, cuda is needed for quantization.""" 'We move the model to cuda.' ) return model elif weights_location is None: raise RuntimeError( F"""`weights_location` needs to be the folder path containing the weights of the model, but we found {weights_location} """ ) else: with init_empty_weights(): a_ : List[Any] = replace_with_bnb_layers( __A , __A , modules_to_not_convert=__A ) a_ : Any = get_quantized_model_device_map( __A , __A , __A , max_memory=__A , no_split_module_classes=__A , ) if offload_state_dict is None and device_map is not None and "disk" in device_map.values(): a_ : Dict = True a_ : Any = any(x in list(device_map.values() ) for x in ['cpu', 'disk'] ) load_checkpoint_in_model( __A , __A , __A , dtype=bnb_quantization_config.torch_dtype , offload_folder=__A , offload_state_dict=__A , keep_in_fpaa_modules=bnb_quantization_config.keep_in_fpaa_modules , offload_abit_bnb=load_in_abit and offload , ) return dispatch_model(__A , device_map=__A , offload_dir=__A ) def SCREAMING_SNAKE_CASE_ ( __A : List[str] , __A : Dict , __A : List[Any]=None , __A : Union[str, Any]=None , __A : Optional[Any]=None ) -> str: """simple docstring""" if device_map is None: if torch.cuda.is_available(): a_ : Dict = {'': torch.cuda.current_device()} else: raise RuntimeError('No GPU found. A GPU is needed for quantization.' ) logger.info('The device_map was not initialized.' 'Setting device_map to `{\'\':torch.cuda.current_device()}`.' ) if isinstance(__A , __A ): if device_map not in ["auto", "balanced", "balanced_low_0", "sequential"]: raise ValueError( 'If passing a string for `device_map`, please choose \'auto\', \'balanced\', \'balanced_low_0\' or ' '\'sequential\'.' ) a_ : Any = {} special_dtypes.update( { name: bnb_quantization_config.torch_dtype for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.skip_modules ) } ) special_dtypes.update( { name: torch.floataa for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.keep_in_fpaa_modules ) } ) a_ : Optional[Any] = {} a_ : Union[str, Any] = special_dtypes a_ : Optional[int] = no_split_module_classes a_ : str = bnb_quantization_config.target_dtype # get max_memory for each device. if device_map != "sequential": a_ : Optional[Any] = get_balanced_memory( __A , low_zero=(device_map == 'balanced_low_0') , max_memory=__A , **__A , ) a_ : int = max_memory a_ : Optional[int] = infer_auto_device_map(__A , **__A ) if isinstance(__A , __A ): # check if don't have any quantized module on the cpu a_ : str = bnb_quantization_config.skip_modules + bnb_quantization_config.keep_in_fpaa_modules a_ : Optional[Any] = { key: device_map[key] for key in device_map.keys() if key not in modules_not_to_convert } for device in ["cpu", "disk"]: if device in device_map_without_some_modules.values(): if bnb_quantization_config.load_in_abit: raise ValueError( '\n Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit\n the quantized model. If you want to dispatch the model on the CPU or the disk while keeping\n these modules in `torch_dtype`, you need to pass a custom `device_map` to\n `load_and_quantize_model`. Check\n https://huggingface.co/docs/accelerate/main/en/usage_guides/quantization#offload-modules-to-cpu-and-disk\n for more details.\n ' ) else: logger.info( 'Some modules are are offloaded to the CPU or the disk. Note that these modules will be converted to 8-bit' ) del device_map_without_some_modules return device_map def SCREAMING_SNAKE_CASE_ ( __A : Tuple , __A : List[str] , __A : List[Any]=None , __A : List[Any]=None ) -> Any: """simple docstring""" if modules_to_not_convert is None: a_ : Union[str, Any] = [] a_ , a_ : List[str] = _replace_with_bnb_layers( __A , __A , __A , __A ) if not has_been_replaced: logger.warning( 'You are loading your model in 8bit or 4bit but no linear modules were found in your model.' ' this can happen for some architectures such as gpt2 that uses Conv1D instead of Linear layers.' ' Please double check your model architecture, or submit an issue on github if you think this is' ' a bug.' ) return model def SCREAMING_SNAKE_CASE_ ( __A : int , __A : int , __A : Tuple=None , __A : Any=None , ) -> Any: """simple docstring""" a_ : int = False for name, module in model.named_children(): if current_key_name is None: a_ : List[str] = [] current_key_name.append(__A ) if isinstance(__A , nn.Linear ) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` a_ : Union[str, Any] = '.'.join(__A ) a_ : Any = True for key in modules_to_not_convert: if ( (key in current_key_name_str) and (key + "." in current_key_name_str) ) or key == current_key_name_str: a_ : int = False break if proceed: # Load bnb module with empty weight and replace ``nn.Linear` module if bnb_quantization_config.load_in_abit: a_ : Union[str, Any] = bnb.nn.LinearabitLt( module.in_features , module.out_features , module.bias is not None , has_fpaa_weights=__A , threshold=bnb_quantization_config.llm_inta_threshold , ) elif bnb_quantization_config.load_in_abit: a_ : Optional[int] = bnb.nn.Linearabit( module.in_features , module.out_features , module.bias is not None , bnb_quantization_config.bnb_abit_compute_dtype , compress_statistics=bnb_quantization_config.bnb_abit_use_double_quant , quant_type=bnb_quantization_config.bnb_abit_quant_type , ) else: raise ValueError('load_in_8bit and load_in_4bit can\'t be both False' ) a_ : Union[str, Any] = module.weight.data if module.bias is not None: a_ : str = module.bias.data bnb_module.requires_grad_(__A ) setattr(__A , __A , __A ) a_ : int = True if len(list(module.children() ) ) > 0: a_ , a_ : List[str] = _replace_with_bnb_layers( __A , __A , __A , __A ) a_ : List[str] = has_been_replaced | _has_been_replaced # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def SCREAMING_SNAKE_CASE_ ( __A : Tuple ) -> Optional[int]: """simple docstring""" with init_empty_weights(): a_ : List[str] = deepcopy(__A ) # this has 0 cost since it is done inside `init_empty_weights` context manager` a_ : List[str] = find_tied_parameters(__A ) # For compatibility with Accelerate < 0.18 if isinstance(__A , __A ): a_ : Union[str, Any] = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: a_ : Any = sum(__A , [] ) a_ : Optional[int] = len(__A ) > 0 # Check if it is a base model a_ : Union[str, Any] = False if hasattr(__A , 'base_model_prefix' ): a_ : Union[str, Any] = not hasattr(__A , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head a_ : Tuple = list(model.named_children() ) a_ : str = [list_modules[-1][0]] # add last module together with tied weights a_ : List[str] = set(__A ) - set(__A ) a_ : Dict = list(set(__A ) ) + list(__A ) # remove ".weight" from the keys a_ : List[str] = ['.weight', '.bias'] a_ : List[Any] = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: a_ : Dict = name.replace(__A , '' ) filtered_module_names.append(__A ) return filtered_module_names def SCREAMING_SNAKE_CASE_ ( __A : Any ) -> Any: """simple docstring""" for m in model.modules(): if isinstance(__A , bnb.nn.Linearabit ): return True return False def SCREAMING_SNAKE_CASE_ ( __A : nn.Module ) -> Union[str, Any]: """simple docstring""" return next(parameter.parameters() ).device def SCREAMING_SNAKE_CASE_ ( __A : Optional[int] , __A : int , __A : Optional[Any] , __A : List[str] , __A : Dict , __A : Tuple , __A : Optional[Any] ) -> Dict: """simple docstring""" if fpaa_statistics is None: set_module_tensor_to_device(__A , __A , 0 , dtype=__A , value=__A ) a_ : Optional[int] = param_name a_ : List[str] = model if "." in tensor_name: a_ : int = tensor_name.split('.' ) for split in splits[:-1]: a_ : int = getattr(__A , __A ) if new_module is None: raise ValueError(F"""{module} has no attribute {split}.""" ) a_ : Optional[Any] = new_module a_ : List[str] = splits[-1] # offload weights a_ : Union[str, Any] = False offload_weight(module._parameters[tensor_name] , __A , __A , index=__A ) if hasattr(module._parameters[tensor_name] , 'SCB' ): offload_weight( module._parameters[tensor_name].SCB , param_name.replace('weight' , 'SCB' ) , __A , index=__A , ) else: offload_weight(__A , __A , __A , index=__A ) offload_weight(__A , param_name.replace('weight' , 'SCB' ) , __A , index=__A ) set_module_tensor_to_device(__A , __A , 'meta' , dtype=__A , value=torch.empty(*param.size() ) )
32
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Optional import pandas as pd import pyarrow as pa import datasets from datasets.table import table_cast @dataclass class __UpperCAmelCase ( datasets.BuilderConfig ): '''simple docstring''' __lowerCAmelCase = None class __UpperCAmelCase ( datasets.ArrowBasedBuilder ): '''simple docstring''' __lowerCAmelCase = PandasConfig def A (self : Tuple ): return datasets.DatasetInfo(features=self.config.features ) def A (self : Optional[int] , _lowerCAmelCase : List[Any] ): if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) A = dl_manager.download_and_extract(self.config.data_files ) if isinstance(_lowerCAmelCase , (str, list, tuple) ): A = data_files if isinstance(_lowerCAmelCase , _lowerCAmelCase ): A = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive A = [dl_manager.iter_files(_lowerCAmelCase ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] A = [] for split_name, files in data_files.items(): if isinstance(_lowerCAmelCase , _lowerCAmelCase ): A = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive A = [dl_manager.iter_files(_lowerCAmelCase ) for file in files] splits.append(datasets.SplitGenerator(name=_lowerCAmelCase , gen_kwargs={"""files""": files} ) ) return splits def A (self : Dict , _lowerCAmelCase : pa.Table ): if self.config.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example A = table_cast(_lowerCAmelCase , self.config.features.arrow_schema ) return pa_table def A (self : List[Any] , _lowerCAmelCase : Optional[Any] ): for i, file in enumerate(itertools.chain.from_iterable(_lowerCAmelCase ) ): with open(_lowerCAmelCase , """rb""" ) as f: A = pa.Table.from_pandas(pd.read_pickle(_lowerCAmelCase ) ) yield i, self._cast_table(_lowerCAmelCase )
258
0
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem __SCREAMING_SNAKE_CASE : Any = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 __SCREAMING_SNAKE_CASE : List[compression.BaseCompressedFileFileSystem] = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(F'''A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.''') fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def snake_case (__lowercase ) -> str: '''simple docstring''' if "://" in dataset_path: _snake_case : Tuple = dataset_path.split("://" )[1] return dataset_path def snake_case (__lowercase ) -> bool: '''simple docstring''' if fs is not None and fs.protocol != "file": return True else: return False def snake_case (__lowercase , __lowercase , __lowercase ) -> str: '''simple docstring''' _snake_case : Optional[int] = not is_remote_filesystem(__lowercase ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(__lowercase ) , fs._strip_protocol(__lowercase ) ) else: fs.mv(__lowercase , __lowercase , recursive=__lowercase ) def snake_case () -> None: '''simple docstring''' if hasattr(fsspec.asyn , "reset_lock" ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: _snake_case : List[str] = None _snake_case : str = None _snake_case : List[Any] = threading.Lock()
284
import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class lowercase_ ( __snake_case ): _lowerCamelCase = 'M-CLIP' def __init__( self , lowercase_=1_024 , lowercase_=768 , **lowercase_ ): _snake_case : str = transformerDimSize _snake_case : Union[str, Any] = imageDimSize super().__init__(**lowercase_ ) class lowercase_ ( __snake_case ): _lowerCamelCase = MCLIPConfig def __init__( self , lowercase_ , *lowercase_ , **lowercase_ ): super().__init__(lowercase_ , *lowercase_ , **lowercase_ ) _snake_case : List[Any] = XLMRobertaModel(lowercase_ ) _snake_case : int = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def UpperCamelCase ( self , lowercase_ , lowercase_ ): _snake_case : Tuple = self.transformer(input_ids=lowercase_ , attention_mask=lowercase_ )[0] _snake_case : Tuple = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(lowercase_ ), embs
284
1
def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> int: return int((input_a, input_a).count(1 ) != 0 ) def SCREAMING_SNAKE_CASE__ ( ) -> None: assert or_gate(0 ,0 ) == 0 assert or_gate(0 ,1 ) == 1 assert or_gate(1 ,0 ) == 1 assert or_gate(1 ,1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
124
def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> str: if number < 0 or shift_amount < 0: raise ValueError("""both inputs must be positive integers""" ) snake_case : Optional[int] = str(bin(lowercase ) ) binary_number += "0" * shift_amount return binary_number def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> str: if number < 0 or shift_amount < 0: raise ValueError("""both inputs must be positive integers""" ) snake_case : Dict = str(bin(lowercase ) )[2:] if shift_amount >= len(lowercase ): return "0b0" snake_case : str = binary_number[: len(lowercase ) - shift_amount] return "0b" + shifted_binary_number def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> str: if number >= 0: # Get binary representation of positive number snake_case : Optional[Any] = """0""" + str(bin(lowercase ) ).strip("""-""" )[2:] else: # Get binary (2's complement) representation of negative number snake_case : Dict = len(bin(lowercase )[3:] ) # Find 2's complement of number snake_case : Optional[Any] = bin(abs(lowercase ) - (1 << binary_number_length) )[3:] snake_case : Tuple = ( """1""" + """0""" * (binary_number_length - len(lowercase )) + binary_number ) if shift_amount >= len(lowercase ): return "0b" + binary_number[0] * len(lowercase ) return ( "0b" + binary_number[0] * shift_amount + binary_number[: len(lowercase ) - shift_amount] ) if __name__ == "__main__": import doctest doctest.testmod()
124
1
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class snake_case__(pl.LightningModule ): """simple docstring""" def __init__( self : str , SCREAMING_SNAKE_CASE : Optional[Any] ): super().__init__() lowercase__ : Union[str, Any] = model lowercase__ : Tuple = 2 lowercase__ : List[str] = nn.Linear(self.model.config.hidden_size , self.num_labels ) def snake_case ( self : List[str] ): pass def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" lowercase__ : List[str] = LongformerModel.from_pretrained(_UpperCAmelCase ) lowercase__ : int = LightningModel(_UpperCAmelCase ) lowercase__ : Union[str, Any] = torch.load(_UpperCAmelCase , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model lowercase__ : Any = LongformerForQuestionAnswering.from_pretrained(_UpperCAmelCase ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(_UpperCAmelCase ) print(F"""Conversion successful. Model saved under {pytorch_dump_folder_path}""" ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
359
from math import ceil, sqrt def __lowerCamelCase ( lowerCamelCase__ = 1_000_000 ): """simple docstring""" lowercase__ : int = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: lowercase__ : List[str] = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: lowercase__ : List[str] = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(f'''{solution() = }''')
121
0
import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset snake_case_ : Any = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class __snake_case ( nn.Module ): def __init__( self : Dict , _snake_case : int): """simple docstring""" super().__init__() UpperCAmelCase_ = torchvision.models.resnetaaa(pretrained=_SCREAMING_SNAKE_CASE) UpperCAmelCase_ = list(model.children())[:-2] UpperCAmelCase_ = nn.Sequential(*_SCREAMING_SNAKE_CASE) UpperCAmelCase_ = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds]) def lowerCamelCase ( self : Tuple , _snake_case : Any): """simple docstring""" UpperCAmelCase_ = self.pool(self.model(_SCREAMING_SNAKE_CASE)) UpperCAmelCase_ = torch.flatten(_SCREAMING_SNAKE_CASE , start_dim=2) UpperCAmelCase_ = out.transpose(1 , 2).contiguous() return out # BxNx2048 class __snake_case ( __SCREAMING_SNAKE_CASE ): def __init__( self : int , _snake_case : Any , _snake_case : List[str] , _snake_case : Union[str, Any] , _snake_case : Any , _snake_case : int): """simple docstring""" UpperCAmelCase_ = [json.loads(_SCREAMING_SNAKE_CASE) for l in open(_SCREAMING_SNAKE_CASE)] UpperCAmelCase_ = os.path.dirname(_SCREAMING_SNAKE_CASE) UpperCAmelCase_ = tokenizer UpperCAmelCase_ = labels UpperCAmelCase_ = len(_SCREAMING_SNAKE_CASE) UpperCAmelCase_ = max_seq_length UpperCAmelCase_ = transforms def __len__( self : int): """simple docstring""" return len(self.data) def __getitem__( self : Optional[Any] , _snake_case : str): """simple docstring""" UpperCAmelCase_ = torch.LongTensor(self.tokenizer.encode(self.data[index]['''text'''] , add_special_tokens=_SCREAMING_SNAKE_CASE)) UpperCAmelCase_ = sentence[0], sentence[1:-1], sentence[-1] UpperCAmelCase_ = sentence[: self.max_seq_length] UpperCAmelCase_ = torch.zeros(self.n_classes) UpperCAmelCase_ = 1 UpperCAmelCase_ = Image.open(os.path.join(self.data_dir , self.data[index]['''img'''])).convert('''RGB''') UpperCAmelCase_ = self.transforms(_SCREAMING_SNAKE_CASE) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = Counter() for row in self.data: label_freqs.update(row['''label''']) return label_freqs def A (__A : Union[str, Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase_ = [len(row['''sentence'''] ) for row in batch] UpperCAmelCase_ = len(__snake_case ), max(__snake_case ) UpperCAmelCase_ = torch.zeros(__snake_case , __snake_case , dtype=torch.long ) UpperCAmelCase_ = torch.zeros(__snake_case , __snake_case , dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(__snake_case , __snake_case ) ): UpperCAmelCase_ = input_row["sentence"] UpperCAmelCase_ = 1 UpperCAmelCase_ = torch.stack([row['''image'''] for row in batch] ) UpperCAmelCase_ = torch.stack([row['''label'''] for row in batch] ) UpperCAmelCase_ = torch.stack([row['''image_start_token'''] for row in batch] ) UpperCAmelCase_ = torch.stack([row['''image_end_token'''] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def A () -> List[Any]: """simple docstring""" return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def A () -> List[str]: """simple docstring""" return transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize( mean=[0.46_777_044, 0.44_531_429, 0.40_661_017] , std=[0.12_221_994, 0.12_145_835, 0.14_380_469] , ), ] )
51
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_5_0, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_0_0, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> Tuple: """simple docstring""" if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=_SCREAMING_SNAKE_CASE , ) assert hasattr(self , "env") def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , _SCREAMING_SNAKE_CASE: Any=1) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F"""{self.env.base_job_name}-single""" , instance_count=_SCREAMING_SNAKE_CASE , instance_type=self.instance_type , debugger_hook_config=_SCREAMING_SNAKE_CASE , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: List[Any]) -> Optional[Any]: """simple docstring""" TrainingJobAnalytics(_SCREAMING_SNAKE_CASE).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""") def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Tuple = self.create_estimator() # run training estimator.fit() # result dataframe __lowerCAmelCase : Tuple = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis __lowerCAmelCase : Optional[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) __lowerCAmelCase : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCAmelCase : Tuple = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 99_9999) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , _SCREAMING_SNAKE_CASE)
269
0
'''simple docstring''' import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" __a ='facebook/bart-large-mnli' __a =( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) __a ='text_classifier' __a =AutoTokenizer __a =AutoModelForSequenceClassification __a =['text', ['text']] __a =['text'] def UpperCamelCase__ ( self : str ): super().setup() _a = self.model.config _a = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("entail" ): _a = int(__a ) if self.entailment_id == -1: raise ValueError("Could not determine the entailment ID from the model config, please pass it at init." ) def UpperCamelCase__ ( self : str , __a : str , __a : Optional[Any] ): _a = labels return self.pre_processor( [text] * len(__a ) , [f'This example is {label}' for label in labels] , return_tensors="pt" , padding="max_length" , ) def UpperCamelCase__ ( self : Dict , __a : Any ): _a = outputs.logits _a = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
360
'''simple docstring''' import requests lowerCAmelCase_ : List[Any] = 'YOUR API KEY' def _lowerCamelCase ( lowercase : str , lowercase : str = giphy_api_key ) -> list: _a = "+".join(query.split() ) _a = F'https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}' _a = requests.get(lowercase ).json()["data"] return [gif["url"] for gif in gifs] if __name__ == "__main__": print('\n'.join(get_gifs('space ship')))
346
0
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Optional[Any], lowerCamelCase : int )-> None: lowerCamelCase__ : List[Any] =value lowerCamelCase__ : Node | None =None lowerCamelCase__ : Node | None =None class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : List[Any], lowerCamelCase : Node )-> None: lowerCamelCase__ : Optional[Any] =tree def snake_case ( self : int, lowerCamelCase : Node | None )-> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self : Optional[Any] )-> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
238
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def snake_case__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple , __lowerCamelCase : Any , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple=True , __lowerCamelCase : Dict="pt" ): """simple docstring""" lowerCamelCase__ : str ={'''add_prefix_space''': True} if isinstance(__lowerCamelCase , __lowerCamelCase ) and not line.startswith(''' ''' ) else {} lowerCamelCase__ : int =padding_side return tokenizer( [line] , max_length=__lowerCamelCase , padding='''max_length''' if pad_to_max_length else None , truncation=__lowerCamelCase , return_tensors=__lowerCamelCase , add_special_tokens=__lowerCamelCase , **__lowerCamelCase , ) def snake_case__ ( __lowerCamelCase : str , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int]=None , ): """simple docstring""" lowerCamelCase__ : Any =input_ids.ne(__lowerCamelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' def __init__( self : str, lowerCamelCase : Union[str, Any], lowerCamelCase : Dict, lowerCamelCase : Union[str, Any], lowerCamelCase : Dict, lowerCamelCase : str="train", lowerCamelCase : List[Any]=None, lowerCamelCase : Tuple=None, lowerCamelCase : List[str]=None, lowerCamelCase : int="", )-> List[Any]: super().__init__() lowerCamelCase__ : Tuple =Path(lowerCamelCase ).joinpath(type_path + '''.source''' ) lowerCamelCase__ : str =Path(lowerCamelCase ).joinpath(type_path + '''.target''' ) lowerCamelCase__ : Dict =self.get_char_lens(self.src_file ) lowerCamelCase__ : Tuple =max_source_length lowerCamelCase__ : Optional[int] =max_target_length assert min(self.src_lens ) > 0, F'''found empty line in {self.src_file}''' lowerCamelCase__ : Dict =tokenizer lowerCamelCase__ : List[str] =prefix if n_obs is not None: lowerCamelCase__ : int =self.src_lens[:n_obs] lowerCamelCase__ : Dict =src_lang lowerCamelCase__ : Tuple =tgt_lang def __len__( self : Dict )-> Optional[int]: return len(self.src_lens ) def __getitem__( self : List[str], lowerCamelCase : Optional[int] )-> Dict[str, torch.Tensor]: lowerCamelCase__ : List[Any] =index + 1 # linecache starts at 1 lowerCamelCase__ : Optional[int] =self.prefix + linecache.getline(str(self.src_file ), lowerCamelCase ).rstrip('''\n''' ) lowerCamelCase__ : Optional[Any] =linecache.getline(str(self.tgt_file ), lowerCamelCase ).rstrip('''\n''' ) assert source_line, F'''empty source line for index {index}''' assert tgt_line, F'''empty tgt line for index {index}''' # Need to add eos token manually for T5 if isinstance(self.tokenizer, lowerCamelCase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right lowerCamelCase__ : Optional[int] =( self.tokenizer.question_encoder if isinstance(self.tokenizer, lowerCamelCase ) else self.tokenizer ) lowerCamelCase__ : Tuple =self.tokenizer.generator if isinstance(self.tokenizer, lowerCamelCase ) else self.tokenizer lowerCamelCase__ : Optional[int] =encode_line(lowerCamelCase, lowerCamelCase, self.max_source_length, '''right''' ) lowerCamelCase__ : str =encode_line(lowerCamelCase, lowerCamelCase, self.max_target_length, '''right''' ) lowerCamelCase__ : str =source_inputs['''input_ids'''].squeeze() lowerCamelCase__ : str =target_inputs['''input_ids'''].squeeze() lowerCamelCase__ : Union[str, Any] =source_inputs['''attention_mask'''].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def snake_case ( lowerCamelCase : Union[str, Any] )-> Optional[int]: return [len(lowerCamelCase ) for x in Path(lowerCamelCase ).open().readlines()] def snake_case ( self : str, lowerCamelCase : str )-> Dict[str, torch.Tensor]: lowerCamelCase__ : List[Any] =torch.stack([x['''input_ids'''] for x in batch] ) lowerCamelCase__ : int =torch.stack([x['''attention_mask'''] for x in batch] ) lowerCamelCase__ : Union[str, Any] =torch.stack([x['''decoder_input_ids'''] for x in batch] ) lowerCamelCase__ : str =( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer, lowerCamelCase ) else self.tokenizer.pad_token_id ) lowerCamelCase__ : List[str] =( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer, lowerCamelCase ) else self.tokenizer.pad_token_id ) lowerCamelCase__ : Optional[int] =trim_batch(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ , lowerCamelCase__ : Any =trim_batch(lowerCamelCase, lowerCamelCase, attention_mask=lowerCamelCase ) lowerCamelCase__ : List[str] ={ '''input_ids''': source_ids, '''attention_mask''': source_mask, '''decoder_input_ids''': y, } return batch _lowercase : Any = getLogger(__name__) def snake_case__ ( __lowerCamelCase : List[List] ): """simple docstring""" return list(itertools.chain.from_iterable(__lowerCamelCase ) ) def snake_case__ ( __lowerCamelCase : str ): """simple docstring""" lowerCamelCase__ : Dict =get_git_info() save_json(__lowerCamelCase , os.path.join(__lowerCamelCase , '''git_log.json''' ) ) def snake_case__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Dict=4 , **__lowerCamelCase : int ): """simple docstring""" with open(__lowerCamelCase , '''w''' ) as f: json.dump(__lowerCamelCase , __lowerCamelCase , indent=__lowerCamelCase , **__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : List[Any] ): """simple docstring""" with open(__lowerCamelCase ) as f: return json.load(__lowerCamelCase ) def snake_case__ ( ): """simple docstring""" lowerCamelCase__ : List[Any] =git.Repo(search_parent_directories=__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] ={ '''repo_id''': str(__lowerCamelCase ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), '''hostname''': str(socket.gethostname() ), } return repo_infos def snake_case__ ( __lowerCamelCase : Callable , __lowerCamelCase : Iterable ): """simple docstring""" return list(map(__lowerCamelCase , __lowerCamelCase ) ) def snake_case__ ( __lowerCamelCase : Any , __lowerCamelCase : List[Any] ): """simple docstring""" with open(__lowerCamelCase , '''wb''' ) as f: return pickle.dump(__lowerCamelCase , __lowerCamelCase ) def snake_case__ ( __lowerCamelCase : str ): """simple docstring""" def remove_articles(__lowerCamelCase : List[Any] ): return re.sub(R'''\b(a|an|the)\b''' , ''' ''' , __lowerCamelCase ) def white_space_fix(__lowerCamelCase : Any ): return " ".join(text.split() ) def remove_punc(__lowerCamelCase : Optional[Any] ): lowerCamelCase__ : Tuple =set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__lowerCamelCase : Any ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__lowerCamelCase ) ) ) ) def snake_case__ ( __lowerCamelCase : List[str] , __lowerCamelCase : List[str] ): """simple docstring""" lowerCamelCase__ : List[str] =normalize_answer(__lowerCamelCase ).split() lowerCamelCase__ : List[str] =normalize_answer(__lowerCamelCase ).split() lowerCamelCase__ : Optional[int] =Counter(__lowerCamelCase ) & Counter(__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =sum(common.values() ) if num_same == 0: return 0 lowerCamelCase__ : Dict =1.0 * num_same / len(__lowerCamelCase ) lowerCamelCase__ : List[str] =1.0 * num_same / len(__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =(2 * precision * recall) / (precision + recall) return fa def snake_case__ ( __lowerCamelCase : Dict , __lowerCamelCase : int ): """simple docstring""" return normalize_answer(__lowerCamelCase ) == normalize_answer(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : List[str] , __lowerCamelCase : List[str] ): """simple docstring""" assert len(__lowerCamelCase ) == len(__lowerCamelCase ) lowerCamelCase__ : Any =0 for hypo, pred in zip(__lowerCamelCase , __lowerCamelCase ): em += exact_match_score(__lowerCamelCase , __lowerCamelCase ) if len(__lowerCamelCase ) > 0: em /= len(__lowerCamelCase ) return {"em": em} def snake_case__ ( __lowerCamelCase : List[str] ): """simple docstring""" return model_prefix.startswith('''rag''' ) def snake_case__ ( __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : str ): """simple docstring""" lowerCamelCase__ : Any ={p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead lowerCamelCase__ : Optional[int] ='''dropout_rate''' for p in extra_params: if getattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if not hasattr(__lowerCamelCase , __lowerCamelCase ) and not hasattr(__lowerCamelCase , equivalent_param[p] ): logger.info('''config doesn\'t have a `{}` attribute'''.format(__lowerCamelCase ) ) delattr(__lowerCamelCase , __lowerCamelCase ) continue lowerCamelCase__ : List[Any] =p if hasattr(__lowerCamelCase , __lowerCamelCase ) else equivalent_param[p] setattr(__lowerCamelCase , __lowerCamelCase , getattr(__lowerCamelCase , __lowerCamelCase ) ) delattr(__lowerCamelCase , __lowerCamelCase ) return hparams, config
238
1
UpperCamelCase_ = { "joule": 1.0, "kilojoule": 1000, "megajoule": 1000000, "gigajoule": 1000000000, "wattsecond": 1.0, "watthour": 3600, "kilowatthour": 3600000, "newtonmeter": 1.0, "calorie_nutr": 4186.8, "kilocalorie_nutr": 4186800.00, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1055.05585, "footpound": 1.3_5_5_8_1_8, } def lowerCamelCase_ ( _a : str , _a : str , _a : float ): '''simple docstring''' if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: UpperCAmelCase_ : Union[str, Any] = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {', '.join(_a )}''' ) raise ValueError(_a ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
59
from sklearn.metrics import matthews_corrcoef import datasets UpperCamelCase_ = ''' Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] ''' UpperCamelCase_ = ''' Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric("matthews_correlation") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results[\'matthews_correlation\'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric("matthews_correlation") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results[\'matthews_correlation\'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric("matthews_correlation") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results[\'matthews_correlation\'], 2)) -0.25 ''' UpperCamelCase_ = '''\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): '''simple docstring''' def A__ ( self: Any ) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) ,reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] ,) def A__ ( self: List[str] ,lowerCamelCase_: int ,lowerCamelCase_: Any ,lowerCamelCase_: Optional[Any]=None ) -> int: return { "matthews_correlation": float(matthews_corrcoef(lowerCamelCase_ ,lowerCamelCase_ ,sample_weight=lowerCamelCase_ ) ), }
59
1
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType __snake_case : Optional[Any] = logging.get_logger(__name__) __snake_case : Optional[Any] = { 'openai/whisper-base': 'https://huggingface.co/openai/whisper-base/resolve/main/config.json', } # fmt: off __snake_case : Any = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1_058, 1_220, 1_267, 1_279, 1_303, 1_343, 1_377, 1_391, 1_635, 1_782, 1_875, 2_162, 2_361, 2_488, 3_467, 4_008, 4_211, 4_600, 4_808, 5_299, 5_855, 6_329, 7_203, 9_609, 9_959, 10_563, 10_786, 11_420, 11_709, 11_907, 13_163, 13_697, 13_700, 14_808, 15_306, 16_410, 16_791, 17_992, 19_203, 19_510, 20_724, 22_305, 22_935, 27_007, 30_109, 30_420, 33_409, 34_949, 40_283, 40_493, 40_549, 47_282, 49_146, 50_257, 50_359, 50_360, 50_361 ] __snake_case : str = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1_350, 1_853, 1_982, 2_460, 2_627, 3_246, 3_253, 3_268, 3_536, 3_846, 3_961, 4_183, 4_667, 6_585, 6_647, 7_273, 9_061, 9_383, 10_428, 10_929, 11_938, 12_033, 12_331, 12_562, 13_793, 14_157, 14_635, 15_265, 15_618, 16_553, 16_604, 18_362, 18_956, 20_075, 21_675, 22_520, 26_130, 26_161, 26_435, 28_279, 29_464, 31_650, 32_302, 32_470, 36_865, 42_863, 47_425, 49_870, 50_254, 50_258, 50_360, 50_361, 50_362 ] class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'whisper' SCREAMING_SNAKE_CASE = ['past_key_values'] SCREAMING_SNAKE_CASE = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self: Optional[int] , _SCREAMING_SNAKE_CASE: Any=5_1865 , _SCREAMING_SNAKE_CASE: Optional[Any]=80 , _SCREAMING_SNAKE_CASE: Optional[int]=6 , _SCREAMING_SNAKE_CASE: Any=4 , _SCREAMING_SNAKE_CASE: Dict=6 , _SCREAMING_SNAKE_CASE: Dict=4 , _SCREAMING_SNAKE_CASE: Optional[Any]=1536 , _SCREAMING_SNAKE_CASE: List[str]=1536 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: List[str]=5_0257 , _SCREAMING_SNAKE_CASE: Optional[Any]=True , _SCREAMING_SNAKE_CASE: List[str]=True , _SCREAMING_SNAKE_CASE: Optional[int]="gelu" , _SCREAMING_SNAKE_CASE: Tuple=256 , _SCREAMING_SNAKE_CASE: str=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE: List[Any]=0.02 , _SCREAMING_SNAKE_CASE: Tuple=False , _SCREAMING_SNAKE_CASE: Union[str, Any]=1500 , _SCREAMING_SNAKE_CASE: str=448 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: Any=5_0256 , _SCREAMING_SNAKE_CASE: List[str]=5_0256 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: List[Any]=[220, 5_0256] , _SCREAMING_SNAKE_CASE: Dict=False , _SCREAMING_SNAKE_CASE: str=256 , _SCREAMING_SNAKE_CASE: List[str]=False , _SCREAMING_SNAKE_CASE: Tuple=0.05 , _SCREAMING_SNAKE_CASE: List[str]=10 , _SCREAMING_SNAKE_CASE: str=2 , _SCREAMING_SNAKE_CASE: Any=0.0 , _SCREAMING_SNAKE_CASE: Optional[int]=10 , _SCREAMING_SNAKE_CASE: int=0 , _SCREAMING_SNAKE_CASE: Any=7 , **_SCREAMING_SNAKE_CASE: List[str] , ) -> List[Any]: """simple docstring""" __lowerCAmelCase : Optional[int] = vocab_size __lowerCAmelCase : Optional[Any] = num_mel_bins __lowerCAmelCase : int = d_model __lowerCAmelCase : List[Any] = encoder_layers __lowerCAmelCase : List[Any] = encoder_attention_heads __lowerCAmelCase : List[str] = decoder_layers __lowerCAmelCase : Tuple = decoder_attention_heads __lowerCAmelCase : Any = decoder_ffn_dim __lowerCAmelCase : Tuple = encoder_ffn_dim __lowerCAmelCase : List[str] = dropout __lowerCAmelCase : Union[str, Any] = attention_dropout __lowerCAmelCase : Union[str, Any] = activation_dropout __lowerCAmelCase : Dict = activation_function __lowerCAmelCase : Tuple = init_std __lowerCAmelCase : str = encoder_layerdrop __lowerCAmelCase : int = decoder_layerdrop __lowerCAmelCase : Optional[int] = use_cache __lowerCAmelCase : Union[str, Any] = encoder_layers __lowerCAmelCase : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True __lowerCAmelCase : int = max_source_positions __lowerCAmelCase : Any = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. __lowerCAmelCase : Dict = classifier_proj_size __lowerCAmelCase : Dict = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowerCAmelCase : int = apply_spec_augment __lowerCAmelCase : Union[str, Any] = mask_time_prob __lowerCAmelCase : str = mask_time_length __lowerCAmelCase : int = mask_time_min_masks __lowerCAmelCase : List[Any] = mask_feature_prob __lowerCAmelCase : Tuple = mask_feature_length __lowerCAmelCase : Any = mask_feature_min_masks __lowerCAmelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , is_encoder_decoder=_SCREAMING_SNAKE_CASE , decoder_start_token_id=_SCREAMING_SNAKE_CASE , suppress_tokens=_SCREAMING_SNAKE_CASE , begin_suppress_tokens=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Mapping[str, Mapping[int, str]]: """simple docstring""" __lowerCAmelCase : List[str] = OrderedDict( [ ("input_features", {0: "batch", 1: "feature_size", 2: "encoder_sequence"}), ]) if self.use_past: __lowerCAmelCase : Tuple = {0: "batch"} else: __lowerCAmelCase : Union[str, Any] = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(_SCREAMING_SNAKE_CASE , direction="inputs") return common_inputs def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: int = -1 , _SCREAMING_SNAKE_CASE: bool = False , _SCREAMING_SNAKE_CASE: Optional["TensorType"] = None , _SCREAMING_SNAKE_CASE: int = 2_2050 , _SCREAMING_SNAKE_CASE: float = 5.0 , _SCREAMING_SNAKE_CASE: int = 220 , ) -> Mapping[str, Any]: """simple docstring""" __lowerCAmelCase : int = OrderedDict() __lowerCAmelCase : Optional[Any] = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=_SCREAMING_SNAKE_CASE , framework=_SCREAMING_SNAKE_CASE , sampling_rate=_SCREAMING_SNAKE_CASE , time_duration=_SCREAMING_SNAKE_CASE , frequency=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[str] = encoder_inputs["input_features"].shape[2] __lowerCAmelCase : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length __lowerCAmelCase : List[Any] = super().generate_dummy_inputs( preprocessor.tokenizer , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = encoder_inputs.pop("input_features") __lowerCAmelCase : List[Any] = decoder_inputs.pop("decoder_input_ids") if "past_key_values" in decoder_inputs: __lowerCAmelCase : int = decoder_inputs.pop("past_key_values") return dummy_inputs @property def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> float: """simple docstring""" return 1e-3
269
"""simple docstring""" import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = FunnelTokenizer SCREAMING_SNAKE_CASE = FunnelTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[int]: """simple docstring""" super().setUp() __lowerCAmelCase : str = [ "<unk>", "<cls>", "<sep>", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] __lowerCAmelCase : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" return FunnelTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , **_SCREAMING_SNAKE_CASE: Any) -> str: """simple docstring""" return FunnelTokenizerFast.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: str) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "UNwant\u00E9d,running" __lowerCAmelCase : str = "unwanted, running" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" __lowerCAmelCase : Any = self.tokenizer_class(self.vocab_file) __lowerCAmelCase : Any = tokenizer.tokenize("UNwant\u00E9d,running") self.assertListEqual(_SCREAMING_SNAKE_CASE , ["un", "##want", "##ed", ",", "runn", "##ing"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , [7, 4, 5, 10, 8, 9]) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_tokenizers(do_lower_case=_SCREAMING_SNAKE_CASE) for tokenizer in tokenizers: __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running") __lowerCAmelCase : Optional[int] = len(inputs["input_ids"]) - 1 self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len) __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running" , "UNwant\u00E9d,running") self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len + [1] * sentence_len)
269
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _snake_case : int = { "configuration_encodec": [ "ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP", "EncodecConfig", ], "feature_extraction_encodec": ["EncodecFeatureExtractor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Any = [ "ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST", "EncodecModel", "EncodecPreTrainedModel", ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys _snake_case : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
134
from __future__ import annotations from functools import lru_cache from math import ceil _snake_case : Tuple = 100 _snake_case : int = set(range(3, NUM_PRIMES, 2)) primes.add(2) _snake_case : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=1_0_0 ) def lowerCAmelCase_ ( __lowerCamelCase ): if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} __snake_case : set[int] = set() __snake_case : int __snake_case : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def lowerCAmelCase_ ( __lowerCamelCase = 5_0_0_0 ): for number_to_partition in range(1 , __lowerCamelCase ): if len(partition(__lowerCamelCase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(f'''{solution() = }''')
134
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class _lowerCamelCase( _a ): lowercase_ : int = """openai/whisper-base""" lowercase_ : Union[str, Any] = ( """This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the """ """transcribed text.""" ) lowercase_ : Any = """transcriber""" lowercase_ : List[Any] = WhisperProcessor lowercase_ : List[str] = WhisperForConditionalGeneration lowercase_ : Any = ["""audio"""] lowercase_ : Union[str, Any] = ["""text"""] def UpperCamelCase ( self, lowerCamelCase) -> Tuple: """simple docstring""" return self.pre_processor(lowerCamelCase, return_tensors='pt').input_features def UpperCamelCase ( self, lowerCamelCase) -> List[str]: """simple docstring""" return self.model.generate(inputs=lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase) -> Optional[Any]: """simple docstring""" return self.pre_processor.batch_decode(lowerCamelCase, skip_special_tokens=lowerCamelCase)[0]
21
"""simple docstring""" import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class a__ ( a_, unittest.TestCase ): __lowerCAmelCase = CpmAntTokenizer __lowerCAmelCase = False def __magic_name__ ( self ): super().setUp() lowercase : List[str] = [ "<d>", "</d>", "<s>", "</s>", "</_>", "<unk>", "<pad>", "</n>", "我", "是", "C", "P", "M", "A", "n", "t", ] lowercase : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) @tooslow def __magic_name__ ( self ): lowercase : Tuple = CpmAntTokenizer.from_pretrained("openbmb/cpm-ant-10b" ) lowercase : str = "今天天气真好!" lowercase : Optional[int] = ["今天", "天气", "真", "好", "!"] lowercase : int = tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) lowercase : Any = "今天天气真好!" lowercase : int = [tokenizer.bos_token] + tokens lowercase : List[str] = [6, 9_802, 14_962, 2_082, 831, 244] self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , _a ) lowercase : str = tokenizer.decode(_a ) self.assertEqual(_a , _a )
202
0
from __future__ import annotations def __lowercase ( a__ , a__ , a__ ) -> float: if days_between_payments <= 0: raise ValueError('days_between_payments must be > 0' ) if daily_interest_rate < 0: raise ValueError('daily_interest_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * daily_interest_rate * days_between_payments def __lowercase ( a__ , a__ , a__ , ) -> float: if number_of_compounding_periods <= 0: raise ValueError('number_of_compounding_periods must be > 0' ) if nominal_annual_interest_rate_percentage < 0: raise ValueError('nominal_annual_interest_rate_percentage must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def __lowercase ( a__ , a__ , a__ , ) -> float: if number_of_years <= 0: raise ValueError('number_of_years must be > 0' ) if nominal_annual_percentage_rate < 0: raise ValueError('nominal_annual_percentage_rate must be >= 0' ) if principal <= 0: raise ValueError('principal must be > 0' ) return compound_interest( a__ , nominal_annual_percentage_rate / 3_65 , number_of_years * 3_65 ) if __name__ == "__main__": import doctest doctest.testmod()
118
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCAmelCase_ ( UpperCamelCase_ , unittest.TestCase ): '''simple docstring''' UpperCamelCase__ : Optional[int] = LayoutLMTokenizer UpperCamelCase__ : Any = LayoutLMTokenizerFast UpperCamelCase__ : Optional[int] = True UpperCamelCase__ : int = True def _A ( self ): '''simple docstring''' super().setUp() __SCREAMING_SNAKE_CASE = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def _A ( self , **_A ): '''simple docstring''' return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_A ) def _A ( self , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = 'UNwant\u00E9d,running' __SCREAMING_SNAKE_CASE = 'unwanted, running' return input_text, output_text def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.tokenizer_class(self.vocab_file ) __SCREAMING_SNAKE_CASE = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_A , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_A ) , [7, 4, 5, 10, 8, 9] ) def _A ( self ): '''simple docstring''' pass
118
1
def _a ( SCREAMING_SNAKE_CASE_ : int = 1_00 ): __lowerCAmelCase = (n * (n + 1) // 2) ** 2 __lowerCAmelCase = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(f'''{solution() = }''')
92
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class UpperCAmelCase : def __init__( self : str , __snake_case : Any ) -> str: _lowerCAmelCase = str(id_ ) _lowerCAmelCase = None _lowerCAmelCase = None _lowerCAmelCase = [] _lowerCAmelCase = {} # {vertex:distance} def __lt__( self : List[str] , __snake_case : Union[str, Any] ) -> Any: return self.key < other.key def __repr__( self : Optional[Any] ) -> Optional[Any]: return self.id def lowercase__ ( self : Union[str, Any] , __snake_case : Tuple ) -> Optional[Any]: self.neighbors.append(__snake_case ) def lowercase__ ( self : Tuple , __snake_case : List[str] , __snake_case : Tuple ) -> Any: _lowerCAmelCase = weight def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , lowerCAmelCase ) graph[b - 1].add_edge(graph[a - 1] , lowerCAmelCase ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = [] for u in graph: _lowerCAmelCase = math.inf _lowerCAmelCase = None _lowerCAmelCase = 0 _lowerCAmelCase = graph[:] while q: _lowerCAmelCase = min(lowerCAmelCase ) q.remove(lowerCAmelCase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): _lowerCAmelCase = u _lowerCAmelCase = u.edges[v.id] for i in range(1 , len(lowerCAmelCase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" for u in graph: _lowerCAmelCase = math.inf _lowerCAmelCase = None _lowerCAmelCase = 0 _lowerCAmelCase = list(lowerCAmelCase ) hq.heapify(lowerCAmelCase ) while h: _lowerCAmelCase = hq.heappop(lowerCAmelCase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): _lowerCAmelCase = u _lowerCAmelCase = u.edges[v.id] hq.heapify(lowerCAmelCase ) for i in range(1 , len(lowerCAmelCase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def UpperCamelCase__ ( ): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
70
0
"""simple docstring""" import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed _lowerCAmelCase : List[str] = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"{bindir}/../../examples/pytorch/translation"): from run_translation import main # noqa set_seed(42) _lowerCAmelCase : str = "sshleifer/student_marian_en_ro_6_1" _lowerCAmelCase : Union[str, Any] = "sshleifer/tiny-mbart" @require_torch class UpperCAmelCase_ ( _UpperCamelCase ): def snake_case_ ( self : Any , A : List[str]=False , A : Optional[Any]=None , A : Union[str, Any]=True , A : Optional[int]=True , A : int=True , A : Any=True , ): _UpperCAmelCase : int = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=A , num_train_epochs=1 , distributed=A , extra_args_str=A , predict_with_generate=A , do_train=A , do_eval=A , do_predict=A , ) _UpperCAmelCase : Any = TrainerState.load_from_json(os.path.join(A , "trainer_state.json" ) ).log_history if not do_eval: return _UpperCAmelCase : Optional[Any] = [log for log in logs if "eval_loss" in log.keys()] _UpperCAmelCase : List[str] = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats _UpperCAmelCase : List[Any] = eval_metrics[-1] assert isinstance(last_step_stats["eval_bleu"] , A ) assert not math.isnan(float(last_step_stats["eval_loss"] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def snake_case_ ( self : Union[str, Any] ): self.run_seqaseq_quick() @require_torch_multi_gpu def snake_case_ ( self : List[Any] ): self.run_seqaseq_quick(distributed=A ) @require_torch_multi_gpu def snake_case_ ( self : Any ): self.run_seqaseq_quick(distributed=A ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def snake_case_ ( self : Union[str, Any] ): self.run_seqaseq_quick(distributed=A , extra_args_str="--sharded_ddp simple" ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def snake_case_ ( self : int ): self.run_seqaseq_quick(distributed=A , extra_args_str="--sharded_ddp simple --fp16" ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def snake_case_ ( self : Tuple ): self.run_seqaseq_quick(distributed=A , extra_args_str="--sharded_ddp zero_dp_2" , predict_with_generate=A ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def snake_case_ ( self : str ): self.run_seqaseq_quick( distributed=A , extra_args_str="--sharded_ddp zero_dp_2 --fp16" , predict_with_generate=A ) @require_apex @require_torch_gpu def snake_case_ ( self : Tuple ): # XXX: apex breaks the trainer if it's run twice e.g. run_seq2seq.main() from the same # program and it breaks other tests that run from the same pytest worker, therefore until this is # sorted out it must be run only in an external program, that is distributed=True in this # test and only under one or more gpus - if we want cpu will need to make a special test # # specifically to the problem traced it to self.optimizer.step() - if it's run 2nd time via # 2nd main() call it botches the future eval. # self.run_seqaseq_quick(distributed=A , extra_args_str="--fp16 --fp16_backend=apex" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=A , extra_args_str="--fp16 --fp16_backend=apex" ) @parameterized.expand(["base", "low", "high", "mixed"] ) @require_torch_multi_gpu def snake_case_ ( self : List[Any] , A : List[str] ): # as each sub-test is slow-ish split into multiple sub-tests to avoid CI timeout _UpperCAmelCase : Optional[Any] = { # test with the default log_level - should be info and thus log info once "base": {"extra_args_str": "", "n_matches": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes "low": {"extra_args_str": "--log_level debug --log_level_replica debug", "n_matches": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica "high": {"extra_args_str": "--log_level error --log_level_replica debug", "n_matches": 1}, # test with high log_level and log_level_replica - should be quiet on all processes "mixed": {"extra_args_str": "--log_level error --log_level_replica error", "n_matches": 0}, } _UpperCAmelCase : int = experiments[experiment_id] _UpperCAmelCase : Dict = {"distributed": True, "predict_with_generate": False, "do_eval": False, "do_predict": False} _UpperCAmelCase : Optional[Any] = "Running training" with CaptureStderr() as cl: self.run_seqaseq_quick(**A , extra_args_str=data["extra_args_str"] ) _UpperCAmelCase : List[str] = len(re.findall(A , cl.err ) ) self.assertEqual(A , data["n_matches"] ) @slow def snake_case_ ( self : List[str] ): _UpperCAmelCase : Any = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=A , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=A , ) # Check metrics _UpperCAmelCase : List[str] = TrainerState.load_from_json(os.path.join(A , "trainer_state.json" ) ).log_history _UpperCAmelCase : List[str] = [log for log in logs if "eval_loss" in log.keys()] _UpperCAmelCase : Optional[int] = eval_metrics[0] _UpperCAmelCase : Optional[Any] = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["eval_bleu"] , A ) # test if do_predict saves generations and metrics _UpperCAmelCase : int = os.listdir(A ) _UpperCAmelCase : Optional[int] = {os.path.basename(A ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def snake_case_ ( self : Tuple ): from transformers.training_args import OptimizerNames def train_and_return_metrics(A : str ) -> Tuple[int, float]: _UpperCAmelCase : List[str] = "--skip_memory_metrics 0" _UpperCAmelCase : str = self.run_trainer( max_len=1_2_8 , model_name=A , learning_rate=3e-4 , num_train_epochs=1 , optim=A , distributed=A , extra_args_str=A , do_eval=A , do_predict=A , n_gpus_to_use=1 , ) # Check metrics _UpperCAmelCase : Any = TrainerState.load_from_json(Path(A , "trainer_state.json" ) ).log_history _UpperCAmelCase : Tuple = int(logs[0]["train_mem_gpu_peaked_delta"] / 2**2_0 ) _UpperCAmelCase : Optional[int] = int(logs[0]["train_mem_gpu_alloc_delta"] / 2**2_0 ) _UpperCAmelCase : Union[str, Any] = logs[0]["train_loss"] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss _UpperCAmelCase : str = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) _UpperCAmelCase : Union[str, Any] = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) _UpperCAmelCase : List[str] = gpu_alloc_mem_orig - gpu_alloc_mem_bnb _UpperCAmelCase : Optional[Any] = gpu_peak_mem_orig + gpu_alloc_mem_orig _UpperCAmelCase : Dict = gpu_peak_mem_bnb + gpu_alloc_mem_bnb _UpperCAmelCase : Any = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings _UpperCAmelCase : Union[str, Any] = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( A , A , "should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got" f' a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and' f' gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB' , ) self.assertGreater( A , A , "should use ~150MB less total gpu memory with BNB, compared to without it for this model but got" f' a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and' f' gpu_total_mem_bnb={gpu_total_mem_bnb}MB' , ) self.assertEqual( A , A , f'loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}' ) def snake_case_ ( self : Optional[Any] , A : int , A : str , A : int , A : float = 3e-3 , A : str = "adafactor" , A : bool = False , A : str = None , A : int = 0 , A : bool = True , A : bool = True , A : bool = True , A : bool = True , A : int = None , ): _UpperCAmelCase : Union[str, Any] = self.test_file_dir / "../fixtures/tests_samples/wmt_en_ro" _UpperCAmelCase : Union[str, Any] = self.get_auto_remove_tmp_dir() _UpperCAmelCase : Optional[Any] = f'\n --model_name_or_path {model_name}\n --train_file {data_dir}/train.json\n --validation_file {data_dir}/val.json\n --test_file {data_dir}/test.json\n --output_dir {output_dir}\n --overwrite_output_dir\n --max_train_samples 8\n --max_source_length {max_len}\n --max_target_length {max_len}\n --do_train\n --num_train_epochs {str(A )}\n --per_device_train_batch_size 4\n --learning_rate {learning_rate}\n --warmup_steps 8\n --logging_steps 0\n --logging_strategy no\n --save_steps {str(A )}\n --group_by_length\n --label_smoothing_factor 0.1\n --target_lang ro_RO\n --source_lang en_XX\n '.split() _UpperCAmelCase : Tuple = f'\n --do_eval\n --per_device_eval_batch_size 4\n --max_eval_samples 8\n --val_max_target_length {max_len}\n --evaluation_strategy steps\n --eval_steps {str(A )}\n '.split() _UpperCAmelCase : List[Any] = "\n --do_predict\n ".split() _UpperCAmelCase : List[str] = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += f'--optim {optim}'.split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: _UpperCAmelCase : Union[str, Any] = get_gpu_count() _UpperCAmelCase : Optional[Any] = get_torch_dist_unique_port() _UpperCAmelCase : List[str] = f'\n -m torch.distributed.run\n --nproc_per_node={n_gpus_to_use}\n --master_port={master_port}\n {self.examples_dir_str}/pytorch/translation/run_translation.py\n '.split() _UpperCAmelCase : Optional[int] = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(A , env=self.get_env() ) else: _UpperCAmelCase : List[str] = ["run_translation.py"] + args with patch.object(A , "argv" , A ): main() return output_dir
364
"""simple docstring""" import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase : Dict = logging.get_logger(__name__) _lowerCAmelCase : List[str] = { "google/owlvit-base-patch32": "https://huggingface.co/google/owlvit-base-patch32/resolve/main/config.json", "google/owlvit-base-patch16": "https://huggingface.co/google/owlvit-base-patch16/resolve/main/config.json", "google/owlvit-large-patch14": "https://huggingface.co/google/owlvit-large-patch14/resolve/main/config.json", } class UpperCAmelCase_ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE : int = 'owlvit_text_model' def __init__( self : int , A : int=4_9_4_0_8 , A : Optional[Any]=5_1_2 , A : Optional[Any]=2_0_4_8 , A : str=1_2 , A : int=8 , A : Tuple=1_6 , A : List[Any]="quick_gelu" , A : Tuple=1e-5 , A : Union[str, Any]=0.0 , A : List[Any]=0.02 , A : str=1.0 , A : str=0 , A : List[str]=4_9_4_0_6 , A : str=4_9_4_0_7 , **A : Optional[Any] , ): super().__init__(pad_token_id=A , bos_token_id=A , eos_token_id=A , **A ) _UpperCAmelCase : Union[str, Any] = vocab_size _UpperCAmelCase : str = hidden_size _UpperCAmelCase : List[Any] = intermediate_size _UpperCAmelCase : Any = num_hidden_layers _UpperCAmelCase : str = num_attention_heads _UpperCAmelCase : List[str] = max_position_embeddings _UpperCAmelCase : List[Any] = hidden_act _UpperCAmelCase : Tuple = layer_norm_eps _UpperCAmelCase : List[str] = attention_dropout _UpperCAmelCase : Optional[Any] = initializer_range _UpperCAmelCase : List[Any] = initializer_factor @classmethod def snake_case_ ( cls : Any , A : Union[str, os.PathLike] , **A : Dict ): cls._set_token_in_kwargs(A ) _UpperCAmelCase , _UpperCAmelCase : List[str] = cls.get_config_dict(A , **A ) # get the text config dict if we are loading from OwlViTConfig if config_dict.get("model_type" ) == "owlvit": _UpperCAmelCase : int = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(A , **A ) class UpperCAmelCase_ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE : Tuple = 'owlvit_vision_model' def __init__( self : Union[str, Any] , A : Optional[int]=7_6_8 , A : int=3_0_7_2 , A : List[str]=1_2 , A : List[str]=1_2 , A : Optional[int]=3 , A : Optional[int]=7_6_8 , A : str=3_2 , A : Tuple="quick_gelu" , A : Dict=1e-5 , A : Optional[int]=0.0 , A : List[Any]=0.02 , A : str=1.0 , **A : Tuple , ): super().__init__(**A ) _UpperCAmelCase : List[str] = hidden_size _UpperCAmelCase : Tuple = intermediate_size _UpperCAmelCase : Optional[Any] = num_hidden_layers _UpperCAmelCase : Dict = num_attention_heads _UpperCAmelCase : Optional[Any] = num_channels _UpperCAmelCase : Union[str, Any] = image_size _UpperCAmelCase : Dict = patch_size _UpperCAmelCase : List[str] = hidden_act _UpperCAmelCase : Union[str, Any] = layer_norm_eps _UpperCAmelCase : Any = attention_dropout _UpperCAmelCase : Tuple = initializer_range _UpperCAmelCase : Tuple = initializer_factor @classmethod def snake_case_ ( cls : Optional[int] , A : Union[str, os.PathLike] , **A : int ): cls._set_token_in_kwargs(A ) _UpperCAmelCase , _UpperCAmelCase : Dict = cls.get_config_dict(A , **A ) # get the vision config dict if we are loading from OwlViTConfig if config_dict.get("model_type" ) == "owlvit": _UpperCAmelCase : Tuple = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(A , **A ) class UpperCAmelCase_ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE : List[str] = 'owlvit' __SCREAMING_SNAKE_CASE : Optional[Any] = True def __init__( self : Optional[Any] , A : Dict=None , A : Tuple=None , A : Optional[Any]=5_1_2 , A : Optional[Any]=2.6_592 , A : int=True , **A : Tuple , ): super().__init__(**A ) if text_config is None: _UpperCAmelCase : List[Any] = {} logger.info("text_config is None. Initializing the OwlViTTextConfig with default values." ) if vision_config is None: _UpperCAmelCase : Tuple = {} logger.info("vision_config is None. initializing the OwlViTVisionConfig with default values." ) _UpperCAmelCase : str = OwlViTTextConfig(**A ) _UpperCAmelCase : int = OwlViTVisionConfig(**A ) _UpperCAmelCase : Optional[Any] = projection_dim _UpperCAmelCase : str = logit_scale_init_value _UpperCAmelCase : Optional[Any] = return_dict _UpperCAmelCase : str = 1.0 @classmethod def snake_case_ ( cls : Dict , A : Union[str, os.PathLike] , **A : Any ): cls._set_token_in_kwargs(A ) _UpperCAmelCase , _UpperCAmelCase : str = cls.get_config_dict(A , **A ) if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(A , **A ) @classmethod def snake_case_ ( cls : Optional[int] , A : Dict , A : Dict , **A : Optional[Any] ): _UpperCAmelCase : Optional[Any] = {} _UpperCAmelCase : int = text_config _UpperCAmelCase : Dict = vision_config return cls.from_dict(A , **A ) def snake_case_ ( self : Optional[int] ): _UpperCAmelCase : str = copy.deepcopy(self.__dict__ ) _UpperCAmelCase : Optional[int] = self.text_config.to_dict() _UpperCAmelCase : Optional[int] = self.vision_config.to_dict() _UpperCAmelCase : List[Any] = self.__class__.model_type return output class UpperCAmelCase_ ( _UpperCamelCase ): @property def snake_case_ ( self : List[str] ): return OrderedDict( [ ("input_ids", {0: "batch", 1: "sequence"}), ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("attention_mask", {0: "batch", 1: "sequence"}), ] ) @property def snake_case_ ( self : Optional[int] ): return OrderedDict( [ ("logits_per_image", {0: "batch"}), ("logits_per_text", {0: "batch"}), ("text_embeds", {0: "batch"}), ("image_embeds", {0: "batch"}), ] ) @property def snake_case_ ( self : str ): return 1e-4 def snake_case_ ( self : str , A : "ProcessorMixin" , A : int = -1 , A : int = -1 , A : Optional["TensorType"] = None , ): _UpperCAmelCase : Optional[Any] = super().generate_dummy_inputs( processor.tokenizer , batch_size=A , seq_length=A , framework=A ) _UpperCAmelCase : Union[str, Any] = super().generate_dummy_inputs( processor.image_processor , batch_size=A , framework=A ) return {**text_input_dict, **image_input_dict} @property def snake_case_ ( self : List[Any] ): return 1_4
202
0
"""simple docstring""" def lowercase_ ( _lowerCamelCase: int = 200 ) -> int: '''simple docstring''' __lowerCamelCase : Union[str, Any] = [1, 2, 5, 10, 20, 50, 100, 200] __lowerCamelCase : Optional[int] = [0] * (pence + 1) __lowerCamelCase : Dict = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowerCamelCase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
135
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer __A = logging.get_logger(__name__) __A = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __A = { '''vocab_file''': { '''google/realm-cc-news-pretrained-embedder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt''' ), '''google/realm-cc-news-pretrained-encoder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt''' ), '''google/realm-cc-news-pretrained-scorer''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt''' ), '''google/realm-cc-news-pretrained-openqa''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt''' ), '''google/realm-orqa-nq-openqa''': '''https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt''', '''google/realm-orqa-nq-reader''': '''https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt''', '''google/realm-orqa-wq-openqa''': '''https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt''', '''google/realm-orqa-wq-reader''': '''https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt''', }, '''tokenizer_file''': { '''google/realm-cc-news-pretrained-embedder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont''' ), '''google/realm-cc-news-pretrained-encoder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json''' ), '''google/realm-cc-news-pretrained-scorer''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json''' ), '''google/realm-cc-news-pretrained-openqa''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json''' ), '''google/realm-orqa-nq-openqa''': ( '''https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json''' ), '''google/realm-orqa-nq-reader''': ( '''https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json''' ), '''google/realm-orqa-wq-openqa''': ( '''https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json''' ), '''google/realm-orqa-wq-reader''': ( '''https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json''' ), }, } __A = { '''google/realm-cc-news-pretrained-embedder''': 512, '''google/realm-cc-news-pretrained-encoder''': 512, '''google/realm-cc-news-pretrained-scorer''': 512, '''google/realm-cc-news-pretrained-openqa''': 512, '''google/realm-orqa-nq-openqa''': 512, '''google/realm-orqa-nq-reader''': 512, '''google/realm-orqa-wq-openqa''': 512, '''google/realm-orqa-wq-reader''': 512, } __A = { '''google/realm-cc-news-pretrained-embedder''': {'''do_lower_case''': True}, '''google/realm-cc-news-pretrained-encoder''': {'''do_lower_case''': True}, '''google/realm-cc-news-pretrained-scorer''': {'''do_lower_case''': True}, '''google/realm-cc-news-pretrained-openqa''': {'''do_lower_case''': True}, '''google/realm-orqa-nq-openqa''': {'''do_lower_case''': True}, '''google/realm-orqa-nq-reader''': {'''do_lower_case''': True}, '''google/realm-orqa-wq-openqa''': {'''do_lower_case''': True}, '''google/realm-orqa-wq-reader''': {'''do_lower_case''': True}, } class _snake_case ( a__ ): snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = PRETRAINED_INIT_CONFIGURATION snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ = RealmTokenizer def __init__( self : Optional[int] , UpperCAmelCase : Any=None , UpperCAmelCase : List[str]=None , UpperCAmelCase : int=True , UpperCAmelCase : List[Any]="[UNK]" , UpperCAmelCase : Tuple="[SEP]" , UpperCAmelCase : List[str]="[PAD]" , UpperCAmelCase : Tuple="[CLS]" , UpperCAmelCase : List[Any]="[MASK]" , UpperCAmelCase : str=True , UpperCAmelCase : Union[str, Any]=None , **UpperCAmelCase : Any , ): super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , do_lower_case=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , tokenize_chinese_chars=UpperCAmelCase , strip_accents=UpperCAmelCase , **UpperCAmelCase , ) __lowerCamelCase : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , UpperCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , UpperCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , UpperCAmelCase ) != tokenize_chinese_chars ): __lowerCamelCase : str = getattr(UpperCAmelCase , normalizer_state.pop("type" ) ) __lowerCamelCase : Any = do_lower_case __lowerCamelCase : List[Any] = strip_accents __lowerCamelCase : Optional[Any] = tokenize_chinese_chars __lowerCamelCase : int = normalizer_class(**UpperCAmelCase ) __lowerCamelCase : List[Any] = do_lower_case def lowerCamelCase__ ( self : Union[str, Any] , UpperCAmelCase : Dict , **UpperCAmelCase : int ): __lowerCamelCase : Optional[int] = PaddingStrategy.MAX_LENGTH __lowerCamelCase : List[Any] = text __lowerCamelCase : Optional[int] = kwargs.pop("text_pair" , UpperCAmelCase ) __lowerCamelCase : List[Any] = kwargs.pop("return_tensors" , UpperCAmelCase ) __lowerCamelCase : Dict = { "input_ids": [], "attention_mask": [], "token_type_ids": [], } for idx, candidate_text in enumerate(UpperCAmelCase ): if batch_text_pair is not None: __lowerCamelCase : List[str] = batch_text_pair[idx] else: __lowerCamelCase : Optional[int] = None __lowerCamelCase : List[str] = super().__call__(UpperCAmelCase , UpperCAmelCase , return_tensors=UpperCAmelCase , **UpperCAmelCase ) __lowerCamelCase : Union[str, Any] = encoded_candidates.get("input_ids" ) __lowerCamelCase : Optional[int] = encoded_candidates.get("attention_mask" ) __lowerCamelCase : int = encoded_candidates.get("token_type_ids" ) if encoded_input_ids is not None: output_data["input_ids"].append(UpperCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(UpperCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(UpperCAmelCase ) __lowerCamelCase : Union[str, Any] = {key: item for key, item in output_data.items() if len(UpperCAmelCase ) != 0} return BatchEncoding(UpperCAmelCase , tensor_type=UpperCAmelCase ) def lowerCamelCase__ ( self : Optional[Any] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[int]=None ): __lowerCamelCase : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCamelCase__ ( self : Optional[Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ): __lowerCamelCase : Tuple = [self.sep_token_id] __lowerCamelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase__ ( self : int , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ): __lowerCamelCase : Any = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase )
135
1
import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A : '''simple docstring''' def __init__( self : Optional[Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Optional[int]=13 , __lowerCAmelCase : List[str]=32 , __lowerCAmelCase : List[Any]=3 , __lowerCAmelCase : Union[str, Any]=4 , __lowerCAmelCase : Any=[10, 20, 30, 40] , __lowerCAmelCase : Dict=[2, 2, 3, 2] , __lowerCAmelCase : Dict=True , __lowerCAmelCase : Optional[Any]=True , __lowerCAmelCase : Optional[int]=37 , __lowerCAmelCase : Union[str, Any]="gelu" , __lowerCAmelCase : int=10 , __lowerCAmelCase : List[str]=0.0_2 , __lowerCAmelCase : Union[str, Any]=["stage2", "stage3", "stage4"] , __lowerCAmelCase : Optional[int]=[2, 3, 4] , __lowerCAmelCase : List[Any]=None , ) -> Dict: """simple docstring""" A__ = parent A__ = batch_size A__ = image_size A__ = num_channels A__ = num_stages A__ = hidden_sizes A__ = depths A__ = is_training A__ = use_labels A__ = intermediate_size A__ = hidden_act A__ = num_labels A__ = initializer_range A__ = out_features A__ = out_indices A__ = scope def a_ ( self : Dict ) -> Tuple: """simple docstring""" A__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.num_labels ) A__ = self.get_config() return config, pixel_values, labels def a_ ( self : Union[str, Any] ) -> Dict: """simple docstring""" return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=__lowerCAmelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def a_ ( self : Any , __lowerCAmelCase : str , __lowerCAmelCase : Any , __lowerCAmelCase : Optional[Any] ) -> str: """simple docstring""" A__ = ConvNextVaModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() A__ = model(__lowerCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def a_ ( self : List[Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Tuple ) -> str: """simple docstring""" A__ = ConvNextVaForImageClassification(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() A__ = model(__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a_ ( self : List[Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Dict ) -> Optional[Any]: """simple docstring""" A__ = ConvNextVaBackbone(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() A__ = model(__lowerCAmelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None A__ = None A__ = ConvNextVaBackbone(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() A__ = model(__lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def a_ ( self : int ) -> List[Any]: """simple docstring""" A__ = self.prepare_config_and_inputs() A__ , A__ , A__ = config_and_inputs A__ = {"""pixel_values""": pixel_values} return config, inputs_dict def a_ ( self : List[Any] ) -> Optional[int]: """simple docstring""" A__ = self.prepare_config_and_inputs() A__ , A__ , A__ = config_and_inputs A__ = {"""pixel_values""": pixel_values, """labels""": labels} return config, inputs_dict @require_torch class A (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' __lowerCamelCase : Optional[int] = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) __lowerCamelCase : List[str] = ( {'''feature-extraction''': ConvNextVaModel, '''image-classification''': ConvNextVaForImageClassification} if is_torch_available() else {} ) __lowerCamelCase : Dict = False __lowerCamelCase : int = False __lowerCamelCase : Optional[int] = False __lowerCamelCase : Dict = False __lowerCamelCase : Optional[Any] = False def a_ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" A__ = ConvNextVaModelTester(self ) A__ = ConfigTester(self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=37 ) def a_ ( self : Optional[int] ) -> Tuple: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def a_ ( self : int ) -> int: """simple docstring""" return @unittest.skip(reason="""ConvNextV2 does not use inputs_embeds""" ) def a_ ( self : Any ) -> str: """simple docstring""" pass @unittest.skip(reason="""ConvNextV2 does not support input and output embeddings""" ) def a_ ( self : List[Any] ) -> Any: """simple docstring""" pass @unittest.skip(reason="""ConvNextV2 does not use feedforward chunking""" ) def a_ ( self : Optional[int] ) -> Any: """simple docstring""" pass def a_ ( self : Dict ) -> Any: """simple docstring""" if not self.model_tester.is_training: return for model_class in self.all_model_classes: A__ , A__ = self.model_tester.prepare_config_and_inputs_with_labels() A__ = True if model_class.__name__ in [ *get_values(__lowerCAmelCase ), *get_values(__lowerCAmelCase ), ]: continue A__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.train() A__ = self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase , return_labels=__lowerCAmelCase ) A__ = model(**__lowerCAmelCase ).loss loss.backward() def a_ ( self : Optional[int] ) -> Dict: """simple docstring""" if not self.model_tester.is_training: return for model_class in self.all_model_classes: A__ , A__ = self.model_tester.prepare_config_and_inputs_with_labels() A__ = False A__ = True if ( model_class.__name__ in [*get_values(__lowerCAmelCase ), *get_values(__lowerCAmelCase )] or not model_class.supports_gradient_checkpointing ): continue A__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.gradient_checkpointing_enable() model.train() A__ = self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase , return_labels=__lowerCAmelCase ) A__ = model(**__lowerCAmelCase ).loss loss.backward() def a_ ( self : List[Any] ) -> int: """simple docstring""" A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(__lowerCAmelCase ) A__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ = [*signature.parameters.keys()] A__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCAmelCase ) def a_ ( self : Optional[Any] ) -> Dict: """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) def a_ ( self : Optional[int] ) -> Dict: """simple docstring""" def check_hidden_states_output(__lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[Any] ): A__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): A__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) A__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A__ = self.model_tester.num_stages self.assertEqual(len(__lowerCAmelCase ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def a_ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCAmelCase ) @slow def a_ ( self : Tuple ) -> List[str]: """simple docstring""" for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = ConvNextVaModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def __lowerCamelCase ( ) -> str: """simple docstring""" A__ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class A (unittest.TestCase ): '''simple docstring''' @cached_property def a_ ( self : Union[str, Any] ) -> Any: """simple docstring""" return AutoImageProcessor.from_pretrained("""facebook/convnextv2-tiny-1k-224""" ) if is_vision_available() else None @slow def a_ ( self : str ) -> Tuple: """simple docstring""" A__ = ConvNextVaForImageClassification.from_pretrained("""facebook/convnextv2-tiny-1k-224""" ).to(__lowerCAmelCase ) A__ = self.default_image_processor A__ = prepare_img() A__ = preprocessor(images=__lowerCAmelCase , return_tensors="""pt""" ).to(__lowerCAmelCase ) # forward pass with torch.no_grad(): A__ = model(**__lowerCAmelCase ) # verify the logits A__ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __lowerCAmelCase ) A__ = torch.tensor([0.9_9_9_6, 0.1_9_6_6, -0.4_3_8_6] ).to(__lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCAmelCase , atol=1e-4 ) )
276
import os import re import shutil import sys import tempfile import unittest import black A : Dict = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, '''utils''')) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. A : Optional[int] = ''' def __init__(self, config): super().__init__() self.transform = BertPredictionHeadTransform(config) # The output weights are the same as the input embeddings, but there is # an output-only bias for each token. self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False) self.bias = nn.Parameter(torch.zeros(config.vocab_size)) # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings` self.decoder.bias = self.bias def forward(self, hidden_states): hidden_states = self.transform(hidden_states) hidden_states = self.decoder(hidden_states) return hidden_states ''' class A (unittest.TestCase ): '''simple docstring''' def a_ ( self : Union[str, Any] ) -> List[str]: """simple docstring""" A__ = tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , """models/bert/""" ) ) A__ = self.transformer_dir shutil.copy( os.path.join(__lowerCAmelCase , """src/transformers/models/bert/modeling_bert.py""" ) , os.path.join(self.transformer_dir , """models/bert/modeling_bert.py""" ) , ) def a_ ( self : str ) -> Optional[int]: """simple docstring""" A__ = """src/transformers""" shutil.rmtree(self.transformer_dir ) def a_ ( self : List[str] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : Tuple=None ) -> Dict: """simple docstring""" A__ = comment + f'\nclass {class_name}(nn.Module):\n' + class_code if overwrite_result is not None: A__ = comment + f'\nclass {class_name}(nn.Module):\n' + overwrite_result A__ = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_19 ) A__ = black.format_str(__lowerCAmelCase , mode=__lowerCAmelCase ) A__ = os.path.join(self.transformer_dir , """new_code.py""" ) with open(__lowerCAmelCase , """w""" , newline="""\n""" ) as f: f.write(__lowerCAmelCase ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(__lowerCAmelCase ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=__lowerCAmelCase ) with open(__lowerCAmelCase , """r""" ) as f: self.assertTrue(f.read() , __lowerCAmelCase ) def a_ ( self : Tuple ) -> List[Any]: """simple docstring""" A__ = check_copies.find_code_in_transformers("""models.bert.modeling_bert.BertLMPredictionHead""" ) self.assertEqual(__lowerCAmelCase , __lowerCAmelCase ) def a_ ( self : Tuple ) -> Any: """simple docstring""" self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead""" , """BertLMPredictionHead""" , REFERENCE_CODE + """\n""" , ) # With no empty line at the end self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead""" , """BertLMPredictionHead""" , __lowerCAmelCase , ) # Copy consistency with rename self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel""" , """TestModelLMPredictionHead""" , re.sub("""Bert""" , """TestModel""" , __lowerCAmelCase ) , ) # Copy consistency with a really long name A__ = """TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason""" self.check_copy_consistency( f'# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}' , f'{long_class_name}LMPredictionHead' , re.sub("""Bert""" , __lowerCAmelCase , __lowerCAmelCase ) , ) # Copy consistency with overwrite self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel""" , """TestModelLMPredictionHead""" , __lowerCAmelCase , overwrite_result=re.sub("""Bert""" , """TestModel""" , __lowerCAmelCase ) , ) def a_ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" A__ = check_copies.LOCALIZED_READMES["""README_zh-hans.md"""] A__ = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the""" """ Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for""" """ Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong""" """ Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1.""" """ **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace),""" """ released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and""" """ lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same""" """ method has been applied to compress GPT2 into""" """ [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into""" """ [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),""" """ Multilingual BERT into""" """ [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German""" """ version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**""" """ (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders""" """ as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang""" """ Luong, Quoc V. Le, Christopher D. Manning.""" ) A__ = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the""" """ Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n""" ) A__ = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the""" """ Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1.""" """ **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文""" """ [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and""" """ lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same""" """ method has been applied to compress GPT2 into""" """ [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into""" """ [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),""" """ Multilingual BERT into""" """ [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German""" """ version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自""" """ Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather""" """ than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le,""" """ Christopher D. Manning 发布。\n""" ) A__ , A__ = check_copies.convert_to_localized_md( __lowerCAmelCase , __lowerCAmelCase , localized_readme["""format_model_list"""] ) self.assertFalse(__lowerCAmelCase ) self.assertEqual(__lowerCAmelCase , __lowerCAmelCase ) A__ , A__ = check_copies.convert_to_localized_md( __lowerCAmelCase , __lowerCAmelCase , localized_readme["""format_model_list"""] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(__lowerCAmelCase ) A__ = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the""" """ Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for""" """ Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong""" """ Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.""" ) A__ = ( """1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and""" """ the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n""" ) A__ = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the""" """ Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n""" ) A__ , A__ = check_copies.convert_to_localized_md( __lowerCAmelCase , __lowerCAmelCase , localized_readme["""format_model_list"""] ) # Check if the model link is synchronized. self.assertEqual(__lowerCAmelCase , __lowerCAmelCase )
276
1
'''simple docstring''' import re from filelock import FileLock try: import nltk __snake_case = True except (ImportError, ModuleNotFoundError): __snake_case = False if NLTK_AVAILABLE: with FileLock('''.lock''') as lock: nltk.download('''punkt''', quiet=True) def a ( __a ) -> str: '''simple docstring''' re.sub('''<n>''' , '''''' , __a ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(__a ) )
97
'''simple docstring''' import csv import tweepy # Twitter API credentials __snake_case = '''''' __snake_case = '''''' __snake_case = '''''' __snake_case = '''''' def a ( __a ) -> None: '''simple docstring''' UpperCamelCase__ :List[Any] = tweepy.OAuthHandler(__a , __a ) auth.set_access_token(__a , __a ) UpperCamelCase__ :List[str] = tweepy.API(__a ) # initialize a list to hold all the tweepy Tweets UpperCamelCase__ :Dict = [] # make initial request for most recent tweets (200 is the maximum allowed count) UpperCamelCase__ :Tuple = api.user_timeline(screen_name=__a , count=200 ) # save most recent tweets alltweets.extend(__a ) # save the id of the oldest tweet less one UpperCamelCase__ :Union[str, Any] = alltweets[-1].id - 1 # keep grabbing tweets until there are no tweets left to grab while len(__a ) > 0: print(f'''getting tweets before {oldest}''' ) # all subsequent requests use the max_id param to prevent duplicates UpperCamelCase__ :Union[str, Any] = api.user_timeline( screen_name=__a , count=200 , max_id=__a ) # save most recent tweets alltweets.extend(__a ) # update the id of the oldest tweet less one UpperCamelCase__ :Tuple = alltweets[-1].id - 1 print(f'''...{len(__a )} tweets downloaded so far''' ) # transform the tweepy tweets into a 2D array that will populate the csv UpperCamelCase__ :int = [[tweet.id_str, tweet.created_at, tweet.text] for tweet in alltweets] # write the csv with open(f'''new_{screen_name}_tweets.csv''' , '''w''' ) as f: UpperCamelCase__ :Tuple = csv.writer(__a ) writer.writerow(['''id''', '''created_at''', '''text'''] ) writer.writerows(__a ) if __name__ == "__main__": # pass in the username of the account you want to download get_all_tweets('''FirePing32''')
97
1
"""simple docstring""" import timeit import numpy as np import datasets from datasets.arrow_writer import ArrowWriter from datasets.features.features import _ArrayXD def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' def wrapper(*lowerCAmelCase , **lowerCAmelCase ): UpperCAmelCase = timeit.default_timer() UpperCAmelCase = func(*lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase = timeit.default_timer() - starttime return delta UpperCAmelCase = func.__name__ return wrapper def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase=100 , lowerCAmelCase=None ): '''simple docstring''' UpperCAmelCase = [] UpperCAmelCase = seq_shapes or {} for i in range(lowerCAmelCase ): UpperCAmelCase = {} for col_id, (k, v) in enumerate(features.items() ): if isinstance(lowerCAmelCase , _ArrayXD ): UpperCAmelCase = np.random.rand(*v.shape ).astype(v.dtype ) elif isinstance(lowerCAmelCase , datasets.Value ): if v.dtype == "string": UpperCAmelCase = '''The small grey turtle was surprisingly fast when challenged.''' else: UpperCAmelCase = np.random.randint(10 , size=1 ).astype(v.dtype ).item() elif isinstance(lowerCAmelCase , datasets.Sequence ): while isinstance(lowerCAmelCase , datasets.Sequence ): UpperCAmelCase = v.feature UpperCAmelCase = seq_shapes[k] UpperCAmelCase = np.random.rand(*lowerCAmelCase ).astype(v.dtype ) UpperCAmelCase = data dummy_data.append((i, example) ) return dummy_data def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=100 , lowerCAmelCase=None ): '''simple docstring''' UpperCAmelCase = generate_examples(lowerCAmelCase , num_examples=lowerCAmelCase , seq_shapes=lowerCAmelCase ) with ArrowWriter(features=lowerCAmelCase , path=lowerCAmelCase ) as writer: for key, record in dummy_data: UpperCAmelCase = features.encode_example(lowerCAmelCase ) writer.write(lowerCAmelCase ) UpperCAmelCase = writer.finalize() if not num_final_examples == num_examples: raise ValueError( F'''Error writing the dataset, wrote {num_final_examples} examples but should have written {num_examples}.''' ) UpperCAmelCase = datasets.Dataset.from_file(filename=lowerCAmelCase , info=datasets.DatasetInfo(features=lowerCAmelCase ) ) return dataset
357
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase_ : str = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Optional[int] = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Optional[Any] = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ : Tuple = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys lowerCAmelCase_ : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
248
0
import re def A ( _lowerCamelCase ): '''simple docstring''' return [char.split() for char in re.split(r"[^ a-z A-Z 0-9 \s]" , str_ )] def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Tuple = split_input(str_ ) return "".join( ["".join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' try: _lowerCAmelCase : Any = split_input(_lowerCamelCase ) if upper: _lowerCAmelCase : Dict = "".join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: _lowerCAmelCase : Dict = "".join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def A ( _lowerCamelCase ): '''simple docstring''' return to_simple_case(_lowerCamelCase ) def A ( _lowerCamelCase ): '''simple docstring''' try: _lowerCAmelCase : Dict = to_simple_case(_lowerCamelCase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' return to_complex_case(_lowerCamelCase , _lowerCamelCase , "_" ) def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' return to_complex_case(_lowerCamelCase , _lowerCamelCase , "-" ) if __name__ == "__main__": __import__("doctest").testmod()
36
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('Temporarily disable the doc tests.' ) @require_torch @require_tf @slow class __magic_name__ ( unittest.TestCase ): def __lowercase ( self : str ,_UpperCAmelCase : Path ,_UpperCAmelCase : Union[str, None] = None ,_UpperCAmelCase : Union[List[str], None] = None ,_UpperCAmelCase : Union[str, List[str], None] = None ,_UpperCAmelCase : bool = True ,): _a : Dict = [file for file in os.listdir(_UpperCAmelCase ) if os.path.isfile(os.path.join(_UpperCAmelCase ,_UpperCAmelCase ) )] if identifier is not None: _a : str = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_UpperCAmelCase ,_UpperCAmelCase ): for n_ in n_identifier: _a : int = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : Dict = ignore_files or [] ignore_files.append('__init__.py' ) _a : List[str] = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_UpperCAmelCase ) if only_modules: _a : Any = file.split('.' )[0] try: _a : Optional[int] = getattr(_UpperCAmelCase ,_UpperCAmelCase ) _a : Dict = doctest.DocTestSuite(_UpperCAmelCase ) _a : Optional[int] = unittest.TextTestRunner().run(_UpperCAmelCase ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : str = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Union[str, Any] ): _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'modeling' _a : Union[str, Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_UpperCAmelCase ,identifier=_UpperCAmelCase ,ignore_files=_UpperCAmelCase ) def __lowercase ( self : int ): _a : str = Path('src/transformers' ) _a : List[str] = 'tokenization' self.analyze_directory(_UpperCAmelCase ,identifier=_UpperCAmelCase ) def __lowercase ( self : int ): _a : Any = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_UpperCAmelCase ,identifier=_UpperCAmelCase ) def __lowercase ( self : Dict ): _a : Tuple = Path('src/transformers' ) _a : Optional[int] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_UpperCAmelCase ,n_identifier=_UpperCAmelCase ) def __lowercase ( self : Optional[Any] ): _a : Union[str, Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_UpperCAmelCase ,ignore_files=_UpperCAmelCase ,only_modules=_UpperCAmelCase )
89
0
"""simple docstring""" import os from typing import BinaryIO, Optional, Union import numpy as np import pyarrow.parquet as pq from .. import Audio, Dataset, Features, Image, NamedSplit, Value, config from ..features.features import FeatureType, _visit from ..formatting import query_table from ..packaged_modules import _PACKAGED_DATASETS_MODULES from ..packaged_modules.parquet.parquet import Parquet from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader def _lowerCamelCase( a ): __a = np.inf def set_batch_size(a ) -> None: nonlocal batch_size if isinstance(A__ , A__ ): __a = min(A__ , config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS ) elif isinstance(A__ , A__ ): __a = min(A__ , config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS ) elif isinstance(A__ , A__ ) and feature.dtype == "binary": __a = min(A__ , config.PARQUET_ROW_GROUP_SIZE_FOR_BINARY_DATASETS ) _visit(A__ , A__ ) return None if batch_size is np.inf else batch_size class snake_case__ ( __lowerCamelCase ): def __init__( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = False , lowerCamelCase = False , lowerCamelCase = None , **lowerCamelCase , ): super().__init__( UpperCamelCase_ , split=UpperCamelCase_ , features=UpperCamelCase_ , cache_dir=UpperCamelCase_ , keep_in_memory=UpperCamelCase_ , streaming=UpperCamelCase_ , num_proc=UpperCamelCase_ , **UpperCamelCase_ , ) __a = path_or_paths if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else {self.split: path_or_paths} __a = _PACKAGED_DATASETS_MODULES["parquet"][1] __a = Parquet( cache_dir=UpperCamelCase_ , data_files=UpperCamelCase_ , features=UpperCamelCase_ , hash=UpperCamelCase_ , **UpperCamelCase_ , ) def a__ ( self ): # Build iterable dataset if self.streaming: __a = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __a = None __a = None __a = None __a = None self.builder.download_and_prepare( download_config=UpperCamelCase_ , download_mode=UpperCamelCase_ , verification_mode=UpperCamelCase_ , base_path=UpperCamelCase_ , num_proc=self.num_proc , ) __a = self.builder.as_dataset( split=self.split , verification_mode=UpperCamelCase_ , in_memory=self.keep_in_memory ) return dataset class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = dataset __a = path_or_buf __a = batch_size or get_writer_batch_size(dataset.features ) __a = parquet_writer_kwargs def a__ ( self ): __a = self.batch_size if self.batch_size else config.DEFAULT_MAX_BATCH_SIZE if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with open(self.path_or_buf , "wb+" ) as buffer: __a = self._write(file_obj=UpperCamelCase_ , batch_size=UpperCamelCase_ , **self.parquet_writer_kwargs ) else: __a = self._write(file_obj=self.path_or_buf , batch_size=UpperCamelCase_ , **self.parquet_writer_kwargs ) return written def a__ ( self , lowerCamelCase , lowerCamelCase , **lowerCamelCase ): __a = 0 __a = parquet_writer_kwargs.pop("path_or_buf" , UpperCamelCase_ ) __a = self.dataset.features.arrow_schema __a = pq.ParquetWriter(UpperCamelCase_ , schema=UpperCamelCase_ , **UpperCamelCase_ ) for offset in logging.tqdm( range(0 , len(self.dataset ) , UpperCamelCase_ ) , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating parquet from Arrow format" , ): __a = query_table( table=self.dataset._data , key=slice(UpperCamelCase_ , offset + batch_size ) , indices=self.dataset._indices if self.dataset._indices is not None else None , ) writer.write_table(UpperCamelCase_ ) written += batch.nbytes writer.close() return written
357
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:int = { """microsoft/focalnet-tiny""": """https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json""", } class snake_case__ ( snake_case_, snake_case_ ): _snake_case : List[Any] = """focalnet""" def __init__( self , lowerCamelCase=224 , lowerCamelCase=4 , lowerCamelCase=3 , lowerCamelCase=96 , lowerCamelCase=False , lowerCamelCase=[192, 384, 768, 768] , lowerCamelCase=[2, 2, 6, 2] , lowerCamelCase=[2, 2, 2, 2] , lowerCamelCase=[3, 3, 3, 3] , lowerCamelCase="gelu" , lowerCamelCase=4.0 , lowerCamelCase=0.0 , lowerCamelCase=0.1 , lowerCamelCase=False , lowerCamelCase=1E-4 , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=0.02 , lowerCamelCase=1E-5 , lowerCamelCase=32 , lowerCamelCase=None , lowerCamelCase=None , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = image_size __a = patch_size __a = num_channels __a = embed_dim __a = use_conv_embed __a = hidden_sizes __a = depths __a = focal_levels __a = focal_windows __a = hidden_act __a = mlp_ratio __a = hidden_dropout_prob __a = drop_path_rate __a = use_layerscale __a = layerscale_value __a = use_post_layernorm __a = use_post_layernorm_in_modulation __a = normalize_modulator __a = initializer_range __a = layer_norm_eps __a = encoder_stride __a = ["stem"] + [F"stage{idx}" for idx in range(1 , len(self.depths ) + 1 )] __a , __a = get_aligned_output_features_output_indices( out_features=lowerCamelCase , out_indices=lowerCamelCase , stage_names=self.stage_names )
268
0
"""simple docstring""" from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class _UpperCamelCase : '''simple docstring''' pass
57
"""simple docstring""" def _lowerCamelCase ( _UpperCamelCase = 6008_5147_5143 ): '''simple docstring''' try: __lowerCAmelCase = int(_UpperCamelCase ) except (TypeError, ValueError): raise TypeError("Parameter n must be int or castable to int." ) if n <= 0: raise ValueError("Parameter n must be greater than or equal to one." ) __lowerCAmelCase = 2 __lowerCAmelCase = 0 if n == 2: return 2 while n > 2: while n % i != 0: i += 1 __lowerCAmelCase = i while n % i == 0: __lowerCAmelCase = n // i i += 1 return int(_UpperCamelCase ) if __name__ == "__main__": print(f'''{solution() = }''')
57
1
'''simple docstring''' def _a( ): '''simple docstring''' return 1 def _a( UpperCamelCase__ : int ): '''simple docstring''' return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def _a( UpperCamelCase__ : int ): '''simple docstring''' return 0 if x < 0 else five_pence(x - 5 ) + two_pence(UpperCamelCase__ ) def _a( UpperCamelCase__ : int ): '''simple docstring''' return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(UpperCamelCase__ ) def _a( UpperCamelCase__ : int ): '''simple docstring''' return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(UpperCamelCase__ ) def _a( UpperCamelCase__ : int ): '''simple docstring''' return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(UpperCamelCase__ ) def _a( UpperCamelCase__ : int ): '''simple docstring''' return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(UpperCamelCase__ ) def _a( UpperCamelCase__ : int ): '''simple docstring''' return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(UpperCamelCase__ ) def _a( UpperCamelCase__ : int = 2_0_0 ): '''simple docstring''' return two_pound(UpperCamelCase__ ) if __name__ == "__main__": print(solution(int(input().strip())))
222
'''simple docstring''' from __future__ import annotations def _a( UpperCamelCase__ : List[str], UpperCamelCase__ : Union[str, Any], UpperCamelCase__ : int, UpperCamelCase__ : List[str] ): # noqa: E741 '''simple docstring''' while r - l > 1: SCREAMING_SNAKE_CASE__ : List[str] =(l + r) // 2 if v[m] >= key: SCREAMING_SNAKE_CASE__ : Dict =m else: SCREAMING_SNAKE_CASE__ : Any =m # noqa: E741 return r def _a( UpperCamelCase__ : list[int] ): '''simple docstring''' if len(UpperCamelCase__ ) == 0: return 0 SCREAMING_SNAKE_CASE__ : List[Any] =[0] * len(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ : Dict =1 SCREAMING_SNAKE_CASE__ : Union[str, Any] =v[0] for i in range(1, len(UpperCamelCase__ ) ): if v[i] < tail[0]: SCREAMING_SNAKE_CASE__ : List[Any] =v[i] elif v[i] > tail[length - 1]: SCREAMING_SNAKE_CASE__ : int =v[i] length += 1 else: SCREAMING_SNAKE_CASE__ : Union[str, Any] =v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
222
1
import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class lowercase__( unittest.TestCase ): """simple docstring""" def _lowercase ( self : str ) -> Optional[int]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase ( self : Optional[int] ) -> Optional[Any]: lowercase_ = 1 lowercase_ = 3 lowercase_ = (3_2, 3_2) lowercase_ = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE_ ) return image @property def _lowercase ( self : Optional[int] ) -> Tuple: torch.manual_seed(0 ) lowercase_ = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , ) return model @property def _lowercase ( self : Optional[int] ) -> int: torch.manual_seed(0 ) lowercase_ = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def _lowercase ( self : Dict ) -> Union[str, Any]: torch.manual_seed(0 ) lowercase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModel(SCREAMING_SNAKE_CASE_ ) @property def _lowercase ( self : int ) -> Optional[int]: def extract(*SCREAMING_SNAKE_CASE_ : List[Any] , **SCREAMING_SNAKE_CASE_ : str ): class lowercase__: """simple docstring""" def __init__( self : Tuple ) -> Optional[int]: lowercase_ = torch.ones([0] ) def _lowercase ( self : Dict , SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Any: self.pixel_values.to(SCREAMING_SNAKE_CASE_ ) return self return Out() return extract def _lowercase ( self : Dict ) -> List[Any]: lowercase_ = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase_ = self.dummy_cond_unet lowercase_ = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=SCREAMING_SNAKE_CASE_ , set_alpha_to_one=SCREAMING_SNAKE_CASE_ , ) lowercase_ = self.dummy_vae lowercase_ = self.dummy_text_encoder lowercase_ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) # make sure here that pndm scheduler skips prk lowercase_ = StableDiffusionPipeline( unet=SCREAMING_SNAKE_CASE_ , scheduler=SCREAMING_SNAKE_CASE_ , vae=SCREAMING_SNAKE_CASE_ , text_encoder=SCREAMING_SNAKE_CASE_ , tokenizer=SCREAMING_SNAKE_CASE_ , safety_checker=SCREAMING_SNAKE_CASE_ , feature_extractor=self.dummy_extractor , ) lowercase_ = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) lowercase_ = "A painting of a squirrel eating a burger" lowercase_ = torch.Generator(device=SCREAMING_SNAKE_CASE_ ).manual_seed(0 ) lowercase_ = sd_pipe([prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' ) lowercase_ = output.images lowercase_ = torch.Generator(device=SCREAMING_SNAKE_CASE_ ).manual_seed(0 ) lowercase_ = sd_pipe( [prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , return_dict=SCREAMING_SNAKE_CASE_ , )[0] lowercase_ = image[0, -3:, -3:, -1] lowercase_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) lowercase_ = np.array([0.57_56, 0.61_18, 0.50_05, 0.50_41, 0.54_71, 0.47_26, 0.49_76, 0.48_65, 0.48_64] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self : int ) -> Optional[int]: lowercase_ = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase_ = self.dummy_cond_unet lowercase_ = PNDMScheduler(skip_prk_steps=SCREAMING_SNAKE_CASE_ ) lowercase_ = self.dummy_vae lowercase_ = self.dummy_text_encoder lowercase_ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) # make sure here that pndm scheduler skips prk lowercase_ = StableDiffusionPipeline( unet=SCREAMING_SNAKE_CASE_ , scheduler=SCREAMING_SNAKE_CASE_ , vae=SCREAMING_SNAKE_CASE_ , text_encoder=SCREAMING_SNAKE_CASE_ , tokenizer=SCREAMING_SNAKE_CASE_ , safety_checker=SCREAMING_SNAKE_CASE_ , feature_extractor=self.dummy_extractor , ) lowercase_ = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) lowercase_ = "A painting of a squirrel eating a burger" lowercase_ = torch.Generator(device=SCREAMING_SNAKE_CASE_ ).manual_seed(0 ) lowercase_ = sd_pipe([prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' ) lowercase_ = output.images lowercase_ = torch.Generator(device=SCREAMING_SNAKE_CASE_ ).manual_seed(0 ) lowercase_ = sd_pipe( [prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , return_dict=SCREAMING_SNAKE_CASE_ , )[0] lowercase_ = image[0, -3:, -3:, -1] lowercase_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) lowercase_ = np.array([0.51_25, 0.57_16, 0.48_28, 0.50_60, 0.56_50, 0.47_68, 0.51_85, 0.48_95, 0.49_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self : Optional[Any] ) -> Any: lowercase_ = StableDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-lms-pipe''' , safety_checker=SCREAMING_SNAKE_CASE_ ) assert isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) assert isinstance(pipe.scheduler , SCREAMING_SNAKE_CASE_ ) assert pipe.safety_checker is None lowercase_ = pipe('''example prompt''' , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(SCREAMING_SNAKE_CASE_ ) lowercase_ = StableDiffusionPipeline.from_pretrained(SCREAMING_SNAKE_CASE_ ) # sanity check that the pipeline still works assert pipe.safety_checker is None lowercase_ = pipe('''example prompt''' , num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def _lowercase ( self : Any ) -> List[Any]: lowercase_ = self.dummy_cond_unet lowercase_ = PNDMScheduler(skip_prk_steps=SCREAMING_SNAKE_CASE_ ) lowercase_ = self.dummy_vae lowercase_ = self.dummy_text_encoder lowercase_ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) # put models in fp16 lowercase_ = unet.half() lowercase_ = vae.half() lowercase_ = bert.half() # make sure here that pndm scheduler skips prk lowercase_ = StableDiffusionPipeline( unet=SCREAMING_SNAKE_CASE_ , scheduler=SCREAMING_SNAKE_CASE_ , vae=SCREAMING_SNAKE_CASE_ , text_encoder=SCREAMING_SNAKE_CASE_ , tokenizer=SCREAMING_SNAKE_CASE_ , safety_checker=SCREAMING_SNAKE_CASE_ , feature_extractor=self.dummy_extractor , ) lowercase_ = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) lowercase_ = "A painting of a squirrel eating a burger" lowercase_ = sd_pipe([prompt] , num_inference_steps=2 , output_type='''np''' ).images assert image.shape == (1, 6_4, 6_4, 3) @nightly @require_torch_gpu class lowercase__( unittest.TestCase ): """simple docstring""" def _lowercase ( self : int ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : str ) -> Tuple: lowercase_ = StableDiffusionPipeline.from_pretrained('''runwayml/stable-diffusion-v1-5''' , safety_checker=SCREAMING_SNAKE_CASE_ ) lowercase_ = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowercase_ = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) lowercase_ = ( "portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle" " coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with" " anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and" " children from bahnhof zoo, detailed " ) lowercase_ = 4_0_0_3_6_6_0_3_4_6 lowercase_ = 7 # without safety guidance (sld_guidance_scale = 0) lowercase_ = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) lowercase_ = sd_pipe( [prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=SCREAMING_SNAKE_CASE_ , num_inference_steps=5_0 , output_type='''np''' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=0 , ) lowercase_ = output.images lowercase_ = image[0, -3:, -3:, -1] lowercase_ = [0.22_78, 0.22_31, 0.22_49, 0.23_33, 0.23_03, 0.18_85, 0.22_73, 0.21_44, 0.21_76] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 # without safety guidance (strong configuration) lowercase_ = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) lowercase_ = sd_pipe( [prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=SCREAMING_SNAKE_CASE_ , num_inference_steps=5_0 , output_type='''np''' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=2_0_0_0 , sld_warmup_steps=7 , sld_threshold=0.0_25 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowercase_ = output.images lowercase_ = image[0, -3:, -3:, -1] lowercase_ = [0.23_83, 0.22_76, 0.2_36, 0.21_92, 0.21_86, 0.20_53, 0.19_71, 0.19_01, 0.17_19] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: lowercase_ = StableDiffusionPipeline.from_pretrained('''runwayml/stable-diffusion-v1-5''' , safety_checker=SCREAMING_SNAKE_CASE_ ) lowercase_ = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowercase_ = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) lowercase_ = "padme amidala taking a bath artwork, safe for work, no nudity" lowercase_ = 2_7_3_4_9_7_1_7_5_5 lowercase_ = 7 lowercase_ = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) lowercase_ = sd_pipe( [prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=SCREAMING_SNAKE_CASE_ , num_inference_steps=5_0 , output_type='''np''' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=0 , ) lowercase_ = output.images lowercase_ = image[0, -3:, -3:, -1] lowercase_ = [0.35_02, 0.36_22, 0.33_96, 0.36_42, 0.34_78, 0.33_18, 0.35, 0.33_48, 0.32_97] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 lowercase_ = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) lowercase_ = sd_pipe( [prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=SCREAMING_SNAKE_CASE_ , num_inference_steps=5_0 , output_type='''np''' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=2_0_0_0 , sld_warmup_steps=7 , sld_threshold=0.0_25 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowercase_ = output.images lowercase_ = image[0, -3:, -3:, -1] lowercase_ = [0.55_31, 0.52_06, 0.48_95, 0.51_56, 0.51_82, 0.47_51, 0.48_02, 0.48_03, 0.44_43] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self : Optional[Any] ) -> List[str]: lowercase_ = StableDiffusionPipeline.from_pretrained('''runwayml/stable-diffusion-v1-5''' ) lowercase_ = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) lowercase_ = ( "the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c." " leyendecker" ) lowercase_ = 1_0_4_4_3_5_5_2_3_4 lowercase_ = 1_2 lowercase_ = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) lowercase_ = sd_pipe( [prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=SCREAMING_SNAKE_CASE_ , num_inference_steps=5_0 , output_type='''np''' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=0 , ) lowercase_ = output.images lowercase_ = image[0, -3:, -3:, -1] lowercase_ = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-7 lowercase_ = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) lowercase_ = sd_pipe( [prompt] , generator=SCREAMING_SNAKE_CASE_ , guidance_scale=SCREAMING_SNAKE_CASE_ , num_inference_steps=5_0 , output_type='''np''' , width=5_1_2 , height=5_1_2 , sld_guidance_scale=2_0_0_0 , sld_warmup_steps=7 , sld_threshold=0.0_25 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowercase_ = output.images lowercase_ = image[0, -3:, -3:, -1] lowercase_ = np.array([0.58_18, 0.62_85, 0.68_35, 0.60_19, 0.6_25, 0.67_54, 0.60_96, 0.63_34, 0.65_61] ) assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
30
from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar lowercase : Optional[Any] = TypeVar('T') class lowerCamelCase__ ( Generic[T]): '''simple docstring''' _A = 42 # Cache store of keys _A = 42 # References of the keys in cache _A = 1_0 # Maximum capacity of cache def __init__( self :Optional[Any] , a :int ) -> None: __UpperCamelCase : Union[str, Any] = deque() __UpperCamelCase : str = set() if not n: __UpperCamelCase : Union[str, Any] = sys.maxsize elif n < 0: raise ValueError("n should be an integer greater than 0." ) else: __UpperCamelCase : Any = n def _lowerCamelCase ( self :Tuple , a :T ) -> None: if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: __UpperCamelCase : int = self.dq_store.pop() self.key_reference.remove(a ) else: self.dq_store.remove(a ) self.dq_store.appendleft(a ) self.key_reference.add(a ) def _lowerCamelCase ( self :Any ) -> None: for k in self.dq_store: print(a ) def __repr__( self :Tuple ) -> str: return f'LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}' if __name__ == "__main__": import doctest doctest.testmod() lowercase : LRUCache[str | int] = LRUCache(4) lru_cache.refer('A') lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer('A') lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
232
0
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def lowerCAmelCase_ ( ) -> Optional[Any]: UpperCamelCase_ = ArgumentParser( description=( "PyTorch TPU distributed training launch helper utility that will spawn up multiple distributed processes" ) ) # Optional arguments for the launch helper parser.add_argument("--num_cores" , type=UpperCamelCase_ , default=1 , help="Number of TPU cores to use (1 or 8)." ) # positional parser.add_argument( "training_script" , type=UpperCamelCase_ , help=( "The full path to the single TPU training " "program/script to be launched in parallel, " "followed by all the arguments for the " "training script" ) , ) # rest from the training program parser.add_argument("training_script_args" , nargs=UpperCamelCase_ ) return parser.parse_args() def lowerCAmelCase_ ( ) -> Optional[Any]: UpperCamelCase_ = parse_args() # Import training_script as a module. UpperCamelCase_ = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) UpperCamelCase_ = script_fpath.stem UpperCamelCase_ = importlib.import_module(UpperCamelCase_ ) # Patch sys.argv UpperCamelCase_ = [args.training_script] + args.training_script_args + ["--tpu_num_cores", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
354
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): _UpperCamelCase : List[Any] = IFImgaImgSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''width''', '''height'''} _UpperCamelCase : List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'''original_image'''} ) _UpperCamelCase : List[Any] = PipelineTesterMixin.required_optional_params - {'''latents'''} def lowercase ( self: List[str] ) -> Any: """simple docstring""" return self._get_superresolution_dummy_components() def lowercase ( self: Any , _SCREAMING_SNAKE_CASE: Any , _SCREAMING_SNAKE_CASE: Optional[int]=0 ) -> List[Any]: """simple docstring""" if str(_SCREAMING_SNAKE_CASE ).startswith("mps" ): UpperCamelCase_ = torch.manual_seed(_SCREAMING_SNAKE_CASE ) else: UpperCamelCase_ = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = floats_tensor((1, 3, 32, 32) , rng=random.Random(_SCREAMING_SNAKE_CASE ) ).to(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = floats_tensor((1, 3, 16, 16) , rng=random.Random(_SCREAMING_SNAKE_CASE ) ).to(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowercase ( self: Any ) -> Union[str, Any]: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def lowercase ( self: int ) -> Tuple: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def lowercase ( self: Optional[Any] ) -> Union[str, Any]: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1e-1 ) def lowercase ( self: List[Any] ) -> Union[str, Any]: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def lowercase ( self: Dict ) -> Any: """simple docstring""" self._test_save_load_local() def lowercase ( self: Any ) -> Dict: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
328
0
from cva import destroyAllWindows, imread, imshow, waitKey def UpperCAmelCase_ ( _A ): '''simple docstring''' SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = img.shape[0], img.shape[1] # converting each pixel's color to its negative for i in range(_A ): for j in range(_A ): SCREAMING_SNAKE_CASE__ = [2_55, 2_55, 2_55] - img[i][j] return img if __name__ == "__main__": # read original image _SCREAMING_SNAKE_CASE : Optional[int] = imread('''image_data/lena.jpg''', 1) # convert to its negative _SCREAMING_SNAKE_CASE : Union[str, Any] = convert_to_negative(img) # show result image imshow('''negative of original image''', img) waitKey(0) destroyAllWindows()
314
from ...configuration_utils import PretrainedConfig _SCREAMING_SNAKE_CASE : Optional[Any] = { '''google/tapas-base-finetuned-sqa''': ( '''https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json''' ), '''google/tapas-base-finetuned-wtq''': ( '''https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json''' ), '''google/tapas-base-finetuned-wikisql-supervised''': ( '''https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json''' ), '''google/tapas-base-finetuned-tabfact''': ( '''https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json''' ), } class UpperCAmelCase__ ( A__ ): """simple docstring""" a = "tapas" def __init__( self : int , __lowerCamelCase : Optional[Any]=3_0522 , __lowerCamelCase : Tuple=768 , __lowerCamelCase : int=12 , __lowerCamelCase : Any=12 , __lowerCamelCase : Union[str, Any]=3072 , __lowerCamelCase : Optional[int]="gelu" , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : Dict=0.1 , __lowerCamelCase : str=1024 , __lowerCamelCase : Union[str, Any]=[3, 256, 256, 2, 256, 256, 10] , __lowerCamelCase : Optional[int]=0.02 , __lowerCamelCase : List[str]=1e-12 , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Optional[Any]=10.0 , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : str=1.0 , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : int=1.0 , __lowerCamelCase : Dict=1.0 , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : int=False , __lowerCamelCase : List[str]="ratio" , __lowerCamelCase : Tuple=None , __lowerCamelCase : List[Any]=None , __lowerCamelCase : List[Any]=64 , __lowerCamelCase : Any=32 , __lowerCamelCase : Tuple=False , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Tuple=False , __lowerCamelCase : Tuple=True , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Optional[Any]=None , **__lowerCamelCase : str , ) -> str: super().__init__(pad_token_id=__lowerCamelCase , **__lowerCamelCase ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_sizes SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps # Fine-tuning task hyperparameters SCREAMING_SNAKE_CASE__ = positive_label_weight SCREAMING_SNAKE_CASE__ = num_aggregation_labels SCREAMING_SNAKE_CASE__ = aggregation_loss_weight SCREAMING_SNAKE_CASE__ = use_answer_as_supervision SCREAMING_SNAKE_CASE__ = answer_loss_importance SCREAMING_SNAKE_CASE__ = use_normalized_answer_loss SCREAMING_SNAKE_CASE__ = huber_loss_delta SCREAMING_SNAKE_CASE__ = temperature SCREAMING_SNAKE_CASE__ = aggregation_temperature SCREAMING_SNAKE_CASE__ = use_gumbel_for_cells SCREAMING_SNAKE_CASE__ = use_gumbel_for_aggregation SCREAMING_SNAKE_CASE__ = average_approximation_function SCREAMING_SNAKE_CASE__ = cell_selection_preference SCREAMING_SNAKE_CASE__ = answer_loss_cutoff SCREAMING_SNAKE_CASE__ = max_num_rows SCREAMING_SNAKE_CASE__ = max_num_columns SCREAMING_SNAKE_CASE__ = average_logits_per_cell SCREAMING_SNAKE_CASE__ = select_one_column SCREAMING_SNAKE_CASE__ = allow_empty_column_selection SCREAMING_SNAKE_CASE__ = init_cell_selection_weights_to_zero SCREAMING_SNAKE_CASE__ = reset_position_index_per_cell SCREAMING_SNAKE_CASE__ = disable_per_token_loss # Aggregation hyperparameters SCREAMING_SNAKE_CASE__ = aggregation_labels SCREAMING_SNAKE_CASE__ = no_aggregation_label_index if isinstance(self.aggregation_labels , __lowerCamelCase ): SCREAMING_SNAKE_CASE__ = {int(__lowerCamelCase ): v for k, v in aggregation_labels.items()}
314
1
def UpperCamelCase ( __magic_name__ : str ) -> bool: """simple docstring""" lowercase__ = [int(__magic_name__ ) for i in ip_va_address.split(""".""" ) if i.isdigit()] return len(__magic_name__ ) == 4 and all(0 <= int(__magic_name__ ) <= 254 for octet in octets ) if __name__ == "__main__": A : List[str] = input().strip() A : int = 'valid' if is_ip_va_address_valid(ip) else 'invalid' print(F'{ip} is a {valid_or_invalid} IP v4 address.')
146
import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor A : List[Any] = logging.get_logger(__name__) class A ( UpperCAmelCase__ ): '''simple docstring''' def __init__(self : List[Any] , *_UpperCAmelCase : int , **_UpperCAmelCase : List[str] ) -> None: """simple docstring""" warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , _UpperCAmelCase , ) super().__init__(*_UpperCAmelCase , **_UpperCAmelCase )
146
1
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE :Optional[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE :str = { 'huggingface/time-series-transformer-tourism-monthly': ( 'https://huggingface.co/huggingface/time-series-transformer-tourism-monthly/resolve/main/config.json' ), # See all TimeSeriesTransformer models at https://huggingface.co/models?filter=time_series_transformer } class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case_ = "time_series_transformer" snake_case_ = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", "num_hidden_layers": "encoder_layers", } def __init__( self : Tuple ,A : Optional[int] = None ,A : Optional[int] = None ,A : str = "student_t" ,A : str = "nll" ,A : int = 1 ,A : List[int] = [1, 2, 3, 4, 5, 6, 7] ,A : Optional[Union[str, bool]] = "mean" ,A : int = 0 ,A : int = 0 ,A : int = 0 ,A : int = 0 ,A : Optional[List[int]] = None ,A : Optional[List[int]] = None ,A : int = 32 ,A : int = 32 ,A : int = 2 ,A : int = 2 ,A : int = 2 ,A : int = 2 ,A : bool = True ,A : str = "gelu" ,A : int = 64 ,A : float = 0.1 ,A : float = 0.1 ,A : float = 0.1 ,A : float = 0.1 ,A : float = 0.1 ,A : int = 1_00 ,A : float = 0.02 ,A : Union[str, Any]=True ,**A : Optional[int] ,): # time series specific configuration __A = prediction_length __A = context_length or prediction_length __A = distribution_output __A = loss __A = input_size __A = num_time_features __A = lags_sequence __A = scaling __A = num_dynamic_real_features __A = num_static_real_features __A = num_static_categorical_features if cardinality and num_static_categorical_features > 0: if len(A ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __A = cardinality else: __A = [0] if embedding_dimension and num_static_categorical_features > 0: if len(A ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __A = embedding_dimension else: __A = [min(50 ,(cat + 1) // 2 ) for cat in self.cardinality] __A = num_parallel_samples # Transformer architecture configuration __A = input_size * len(A ) + self._number_of_features __A = d_model __A = encoder_attention_heads __A = decoder_attention_heads __A = encoder_ffn_dim __A = decoder_ffn_dim __A = encoder_layers __A = decoder_layers __A = dropout __A = attention_dropout __A = activation_dropout __A = encoder_layerdrop __A = decoder_layerdrop __A = activation_function __A = init_std __A = use_cache super().__init__(is_encoder_decoder=A ,**A ) @property def UpperCamelCase_ ( self : Optional[Any] ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
15
import importlib.util import json import os import warnings from dataclasses import dataclass, field import torch from ..training_args import TrainingArguments from ..utils import cached_property, is_sagemaker_dp_enabled, logging __A =logging.get_logger(__name__) def lowerCamelCase_ ( ): # Get the sagemaker specific mp parameters from smp_options variable. lowerCamelCase_ = os.getenv("SM_HP_MP_PARAMETERS" , "{}" ) try: # Parse it and check the field "partitions" is included, it is required for model parallel. lowerCamelCase_ = json.loads(lowerCamelCase__ ) if "partitions" not in smp_options: return False except json.JSONDecodeError: return False # Get the sagemaker specific framework parameters from mpi_options variable. lowerCamelCase_ = os.getenv("SM_FRAMEWORK_PARAMS" , "{}" ) try: # Parse it and check the field "sagemaker_distributed_dataparallel_enabled". lowerCamelCase_ = json.loads(lowerCamelCase__ ) if not mpi_options.get("sagemaker_mpi_enabled" , lowerCamelCase__ ): return False except json.JSONDecodeError: return False # Lastly, check if the `smdistributed` module is present. return importlib.util.find_spec("smdistributed" ) is not None if is_sagemaker_model_parallel_available(): import smdistributed.modelparallel.torch as smp smp.init() @dataclass class _SCREAMING_SNAKE_CASE ( snake_case_ ): lowerCAmelCase__ = field( default='' , metadata={'help': 'Used by the SageMaker launcher to send mp-specific args. Ignored in SageMakerTrainer'} , ) def SCREAMING_SNAKE_CASE_( self ) -> Tuple: super().__post_init__() warnings.warn( "`SageMakerTrainingArguments` is deprecated and will be removed in v5 of Transformers. You can use " "`TrainingArguments` instead." , lowercase , ) @cached_property def SCREAMING_SNAKE_CASE_( self ) -> "torch.device": logger.info("PyTorch: setting up devices" ) if torch.distributed.is_available() and torch.distributed.is_initialized() and self.local_rank == -1: logger.warning( "torch.distributed process group is initialized, but local_rank == -1. " "In order to use Torch DDP, launch your script with `python -m torch.distributed.launch" ) if self.no_cuda: lowerCamelCase_ = torch.device("cpu" ) lowerCamelCase_ = 0 elif is_sagemaker_model_parallel_available(): lowerCamelCase_ = smp.local_rank() lowerCamelCase_ = torch.device("cuda" , lowercase ) lowerCamelCase_ = 1 elif is_sagemaker_dp_enabled(): import smdistributed.dataparallel.torch.torch_smddp # noqa: F401 torch.distributed.init_process_group(backend="smddp" , timeout=self.ddp_timeout_delta ) lowerCamelCase_ = int(os.getenv("SMDATAPARALLEL_LOCAL_RANK" ) ) lowerCamelCase_ = torch.device("cuda" , self.local_rank ) lowerCamelCase_ = 1 elif self.local_rank == -1: # if n_gpu is > 1 we'll use nn.DataParallel. # If you only want to use a specific subset of GPUs use `CUDA_VISIBLE_DEVICES=0` # Explicitly set CUDA to the first (index 0) CUDA device, otherwise `set_device` will # trigger an error that a device index is missing. Index 0 takes into account the # GPUs available in the environment, so `CUDA_VISIBLE_DEVICES=1,2` with `cuda:0` # will use the first GPU in that env, i.e. GPU#1 lowerCamelCase_ = torch.device("cuda:0" if torch.cuda.is_available() else "cpu" ) # Sometimes the line in the postinit has not been run before we end up here, so just checking we're not at # the default value. lowerCamelCase_ = torch.cuda.device_count() else: # Here, we'll use torch.distributed. # Initializes the distributed backend which will take care of synchronizing nodes/GPUs if not torch.distributed.is_initialized(): torch.distributed.init_process_group(backend="nccl" , timeout=self.ddp_timeout_delta ) lowerCamelCase_ = torch.device("cuda" , self.local_rank ) lowerCamelCase_ = 1 if device.type == "cuda": torch.cuda.set_device(lowercase ) return device @property def SCREAMING_SNAKE_CASE_( self ) -> Tuple: if is_sagemaker_model_parallel_available(): return smp.dp_size() return super().world_size @property def SCREAMING_SNAKE_CASE_( self ) -> List[str]: return not is_sagemaker_model_parallel_available() @property def SCREAMING_SNAKE_CASE_( self ) -> Dict: return False
19
0
"""simple docstring""" from collections.abc import Generator from math import sin def UpperCAmelCase__ (snake_case__ : bytes ): """simple docstring""" if len(snake_case__ ) != 32: raise ValueError("""Input must be of length 32""" ) _snake_case : Optional[int] = B"""""" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" if i < 0: raise ValueError("""Input must be non-negative""" ) _snake_case : Optional[Any] = format(snake_case__ , """08x""" )[-8:] _snake_case : Optional[Any] = B"""""" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("""utf-8""" ) return little_endian_hex def UpperCAmelCase__ (snake_case__ : bytes ): """simple docstring""" _snake_case : Union[str, Any] = B"""""" for char in message: bit_string += format(snake_case__ , """08b""" ).encode("""utf-8""" ) _snake_case : List[Any] = format(len(snake_case__ ) , """064b""" ).encode("""utf-8""" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(snake_case__ ) % 5_12 != 4_48: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def UpperCAmelCase__ (snake_case__ : bytes ): """simple docstring""" if len(snake_case__ ) % 5_12 != 0: raise ValueError("""Input must have length that's a multiple of 512""" ) for pos in range(0 , len(snake_case__ ) , 5_12 ): _snake_case : List[str] = bit_string[pos : pos + 5_12] _snake_case : List[str] = [] for i in range(0 , 5_12 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def UpperCAmelCase__ (snake_case__ : int ): """simple docstring""" if i < 0: raise ValueError("""Input must be non-negative""" ) _snake_case : Optional[int] = format(snake_case__ , """032b""" ) _snake_case : str = """""" for c in i_str: new_str += "1" if c == "0" else "0" return int(snake_case__ , 2 ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : int ): """simple docstring""" return (a + b) % 2**32 def UpperCAmelCase__ (snake_case__ : int , snake_case__ : int ): """simple docstring""" if i < 0: raise ValueError("""Input must be non-negative""" ) if shift < 0: raise ValueError("""Shift must be non-negative""" ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCAmelCase__ (snake_case__ : bytes ): """simple docstring""" _snake_case : Any = preprocess(snake_case__ ) _snake_case : Optional[Any] = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states _snake_case : Union[str, Any] = 0x6745_2301 _snake_case : List[Any] = 0xEFCD_AB89 _snake_case : Optional[Any] = 0x98BA_DCFE _snake_case : Optional[int] = 0x1032_5476 _snake_case : Tuple = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(snake_case__ ): _snake_case : Tuple = aa _snake_case : str = ba _snake_case : int = ca _snake_case : Dict = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f _snake_case : int = d ^ (b & (c ^ d)) _snake_case : Dict = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f _snake_case : Tuple = c ^ (d & (b ^ c)) _snake_case : int = (5 * i + 1) % 16 elif i <= 47: _snake_case : List[Any] = b ^ c ^ d _snake_case : Union[str, Any] = (3 * i + 5) % 16 else: _snake_case : Tuple = c ^ (b | not_aa(snake_case__ )) _snake_case : Optional[int] = (7 * i) % 16 _snake_case : Dict = (f + a + added_consts[i] + block_words[g]) % 2**32 _snake_case : List[str] = d _snake_case : List[Any] = c _snake_case : str = b _snake_case : List[str] = sum_aa(snake_case__ , left_rotate_aa(snake_case__ , shift_amounts[i] ) ) # Add hashed chunk to running total _snake_case : Union[str, Any] = sum_aa(snake_case__ , snake_case__ ) _snake_case : str = sum_aa(snake_case__ , snake_case__ ) _snake_case : Any = sum_aa(snake_case__ , snake_case__ ) _snake_case : List[str] = sum_aa(snake_case__ , snake_case__ ) _snake_case : Any = reformat_hex(snake_case__ ) + reformat_hex(snake_case__ ) + reformat_hex(snake_case__ ) + reformat_hex(snake_case__ ) return digest if __name__ == "__main__": import doctest doctest.testmod()
132
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A_ = logging.get_logger(__name__) A_ = { '''microsoft/swin-tiny-patch4-window7-224''': ( '''https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json''' ), # See all Swin models at https://huggingface.co/models?filter=swin } class lowercase( __a , __a ): '''simple docstring''' lowercase__ = "swin" lowercase__ = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self: Any, a_: List[str]=224, a_: List[Any]=4, a_: List[Any]=3, a_: Dict=96, a_: List[str]=[2, 2, 6, 2], a_: int=[3, 6, 12, 24], a_: int=7, a_: str=4.0, a_: Optional[Any]=True, a_: Dict=0.0, a_: List[Any]=0.0, a_: List[str]=0.1, a_: Union[str, Any]="gelu", a_: Dict=False, a_: Union[str, Any]=0.02, a_: Optional[int]=1E-5, a_: Optional[int]=32, a_: Tuple=None, a_: Union[str, Any]=None, **a_: Any, ): '''simple docstring''' super().__init__(**a_ ) _snake_case : Any = image_size _snake_case : List[Any] = patch_size _snake_case : Tuple = num_channels _snake_case : str = embed_dim _snake_case : Union[str, Any] = depths _snake_case : int = len(a_ ) _snake_case : Union[str, Any] = num_heads _snake_case : List[str] = window_size _snake_case : str = mlp_ratio _snake_case : Union[str, Any] = qkv_bias _snake_case : Dict = hidden_dropout_prob _snake_case : str = attention_probs_dropout_prob _snake_case : Union[str, Any] = drop_path_rate _snake_case : Optional[int] = hidden_act _snake_case : str = use_absolute_embeddings _snake_case : Tuple = layer_norm_eps _snake_case : List[Any] = initializer_range _snake_case : Optional[Any] = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _snake_case : Any = int(embed_dim * 2 ** (len(a_ ) - 1) ) _snake_case : Any = ["""stem"""] + [f"stage{idx}" for idx in range(1, len(a_ ) + 1 )] _snake_case , _snake_case : List[str] = get_aligned_output_features_output_indices( out_features=a_, out_indices=a_, stage_names=self.stage_names ) class lowercase( __a ): '''simple docstring''' lowercase__ = version.parse("1.11" ) @property def UpperCamelCase_ ( self: Any ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' return 1E-4
132
1
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer SCREAMING_SNAKE_CASE_: Union[str, Any] =['gpt2'] SCREAMING_SNAKE_CASE_: Optional[int] ='gpt2' if is_tf_available(): class __A ( tf.Module ): def __init__(self : Union[str, Any] , __a : Union[str, Any] ): super().__init__() UpperCAmelCase_ = tokenizer UpperCAmelCase_ = AutoConfig.from_pretrained(__a ) UpperCAmelCase_ = TFGPTaLMHeadModel.from_config(__a ) @tf.function(input_signature=(tf.TensorSpec((None,) , tf.string , name="text" ),) ) def _lowercase (self : Tuple , __a : List[str] ): UpperCAmelCase_ = self.tokenizer(__a ) UpperCAmelCase_ = tokenized["input_ids"].to_tensor() UpperCAmelCase_ = tf.cast(input_ids_dense > 0 , tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) UpperCAmelCase_ = self.model(input_ids=__a , attention_mask=__a )["logits"] return outputs @require_tf @require_keras_nlp class __A ( unittest.TestCase ): def _lowercase (self : List[Any] ): super().setUp() UpperCAmelCase_ = [GPTaTokenizer.from_pretrained(__a ) for checkpoint in (TOKENIZER_CHECKPOINTS)] UpperCAmelCase_ = [TFGPTaTokenizer.from_pretrained(__a ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) UpperCAmelCase_ = [ "This is a straightforward English test sentence.", "This one has some weird characters\rto\nsee\r\nif those\u00E9break things.", "Now we're going to add some Chinese: 一 二 三 一二三", "And some much more rare Chinese: 齉 堃 齉堃", "Je vais aussi écrire en français pour tester les accents", "Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ", ] UpperCAmelCase_ = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def _lowercase (self : List[str] ): for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in self.test_sentences: UpperCAmelCase_ = tokenizer([test_inputs] , return_tensors="tf" ) UpperCAmelCase_ = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors UpperCAmelCase_ = python_outputs[key].numpy() UpperCAmelCase_ = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(__a , tf.intaa ) == tf_outputs_values ) ) @slow def _lowercase (self : Optional[Any] ): for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase_ = tf.function(__a ) for test_inputs in self.test_sentences: UpperCAmelCase_ = tf.constant(__a ) UpperCAmelCase_ = compiled_tokenizer(__a ) UpperCAmelCase_ = tf_tokenizer(__a ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def _lowercase (self : Tuple ): for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase_ = ModelToSave(tokenizer=__a ) UpperCAmelCase_ = tf.convert_to_tensor([self.test_sentences[0]] ) UpperCAmelCase_ = model.serving(__a ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: UpperCAmelCase_ = Path(__a ) / "saved.model" tf.saved_model.save(__a , __a , signatures={"serving_default": model.serving} ) UpperCAmelCase_ = tf.saved_model.load(__a ) UpperCAmelCase_ = loaded_model.signatures["serving_default"](__a )["output_0"] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def _lowercase (self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase_ = tf.convert_to_tensor([self.test_sentences[0]] ) UpperCAmelCase_ = tf_tokenizer(__a ) # Build model with some sample inputs UpperCAmelCase_ = tf_tokenizer.get_config() UpperCAmelCase_ = TFGPTaTokenizer.from_config(__a ) UpperCAmelCase_ = model_from_config(__a ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def _lowercase (self : Tuple ): for tf_tokenizer in self.tf_tokenizers: # for the test to run UpperCAmelCase_ = 123123 for max_length in [3, 5, 1024]: UpperCAmelCase_ = tf.convert_to_tensor([self.test_sentences[0]] ) UpperCAmelCase_ = tf_tokenizer(__a , max_length=__a ) UpperCAmelCase_ = out["input_ids"].numpy().shape[1] assert out_length == max_length
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class __A ( unittest.TestCase ): def _lowercase (self : Tuple ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase (self : str ): UpperCAmelCase_ = 1 UpperCAmelCase_ = 3 UpperCAmelCase_ = (32, 32) UpperCAmelCase_ = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(__a ) return image @property def _lowercase (self : int ): torch.manual_seed(0 ) UpperCAmelCase_ = UNetaDConditionModel( block_out_channels=(32, 32, 64) , layers_per_block=2 , sample_size=32 , in_channels=7 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , attention_head_dim=8 , use_linear_projection=__a , only_cross_attention=(True, True, False) , num_class_embeds=100 , ) return model @property def _lowercase (self : Any ): torch.manual_seed(0 ) UpperCAmelCase_ = AutoencoderKL( block_out_channels=[32, 32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def _lowercase (self : Optional[Any] ): torch.manual_seed(0 ) UpperCAmelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=512 , ) return CLIPTextModel(__a ) def _lowercase (self : Any ): UpperCAmelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ = self.dummy_cond_unet_upscale UpperCAmelCase_ = DDPMScheduler() UpperCAmelCase_ = DDIMScheduler(prediction_type="v_prediction" ) UpperCAmelCase_ = self.dummy_vae UpperCAmelCase_ = self.dummy_text_encoder UpperCAmelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCAmelCase_ = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCAmelCase_ = Image.fromarray(np.uinta(__a ) ).convert("RGB" ).resize((64, 64) ) # make sure here that pndm scheduler skips prk UpperCAmelCase_ = StableDiffusionUpscalePipeline( unet=__a , low_res_scheduler=__a , scheduler=__a , vae=__a , text_encoder=__a , tokenizer=__a , max_noise_level=350 , ) UpperCAmelCase_ = sd_pipe.to(__a ) sd_pipe.set_progress_bar_config(disable=__a ) UpperCAmelCase_ = "A painting of a squirrel eating a burger" UpperCAmelCase_ = torch.Generator(device=__a ).manual_seed(0 ) UpperCAmelCase_ = sd_pipe( [prompt] , image=__a , generator=__a , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="np" , ) UpperCAmelCase_ = output.images UpperCAmelCase_ = torch.Generator(device=__a ).manual_seed(0 ) UpperCAmelCase_ = sd_pipe( [prompt] , image=__a , generator=__a , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="np" , return_dict=__a , )[0] UpperCAmelCase_ = image[0, -3:, -3:, -1] UpperCAmelCase_ = image_from_tuple[0, -3:, -3:, -1] UpperCAmelCase_ = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) UpperCAmelCase_ = np.array([0.31_13, 0.39_10, 0.42_72, 0.48_59, 0.50_61, 0.46_52, 0.53_62, 0.57_15, 0.56_61] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase (self : Optional[int] ): UpperCAmelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ = self.dummy_cond_unet_upscale UpperCAmelCase_ = DDPMScheduler() UpperCAmelCase_ = DDIMScheduler(prediction_type="v_prediction" ) UpperCAmelCase_ = self.dummy_vae UpperCAmelCase_ = self.dummy_text_encoder UpperCAmelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCAmelCase_ = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCAmelCase_ = Image.fromarray(np.uinta(__a ) ).convert("RGB" ).resize((64, 64) ) # make sure here that pndm scheduler skips prk UpperCAmelCase_ = StableDiffusionUpscalePipeline( unet=__a , low_res_scheduler=__a , scheduler=__a , vae=__a , text_encoder=__a , tokenizer=__a , max_noise_level=350 , ) UpperCAmelCase_ = sd_pipe.to(__a ) sd_pipe.set_progress_bar_config(disable=__a ) UpperCAmelCase_ = "A painting of a squirrel eating a burger" UpperCAmelCase_ = sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="np" , ) UpperCAmelCase_ = output.images assert image.shape[0] == 2 UpperCAmelCase_ = torch.Generator(device=__a ).manual_seed(0 ) UpperCAmelCase_ = sd_pipe( [prompt] , image=__a , generator=__a , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="np" , ) UpperCAmelCase_ = output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _lowercase (self : str ): UpperCAmelCase_ = self.dummy_cond_unet_upscale UpperCAmelCase_ = DDPMScheduler() UpperCAmelCase_ = DDIMScheduler(prediction_type="v_prediction" ) UpperCAmelCase_ = self.dummy_vae UpperCAmelCase_ = self.dummy_text_encoder UpperCAmelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCAmelCase_ = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCAmelCase_ = Image.fromarray(np.uinta(__a ) ).convert("RGB" ).resize((64, 64) ) # put models in fp16, except vae as it overflows in fp16 UpperCAmelCase_ = unet.half() UpperCAmelCase_ = text_encoder.half() # make sure here that pndm scheduler skips prk UpperCAmelCase_ = StableDiffusionUpscalePipeline( unet=__a , low_res_scheduler=__a , scheduler=__a , vae=__a , text_encoder=__a , tokenizer=__a , max_noise_level=350 , ) UpperCAmelCase_ = sd_pipe.to(__a ) sd_pipe.set_progress_bar_config(disable=__a ) UpperCAmelCase_ = "A painting of a squirrel eating a burger" UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = sd_pipe( [prompt] , image=__a , generator=__a , num_inference_steps=2 , output_type="np" , ).images UpperCAmelCase_ = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class __A ( unittest.TestCase ): def _lowercase (self : List[str] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase (self : List[Any] ): UpperCAmelCase_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-upscale/low_res_cat.png" ) UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale" "/upsampled_cat.npy" ) UpperCAmelCase_ = "stabilityai/stable-diffusion-x4-upscaler" UpperCAmelCase_ = StableDiffusionUpscalePipeline.from_pretrained(__a ) pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) pipe.enable_attention_slicing() UpperCAmelCase_ = "a cat sitting on a park bench" UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = pipe( prompt=__a , image=__a , generator=__a , output_type="np" , ) UpperCAmelCase_ = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-3 def _lowercase (self : Tuple ): UpperCAmelCase_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-upscale/low_res_cat.png" ) UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale" "/upsampled_cat_fp16.npy" ) UpperCAmelCase_ = "stabilityai/stable-diffusion-x4-upscaler" UpperCAmelCase_ = StableDiffusionUpscalePipeline.from_pretrained( __a , torch_dtype=torch.floataa , ) pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) pipe.enable_attention_slicing() UpperCAmelCase_ = "a cat sitting on a park bench" UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = pipe( prompt=__a , image=__a , generator=__a , output_type="np" , ) UpperCAmelCase_ = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def _lowercase (self : List[Any] ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() UpperCAmelCase_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-upscale/low_res_cat.png" ) UpperCAmelCase_ = "stabilityai/stable-diffusion-x4-upscaler" UpperCAmelCase_ = StableDiffusionUpscalePipeline.from_pretrained( __a , torch_dtype=torch.floataa , ) pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() UpperCAmelCase_ = "a cat sitting on a park bench" UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = pipe( prompt=__a , image=__a , generator=__a , num_inference_steps=5 , output_type="np" , ) UpperCAmelCase_ = torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 10**9
1
1
"""simple docstring""" import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __a (lowercase__ , unittest.TestCase): '''simple docstring''' _SCREAMING_SNAKE_CASE :str = LxmertTokenizer _SCREAMING_SNAKE_CASE :int = LxmertTokenizerFast _SCREAMING_SNAKE_CASE :Union[str, Any] = True _SCREAMING_SNAKE_CASE :List[Any] = True def _a ( self ) -> List[str]: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ : Dict = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] SCREAMING_SNAKE_CASE__ : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def _a ( self , _a ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = 'UNwant\u00E9d,running' SCREAMING_SNAKE_CASE__ : List[Any] = 'unwanted, running' return input_text, output_text def _a ( self ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = self.tokenizer_class(self.vocab_file ) SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(_a , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [7, 4, 5, 10, 8, 9] ) def _a ( self ) -> Union[str, Any]: """simple docstring""" if not self.test_rust_tokenizer: return SCREAMING_SNAKE_CASE__ : str = self.get_tokenizer() SCREAMING_SNAKE_CASE__ : Any = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ : List[str] = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE__ : List[Any] = tokenizer.tokenize(_a ) SCREAMING_SNAKE_CASE__ : Optional[int] = rust_tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) SCREAMING_SNAKE_CASE__ : Optional[int] = tokenizer.encode(_a , add_special_tokens=_a ) SCREAMING_SNAKE_CASE__ : Optional[Any] = rust_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ : Dict = tokenizer.encode(_a ) SCREAMING_SNAKE_CASE__ : Optional[Any] = rust_tokenizer.encode(_a ) self.assertListEqual(_a , _a )
355
"""simple docstring""" import math from collections.abc import Iterator from itertools import takewhile def _lowercase ( __lowerCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(__lowerCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _lowercase ( ) -> Iterator[int]: SCREAMING_SNAKE_CASE__ : List[Any] = 2 while True: if is_prime(__lowerCAmelCase ): yield num num += 1 def _lowercase ( __lowerCAmelCase = 200_0000 ) -> int: return sum(takewhile(lambda __lowerCAmelCase : x < n , prime_generator() ) ) if __name__ == "__main__": print(f'{solution() = }')
56
0
def _a ( UpperCamelCase_ : list[int] , UpperCamelCase_ : list[int] ) -> None: """simple docstring""" lowerCAmelCase__ = len(UpperCamelCase_ ) print("The following activities are selected:" ) # The first activity is always selected lowerCAmelCase__ = 0 print(UpperCamelCase_ , end="," ) # Consider rest of the activities for j in range(UpperCamelCase_ ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(UpperCamelCase_ , end="," ) lowerCAmelCase__ = j if __name__ == "__main__": import doctest doctest.testmod() a_ = [1, 3, 0, 5, 8, 5] a_ = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
340
import collections import importlib.util import os import re from pathlib import Path a_ = '''src/transformers''' # Matches is_xxx_available() a_ = re.compile(r'''is\_([a-z_]*)_available()''') # Catches a one-line _import_struct = {xxx} a_ = re.compile(r'''^_import_structure\s+=\s+\{([^\}]+)\}''') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] a_ = re.compile(r'''\s+"\S*":\s+\[([^\]]*)\]''') # Catches a line if not is_foo_available a_ = re.compile(r'''^\s*if\s+not\s+is\_[a-z_]*\_available\(\)''') # Catches a line _import_struct["bla"].append("foo") a_ = re.compile(r'''^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)''') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] a_ = re.compile(r'''^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]''') # Catches a line with an object between quotes and a comma: "MyModel", a_ = re.compile('''^\s+"([^"]+)",''') # Catches a line with objects between brackets only: ["foo", "bar"], a_ = re.compile('''^\s+\[([^\]]+)\]''') # Catches a line with from foo import bar, bla, boo a_ = re.compile(r'''\s+from\s+\S*\s+import\s+([^\(\s].*)\n''') # Catches a line with try: a_ = re.compile(r'''^\s*try:''') # Catches a line with else: a_ = re.compile(r'''^\s*else:''') def _a ( UpperCamelCase_ : Union[str, Any] ) -> List[str]: """simple docstring""" if _re_test_backend.search(UpperCamelCase_ ) is None: return None lowerCAmelCase__ = [b[0] for b in _re_backend.findall(UpperCamelCase_ )] backends.sort() return "_and_".join(UpperCamelCase_ ) def _a ( UpperCamelCase_ : Optional[int] ) -> Tuple: """simple docstring""" with open(UpperCamelCase_ , "r" , encoding="utf-8" , newline="\n" ) as f: lowerCAmelCase__ = f.readlines() lowerCAmelCase__ = 0 while line_index < len(UpperCamelCase_ ) and not lines[line_index].startswith("_import_structure = {" ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(UpperCamelCase_ ): return None # First grab the objects without a specific backend in _import_structure lowerCAmelCase__ = [] while not lines[line_index].startswith("if TYPE_CHECKING" ) and find_backend(lines[line_index] ) is None: lowerCAmelCase__ = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(UpperCamelCase_ ): lowerCAmelCase__ = _re_one_line_import_struct.search(UpperCamelCase_ ).groups()[0] lowerCAmelCase__ = re.findall("\[([^\]]+)\]" , UpperCamelCase_ ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(", " )] ) line_index += 1 continue lowerCAmelCase__ = _re_import_struct_key_value.search(UpperCamelCase_ ) if single_line_import_search is not None: lowerCAmelCase__ = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(", " ) if len(UpperCamelCase_ ) > 0] objects.extend(UpperCamelCase_ ) elif line.startswith(" " * 8 + "\"" ): objects.append(line[9:-3] ) line_index += 1 lowerCAmelCase__ = {"none": objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith("if TYPE_CHECKING" ): # If the line is an if not is_backend_available, we grab all objects associated. lowerCAmelCase__ = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowerCAmelCase__ = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowerCAmelCase__ = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(" " * 4 ): lowerCAmelCase__ = lines[line_index] if _re_import_struct_add_one.search(UpperCamelCase_ ) is not None: objects.append(_re_import_struct_add_one.search(UpperCamelCase_ ).groups()[0] ) elif _re_import_struct_add_many.search(UpperCamelCase_ ) is not None: lowerCAmelCase__ = _re_import_struct_add_many.search(UpperCamelCase_ ).groups()[0].split(", " ) lowerCAmelCase__ = [obj[1:-1] for obj in imports if len(UpperCamelCase_ ) > 0] objects.extend(UpperCamelCase_ ) elif _re_between_brackets.search(UpperCamelCase_ ) is not None: lowerCAmelCase__ = _re_between_brackets.search(UpperCamelCase_ ).groups()[0].split(", " ) lowerCAmelCase__ = [obj[1:-1] for obj in imports if len(UpperCamelCase_ ) > 0] objects.extend(UpperCamelCase_ ) elif _re_quote_object.search(UpperCamelCase_ ) is not None: objects.append(_re_quote_object.search(UpperCamelCase_ ).groups()[0] ) elif line.startswith(" " * 8 + "\"" ): objects.append(line[9:-3] ) elif line.startswith(" " * 12 + "\"" ): objects.append(line[13:-3] ) line_index += 1 lowerCAmelCase__ = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend lowerCAmelCase__ = [] while ( line_index < len(UpperCamelCase_ ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith("else" ) ): lowerCAmelCase__ = lines[line_index] lowerCAmelCase__ = _re_import.search(UpperCamelCase_ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", " ) ) elif line.startswith(" " * 8 ): objects.append(line[8:-2] ) line_index += 1 lowerCAmelCase__ = {"none": objects} # Let's continue with backend-specific objects while line_index < len(UpperCamelCase_ ): # If the line is an if is_backend_available, we grab all objects associated. lowerCAmelCase__ = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowerCAmelCase__ = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowerCAmelCase__ = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(" " * 8 ): lowerCAmelCase__ = lines[line_index] lowerCAmelCase__ = _re_import.search(UpperCamelCase_ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", " ) ) elif line.startswith(" " * 12 ): objects.append(line[12:-2] ) line_index += 1 lowerCAmelCase__ = objects else: line_index += 1 return import_dict_objects, type_hint_objects def _a ( UpperCamelCase_ : int , UpperCamelCase_ : Optional[Any] ) -> str: """simple docstring""" def find_duplicates(UpperCamelCase_ : str ): return [k for k, v in collections.Counter(UpperCamelCase_ ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] lowerCAmelCase__ = [] for key in import_dict_objects.keys(): lowerCAmelCase__ = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F"Duplicate _import_structure definitions for: {duplicate_imports}" ) lowerCAmelCase__ = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F"Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}" ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): lowerCAmelCase__ = "base imports" if key == "none" else F"{key} backend" errors.append(F"Differences for {name}:" ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F" {a} in TYPE_HINT but not in _import_structure." ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F" {a} in _import_structure but not in TYPE_HINT." ) return errors def _a ( ) -> List[Any]: """simple docstring""" lowerCAmelCase__ = [] for root, _, files in os.walk(UpperCamelCase_ ): if "__init__.py" in files: lowerCAmelCase__ = os.path.join(UpperCamelCase_ , "__init__.py" ) lowerCAmelCase__ = parse_init(UpperCamelCase_ ) if objects is not None: lowerCAmelCase__ = analyze_results(*UpperCamelCase_ ) if len(UpperCamelCase_ ) > 0: lowerCAmelCase__ = F"Problem in {fname}, both halves do not define the same objects.\n{errors[0]}" failures.append("\n".join(UpperCamelCase_ ) ) if len(UpperCamelCase_ ) > 0: raise ValueError("\n\n".join(UpperCamelCase_ ) ) def _a ( ) -> str: """simple docstring""" lowerCAmelCase__ = [] for path, directories, files in os.walk(UpperCamelCase_ ): for folder in directories: # Ignore private modules if folder.startswith("_" ): directories.remove(UpperCamelCase_ ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(UpperCamelCase_ ) / folder).glob("*.py" ) ) ) == 0: continue lowerCAmelCase__ = str((Path(UpperCamelCase_ ) / folder).relative_to(UpperCamelCase_ ) ) lowerCAmelCase__ = short_path.replace(os.path.sep , "." ) submodules.append(UpperCamelCase_ ) for fname in files: if fname == "__init__.py": continue lowerCAmelCase__ = str((Path(UpperCamelCase_ ) / fname).relative_to(UpperCamelCase_ ) ) lowerCAmelCase__ = short_path.replace(".py" , "" ).replace(os.path.sep , "." ) if len(submodule.split("." ) ) == 1: submodules.append(UpperCamelCase_ ) return submodules a_ = [ '''convert_pytorch_checkpoint_to_tf2''', '''modeling_flax_pytorch_utils''', ] def _a ( ) -> int: """simple docstring""" lowerCAmelCase__ = importlib.util.spec_from_file_location( "transformers" , os.path.join(UpperCamelCase_ , "__init__.py" ) , submodule_search_locations=[PATH_TO_TRANSFORMERS] , ) lowerCAmelCase__ = spec.loader.load_module() lowerCAmelCase__ = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in transformers._import_structure.keys() ] if len(UpperCamelCase_ ) > 0: lowerCAmelCase__ = "\n".join(F"- {module}" for module in module_not_registered ) raise ValueError( "The following submodules are not properly registered in the main init of Transformers:\n" F"{list_of_modules}\n" "Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value." ) if __name__ == "__main__": check_all_inits() check_submodules()
340
1
from math import factorial def __lowerCamelCase ( lowerCamelCase__ : int , lowerCamelCase__ : int ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase__ ) // (factorial(lowerCamelCase__ ) * factorial(n - k )) if __name__ == "__main__": print( "The number of five-card hands possible from a standard", f"""fifty-two card deck is: {combinations(52, 5)}\n""", ) print( "If a class of 40 students must be arranged into groups of", f"""4 for group projects, there are {combinations(40, 4)} ways""", "to arrange them.\n", ) print( "If 10 teams are competing in a Formula One race, there", f"""are {combinations(10, 3)} ways that first, second and""", "third place can be awarded.", )
66
import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) UpperCAmelCase : Optional[int] = logging.getLogger(__name__) UpperCAmelCase : Dict = tf.data.AUTOTUNE def __lowerCamelCase ( ): '''simple docstring''' lowerCamelCase = argparse.ArgumentParser(description="""Train a masked language model on TPU.""" ) parser.add_argument( """--pretrained_model_config""" , type=lowerCamelCase__ , default="""roberta-base""" , help="""The model config to use. Note that we don't copy the model's weights, only the config!""" , ) parser.add_argument( """--tokenizer""" , type=lowerCamelCase__ , default="""unigram-tokenizer-wikitext""" , help="""The name of the tokenizer to load. We use the pretrained tokenizer to initialize the model's vocab size.""" , ) parser.add_argument( """--per_replica_batch_size""" , type=lowerCamelCase__ , default=8 , help="""Batch size per TPU core.""" , ) parser.add_argument( """--no_tpu""" , action="""store_true""" , help="""If set, run on CPU and don't try to initialize a TPU. Useful for debugging on non-TPU instances.""" , ) parser.add_argument( """--tpu_name""" , type=lowerCamelCase__ , help="""Name of TPU resource to initialize. Should be blank on Colab, and 'local' on TPU VMs.""" , default="""local""" , ) parser.add_argument( """--tpu_zone""" , type=lowerCamelCase__ , help="""Google cloud zone that TPU resource is located in. Only used for non-Colab TPU nodes.""" , ) parser.add_argument( """--gcp_project""" , type=lowerCamelCase__ , help="""Google cloud project name. Only used for non-Colab TPU nodes.""" ) parser.add_argument( """--bfloat16""" , action="""store_true""" , help="""Use mixed-precision bfloat16 for training. This is the recommended lower-precision format for TPU.""" , ) parser.add_argument( """--train_dataset""" , type=lowerCamelCase__ , help="""Path to training dataset to load. If the path begins with `gs://`""" """ then the dataset will be loaded from a Google Cloud Storage bucket.""" , ) parser.add_argument( """--shuffle_buffer_size""" , type=lowerCamelCase__ , default=2**18 , help="""Size of the shuffle buffer (in samples)""" , ) parser.add_argument( """--eval_dataset""" , type=lowerCamelCase__ , help="""Path to evaluation dataset to load. If the path begins with `gs://`""" """ then the dataset will be loaded from a Google Cloud Storage bucket.""" , ) parser.add_argument( """--num_epochs""" , type=lowerCamelCase__ , default=1 , help="""Number of epochs to train for.""" , ) parser.add_argument( """--learning_rate""" , type=lowerCamelCase__ , default=1E-4 , help="""Learning rate to use for training.""" , ) parser.add_argument( """--weight_decay_rate""" , type=lowerCamelCase__ , default=1E-3 , help="""Weight decay rate to use for training.""" , ) parser.add_argument( """--max_length""" , type=lowerCamelCase__ , default=512 , help="""Maximum length of tokenized sequences. Should match the setting used in prepare_tfrecord_shards.py""" , ) parser.add_argument( """--mlm_probability""" , type=lowerCamelCase__ , default=0.1_5 , help="""Fraction of tokens to mask during training.""" , ) parser.add_argument("""--output_dir""" , type=lowerCamelCase__ , required=lowerCamelCase__ , help="""Path to save model checkpoints to.""" ) parser.add_argument("""--hub_model_id""" , type=lowerCamelCase__ , help="""Model ID to upload to on the Hugging Face Hub.""" ) lowerCamelCase = parser.parse_args() return args def __lowerCamelCase ( lowerCamelCase__ : List[str] ): '''simple docstring''' try: if args.tpu_name: lowerCamelCase = tf.distribute.cluster_resolver.TPUClusterResolver( args.tpu_name , zone=args.tpu_zone , project=args.gcp_project ) else: lowerCamelCase = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: raise RuntimeError( """Couldn't connect to TPU! Most likely you need to specify --tpu_name, --tpu_zone, or """ """--gcp_project. When running on a TPU VM, use --tpu_name local.""" ) tf.config.experimental_connect_to_cluster(lowerCamelCase__ ) tf.tpu.experimental.initialize_tpu_system(lowerCamelCase__ ) return tpu def __lowerCamelCase ( lowerCamelCase__ : List[str] ): '''simple docstring''' lowerCamelCase = 0 for file in file_list: lowerCamelCase = file.split("""/""" )[-1] lowerCamelCase = re.search(R"""-\d+-(\d+)\.tfrecord""" , lowerCamelCase__ ).group(1 ) lowerCamelCase = int(lowerCamelCase__ ) num_samples += sample_count return num_samples def __lowerCamelCase ( lowerCamelCase__ : List[str] , lowerCamelCase__ : Any , lowerCamelCase__ : Dict , lowerCamelCase__ : Tuple , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : int=None ): '''simple docstring''' lowerCamelCase = count_samples(lowerCamelCase__ ) lowerCamelCase = tf.data.Dataset.from_tensor_slices(lowerCamelCase__ ) if shuffle: lowerCamelCase = dataset.shuffle(len(lowerCamelCase__ ) ) lowerCamelCase = tf.data.TFRecordDataset(lowerCamelCase__ , num_parallel_reads=lowerCamelCase__ ) # TF can't infer the total sample count because it doesn't read all the records yet, so we assert it here lowerCamelCase = dataset.apply(tf.data.experimental.assert_cardinality(lowerCamelCase__ ) ) lowerCamelCase = dataset.map(lowerCamelCase__ , num_parallel_calls=lowerCamelCase__ ) if shuffle: assert shuffle_buffer_size is not None lowerCamelCase = dataset.shuffle(args.shuffle_buffer_size ) lowerCamelCase = dataset.batch(lowerCamelCase__ , drop_remainder=lowerCamelCase__ ) lowerCamelCase = dataset.map(lowerCamelCase__ , num_parallel_calls=lowerCamelCase__ ) lowerCamelCase = dataset.prefetch(lowerCamelCase__ ) return dataset def __lowerCamelCase ( lowerCamelCase__ : Any ): '''simple docstring''' if not args.no_tpu: lowerCamelCase = initialize_tpu(lowerCamelCase__ ) lowerCamelCase = tf.distribute.TPUStrategy(lowerCamelCase__ ) else: lowerCamelCase = tf.distribute.OneDeviceStrategy(device="""/gpu:0""" ) if args.bfloataa: tf.keras.mixed_precision.set_global_policy("""mixed_bfloat16""" ) lowerCamelCase = AutoTokenizer.from_pretrained(args.tokenizer ) lowerCamelCase = AutoConfig.from_pretrained(args.pretrained_model_config ) lowerCamelCase = tokenizer.vocab_size lowerCamelCase = tf.io.gfile.glob(os.path.join(args.train_dataset , """*.tfrecord""" ) ) if not training_records: raise ValueError(f'No .tfrecord files found in {args.train_dataset}.' ) lowerCamelCase = tf.io.gfile.glob(os.path.join(args.eval_dataset , """*.tfrecord""" ) ) if not eval_records: raise ValueError(f'No .tfrecord files found in {args.eval_dataset}.' ) lowerCamelCase = count_samples(lowerCamelCase__ ) lowerCamelCase = num_train_samples // (args.per_replica_batch_size * strategy.num_replicas_in_sync) lowerCamelCase = steps_per_epoch * args.num_epochs with strategy.scope(): lowerCamelCase = TFAutoModelForMaskedLM.from_config(lowerCamelCase__ ) model(model.dummy_inputs ) # Pass some dummy inputs through the model to ensure all the weights are built lowerCamelCase , lowerCamelCase = create_optimizer( num_train_steps=lowerCamelCase__ , num_warmup_steps=total_train_steps // 20 , init_lr=args.learning_rate , weight_decay_rate=args.weight_decay_rate , ) # Transformers models compute the right loss for their task by default when labels are passed, and will # use this for training unless you specify your own loss function in compile(). model.compile(optimizer=lowerCamelCase__ , metrics=["""accuracy"""] ) def decode_fn(lowerCamelCase__ : Optional[Any] ): lowerCamelCase = { """input_ids""": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), """attention_mask""": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), } return tf.io.parse_single_example(lowerCamelCase__ , lowerCamelCase__ ) # Many of the data collators in Transformers are TF-compilable when return_tensors == "tf", so we can # use their methods in our data pipeline. lowerCamelCase = DataCollatorForLanguageModeling( tokenizer=lowerCamelCase__ , mlm_probability=args.mlm_probability , mlm=lowerCamelCase__ , return_tensors="""tf""" ) def mask_with_collator(lowerCamelCase__ : List[Any] ): # TF really needs an isin() function lowerCamelCase = ( ~tf.cast(batch["""attention_mask"""] , tf.bool ) | (batch["""input_ids"""] == tokenizer.cls_token_id) | (batch["""input_ids"""] == tokenizer.sep_token_id) ) lowerCamelCase , lowerCamelCase = data_collator.tf_mask_tokens( batch["""input_ids"""] , vocab_size=len(lowerCamelCase__ ) , mask_token_id=tokenizer.mask_token_id , special_tokens_mask=lowerCamelCase__ , ) return batch lowerCamelCase = args.per_replica_batch_size * strategy.num_replicas_in_sync lowerCamelCase = prepare_dataset( lowerCamelCase__ , decode_fn=lowerCamelCase__ , mask_fn=lowerCamelCase__ , batch_size=lowerCamelCase__ , shuffle=lowerCamelCase__ , shuffle_buffer_size=args.shuffle_buffer_size , ) lowerCamelCase = prepare_dataset( lowerCamelCase__ , decode_fn=lowerCamelCase__ , mask_fn=lowerCamelCase__ , batch_size=lowerCamelCase__ , shuffle=lowerCamelCase__ , ) lowerCamelCase = [] if args.hub_model_id: callbacks.append( PushToHubCallback(output_dir=args.output_dir , hub_model_id=args.hub_model_id , tokenizer=lowerCamelCase__ ) ) model.fit( lowerCamelCase__ , validation_data=lowerCamelCase__ , epochs=args.num_epochs , callbacks=lowerCamelCase__ , ) model.save_pretrained(args.output_dir ) if __name__ == "__main__": UpperCAmelCase : Tuple = parse_args() main(args)
66
1
'''simple docstring''' import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def UpperCamelCase_ ( A__ : Any , A__ : str , A__ : str , A__ : Path , A__ : str = None , A__ : str = None , A__ : str = None , ): '''simple docstring''' if config_name_or_path is None: lowerCAmelCase_ : Dict = """facebook/rag-token-base""" if model_type == """rag_token""" else """facebook/rag-sequence-base""" if generator_tokenizer_name_or_path is None: lowerCAmelCase_ : Optional[int] = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: lowerCAmelCase_ : Dict = question_encoder_name_or_path lowerCAmelCase_ : Tuple = RagTokenForGeneration if model_type == """rag_token""" else RagSequenceForGeneration # Save model. lowerCAmelCase_ : List[str] = RagConfig.from_pretrained(A__ ) lowerCAmelCase_ : Optional[Any] = AutoConfig.from_pretrained(A__ ) lowerCAmelCase_ : Tuple = AutoConfig.from_pretrained(A__ ) lowerCAmelCase_ : int = gen_config lowerCAmelCase_ : Dict = question_encoder_config lowerCAmelCase_ : str = model_class.from_pretrained_question_encoder_generator( A__ , A__ , config=A__ ) rag_model.save_pretrained(A__ ) # Sanity check. model_class.from_pretrained(A__ ) # Save tokenizers. lowerCAmelCase_ : str = AutoTokenizer.from_pretrained(A__ ) gen_tokenizer.save_pretrained(dest_dir / """generator_tokenizer/""" ) lowerCAmelCase_ : List[str] = AutoTokenizer.from_pretrained(A__ ) question_encoder_tokenizer.save_pretrained(dest_dir / """question_encoder_tokenizer/""" ) if __name__ == "__main__": __A : str = argparse.ArgumentParser() parser.add_argument( "--model_type", choices=["rag_sequence", "rag_token"], required=True, type=str, help="RAG model type: rag_sequence, rag_token", ) parser.add_argument("--dest", type=str, required=True, help="Path to the output checkpoint directory.") parser.add_argument("--generator_name_or_path", type=str, required=True, help="Generator model identifier") parser.add_argument( "--question_encoder_name_or_path", type=str, required=True, help="Question encoder model identifier" ) parser.add_argument( "--generator_tokenizer_name_or_path", type=str, help="Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``", ) parser.add_argument( "--question_encoder_tokenizer_name_or_path", type=str, help="Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``", ) parser.add_argument( "--config_name_or_path", type=str, help=( "Identifier of the model config to use, if not provided, resolves to a base config for a given" " ``model_type``" ), ) __A : List[str] = parser.parse_args() __A : Union[str, Any] = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
120
'''simple docstring''' import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class __snake_case ( _SCREAMING_SNAKE_CASE): """simple docstring""" lowercase = ['image_processor', 'tokenizer'] lowercase = 'AutoImageProcessor' lowercase = 'AutoTokenizer' def __init__( self : int , lowerCamelCase : List[str]=None , lowerCamelCase : Union[str, Any]=None , **lowerCamelCase : str ) -> Tuple: lowerCAmelCase_ : str = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , lowerCamelCase , ) lowerCAmelCase_ : Tuple = kwargs.pop("""feature_extractor""" ) lowerCAmelCase_ : Tuple = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(lowerCamelCase , lowerCamelCase ) lowerCAmelCase_ : Optional[int] = self.image_processor lowerCAmelCase_ : Any = False def __call__( self : List[Any] , *lowerCamelCase : str , **lowerCamelCase : Tuple ) -> Union[str, Any]: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*lowerCamelCase , **lowerCamelCase ) lowerCAmelCase_ : Any = kwargs.pop("""images""" , lowerCamelCase ) lowerCAmelCase_ : Dict = kwargs.pop("""text""" , lowerCamelCase ) if len(lowerCamelCase ) > 0: lowerCAmelCase_ : str = args[0] lowerCAmelCase_ : Dict = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: lowerCAmelCase_ : Any = self.image_processor(lowerCamelCase , *lowerCamelCase , **lowerCamelCase ) if text is not None: lowerCAmelCase_ : str = self.tokenizer(lowerCamelCase , **lowerCamelCase ) if text is None: return inputs elif images is None: return encodings else: lowerCAmelCase_ : Dict = encodings["""input_ids"""] return inputs def __lowercase ( self : str , *lowerCamelCase : Union[str, Any] , **lowerCamelCase : Union[str, Any] ) -> List[str]: return self.tokenizer.batch_decode(*lowerCamelCase , **lowerCamelCase ) def __lowercase ( self : List[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> List[str]: return self.tokenizer.decode(*lowerCamelCase , **lowerCamelCase ) @contextmanager def __lowercase ( self : str ) -> Union[str, Any]: warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) lowerCAmelCase_ : List[str] = True lowerCAmelCase_ : Optional[Any] = self.tokenizer yield lowerCAmelCase_ : List[Any] = self.image_processor lowerCAmelCase_ : List[str] = False def __lowercase ( self : Dict , lowerCamelCase : List[str] , lowerCamelCase : str=False , lowerCamelCase : List[Any]=None ) -> Optional[int]: if added_vocab is None: lowerCAmelCase_ : str = self.tokenizer.get_added_vocab() lowerCAmelCase_ : Union[str, Any] = {} while tokens: lowerCAmelCase_ : Dict = re.search(R"""<s_(.*?)>""" , lowerCamelCase , re.IGNORECASE ) if start_token is None: break lowerCAmelCase_ : Tuple = start_token.group(1 ) lowerCAmelCase_ : Tuple = re.search(RF'</s_{key}>' , lowerCamelCase , re.IGNORECASE ) lowerCAmelCase_ : Tuple = start_token.group() if end_token is None: lowerCAmelCase_ : str = tokens.replace(lowerCamelCase , """""" ) else: lowerCAmelCase_ : List[str] = end_token.group() lowerCAmelCase_ : Dict = re.escape(lowerCamelCase ) lowerCAmelCase_ : int = re.escape(lowerCamelCase ) lowerCAmelCase_ : Dict = re.search(F'{start_token_escaped}(.*?){end_token_escaped}' , lowerCamelCase , re.IGNORECASE ) if content is not None: lowerCAmelCase_ : str = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node lowerCAmelCase_ : str = self.tokenajson(lowerCamelCase , is_inner_value=lowerCamelCase , added_vocab=lowerCamelCase ) if value: if len(lowerCamelCase ) == 1: lowerCAmelCase_ : List[Any] = value[0] lowerCAmelCase_ : Optional[Any] = value else: # leaf nodes lowerCAmelCase_ : List[str] = [] for leaf in content.split(R"""<sep/>""" ): lowerCAmelCase_ : Union[str, Any] = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": lowerCAmelCase_ : Any = leaf[1:-2] # for categorical special tokens output[key].append(lowerCamelCase ) if len(output[key] ) == 1: lowerCAmelCase_ : Optional[Any] = output[key][0] lowerCAmelCase_ : List[Any] = tokens[tokens.find(lowerCamelCase ) + len(lowerCamelCase ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=lowerCamelCase , added_vocab=lowerCamelCase ) if len(lowerCamelCase ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def __lowercase ( self : Dict ) -> int: warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , lowerCamelCase , ) return self.image_processor_class @property def __lowercase ( self : Dict ) -> Optional[Any]: warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , lowerCamelCase , ) return self.image_processor
120
1
def lowerCamelCase__ ( UpperCamelCase__ : int ) -> bool: '''simple docstring''' return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print("""Program to check whether a number is a Perfect number or not...""") UpperCAmelCase_ = int(input("""Enter number: """).strip()) print(F"{number} is {'' if perfect(number) else 'not '}a Perfect Number.")
295
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging UpperCAmelCase_ = logging.get_logger(__name__) logging.set_verbosity_info() def lowerCamelCase__ ( UpperCamelCase__ : str , UpperCamelCase__ : str ) -> Union[str, Any]: '''simple docstring''' if "xprophetnet" in prophetnet_checkpoint_path: _snake_case = XLMProphetNetForConditionalGenerationOld.from_pretrained(UpperCamelCase__ ) _snake_case , _snake_case = XLMProphetNetForConditionalGeneration.from_pretrained( UpperCamelCase__ , output_loading_info=UpperCamelCase__ ) else: _snake_case = ProphetNetForConditionalGenerationOld.from_pretrained(UpperCamelCase__ ) _snake_case , _snake_case = ProphetNetForConditionalGeneration.from_pretrained( UpperCamelCase__ , output_loading_info=UpperCamelCase__ ) _snake_case = ['key_proj', 'value_proj', 'query_proj'] _snake_case = { 'self_attn': 'ngram_self_attn', 'cross_attn': 'encoder_attn', 'cross_attn_layer_norm': 'encoder_attn_layer_norm', 'feed_forward_layer_norm': 'final_layer_norm', 'feed_forward': '', 'intermediate': 'fc1', 'output': 'fc2', 'key_proj': 'k_proj', 'query_proj': 'q_proj', 'value_proj': 'v_proj', 'word_embeddings': 'embed_tokens', 'embeddings_layer_norm': 'emb_layer_norm', 'relative_pos_embeddings': 'relative_linear', 'ngram_embeddings': 'ngram_input_embed', 'position_embeddings': 'embed_positions', } for key in loading_info["missing_keys"]: _snake_case = key.split('.' ) if attributes[0] == "lm_head": _snake_case = prophet _snake_case = prophet_old else: _snake_case = prophet.prophetnet _snake_case = prophet_old.model _snake_case = False for attribute in attributes: if attribute in mapping: _snake_case = mapping[attribute] if not hasattr(UpperCamelCase__ , UpperCamelCase__ ) and len(UpperCamelCase__ ) > 0: _snake_case = attribute elif hasattr(UpperCamelCase__ , UpperCamelCase__ ): _snake_case = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" _snake_case = old_model.weight logger.info(F'''{attribute} is initialized.''' ) _snake_case = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" _snake_case = old_model.bias logger.info(F'''{attribute} is initialized''' ) _snake_case = True break elif attribute in special_keys and hasattr(UpperCamelCase__ , 'in_proj_weight' ): _snake_case = old_model.in_proj_weight.shape[0] // 3 _snake_case = getattr(UpperCamelCase__ , UpperCamelCase__ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": _snake_case = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) _snake_case = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": _snake_case = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) _snake_case = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": _snake_case = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) _snake_case = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) _snake_case = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." _snake_case = nn.Parameter(old_model.embed_positions.weight[:512, :] ) _snake_case = True break if attribute.isdigit(): _snake_case = model[int(UpperCamelCase__ )] _snake_case = old_model[int(UpperCamelCase__ )] else: _snake_case = getattr(UpperCamelCase__ , UpperCamelCase__ ) if old_attribute == "": _snake_case = old_model else: if not hasattr(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError(F'''{old_model} does not have {old_attribute}''' ) _snake_case = getattr(UpperCamelCase__ , UpperCamelCase__ ) if not is_key_init: raise ValueError(F'''{key} was not correctly initialized!''' ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) prophet.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--prophetnet_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) UpperCAmelCase_ = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
295
1
'''simple docstring''' from ....configuration_utils import PretrainedConfig from ....utils import logging _A : Dict = logging.get_logger(__name__) # TODO: upload to AWS _A : Tuple = { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json''' ), } class _lowercase ( UpperCAmelCase__ ): '''simple docstring''' _SCREAMING_SNAKE_CASE : List[Any] = """retribert""" def __init__( self : str , SCREAMING_SNAKE_CASE__ : str=3_05_22 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=7_68 , SCREAMING_SNAKE_CASE__ : Dict=8 , SCREAMING_SNAKE_CASE__ : Optional[Any]=12 , SCREAMING_SNAKE_CASE__ : int=30_72 , SCREAMING_SNAKE_CASE__ : int="gelu" , SCREAMING_SNAKE_CASE__ : Tuple=0.1 , SCREAMING_SNAKE_CASE__ : List[str]=0.1 , SCREAMING_SNAKE_CASE__ : List[str]=5_12 , SCREAMING_SNAKE_CASE__ : int=2 , SCREAMING_SNAKE_CASE__ : Dict=0.0_2 , SCREAMING_SNAKE_CASE__ : Optional[int]=1e-1_2 , SCREAMING_SNAKE_CASE__ : Any=True , SCREAMING_SNAKE_CASE__ : Dict=1_28 , SCREAMING_SNAKE_CASE__ : List[str]=0 , **SCREAMING_SNAKE_CASE__ : List[str] , ) -> Optional[int]: super().__init__(pad_token_id=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = hidden_act __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = share_encoders __lowerCAmelCase = projection_dim
229
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _lowercase ( unittest.TestCase ): '''simple docstring''' @property def a ( self : Any ) -> int: torch.manual_seed(0 ) __lowerCAmelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) return model def a ( self : Dict ) -> int: __lowerCAmelCase = self.dummy_uncond_unet __lowerCAmelCase = PNDMScheduler() __lowerCAmelCase = PNDMPipeline(unet=SCREAMING_SNAKE_CASE__ , scheduler=SCREAMING_SNAKE_CASE__ ) pndm.to(SCREAMING_SNAKE_CASE__ ) pndm.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = pndm(generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=20 , output_type="""numpy""" ).images __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = pndm(generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=20 , output_type="""numpy""" , return_dict=SCREAMING_SNAKE_CASE__ )[0] __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class _lowercase ( unittest.TestCase ): '''simple docstring''' def a ( self : Any ) -> Any: __lowerCAmelCase = """google/ddpm-cifar10-32""" __lowerCAmelCase = UNetaDModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = PNDMScheduler() __lowerCAmelCase = PNDMPipeline(unet=SCREAMING_SNAKE_CASE__ , scheduler=SCREAMING_SNAKE_CASE__ ) pndm.to(SCREAMING_SNAKE_CASE__ ) pndm.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = torch.manual_seed(0 ) __lowerCAmelCase = pndm(generator=SCREAMING_SNAKE_CASE__ , output_type="""numpy""" ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCAmelCase = np.array([0.1_5_6_4, 0.1_4_6_4_5, 0.1_4_0_6, 0.1_4_7_1_5, 0.1_2_4_2_5, 0.1_4_0_4_5, 0.1_3_1_1_5, 0.1_2_1_7_5, 0.1_2_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
229
1
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig _snake_case : int = logging.get_logger(__name__) # General docstring _snake_case : int = 'MobileNetV1Config' # Base docstring _snake_case : Optional[int] = 'google/mobilenet_v1_1.0_224' _snake_case : List[str] = [1, 1024, 7, 7] # Image classification docstring _snake_case : Optional[Any] = 'google/mobilenet_v1_1.0_224' _snake_case : List[str] = 'tabby, tabby cat' _snake_case : List[Any] = [ 'google/mobilenet_v1_1.0_224', 'google/mobilenet_v1_0.75_192', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def a_ ( lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Tuple, lowerCAmelCase_ : Optional[Any]=None ): __lowerCAmelCase = {} if isinstance(lowerCAmelCase_, lowerCAmelCase_ ): __lowerCAmelCase = model.mobilenet_va else: __lowerCAmelCase = model __lowerCAmelCase = 'MobilenetV1/Conv2d_0/' __lowerCAmelCase = backbone.conv_stem.convolution.weight __lowerCAmelCase = backbone.conv_stem.normalization.bias __lowerCAmelCase = backbone.conv_stem.normalization.weight __lowerCAmelCase = backbone.conv_stem.normalization.running_mean __lowerCAmelCase = backbone.conv_stem.normalization.running_var for i in range(13 ): __lowerCAmelCase = i + 1 __lowerCAmelCase = i * 2 __lowerCAmelCase = backbone.layer[pt_index] __lowerCAmelCase = F"""MobilenetV1/Conv2d_{tf_index}_depthwise/""" __lowerCAmelCase = pointer.convolution.weight __lowerCAmelCase = pointer.normalization.bias __lowerCAmelCase = pointer.normalization.weight __lowerCAmelCase = pointer.normalization.running_mean __lowerCAmelCase = pointer.normalization.running_var __lowerCAmelCase = backbone.layer[pt_index + 1] __lowerCAmelCase = F"""MobilenetV1/Conv2d_{tf_index}_pointwise/""" __lowerCAmelCase = pointer.convolution.weight __lowerCAmelCase = pointer.normalization.bias __lowerCAmelCase = pointer.normalization.weight __lowerCAmelCase = pointer.normalization.running_mean __lowerCAmelCase = pointer.normalization.running_var if isinstance(lowerCAmelCase_, lowerCAmelCase_ ): __lowerCAmelCase = 'MobilenetV1/Logits/Conv2d_1c_1x1/' __lowerCAmelCase = model.classifier.weight __lowerCAmelCase = model.classifier.bias return tf_to_pt_map def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : Tuple, lowerCAmelCase_ : str ): try: import numpy as np import tensorflow as tf except ImportError: logger.error( 'Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ' 'https://www.tensorflow.org/install/ for installation instructions.' ) raise # Load weights from TF model __lowerCAmelCase = tf.train.list_variables(lowerCAmelCase_ ) __lowerCAmelCase = {} for name, shape in init_vars: logger.info(F"""Loading TF weight {name} with shape {shape}""" ) __lowerCAmelCase = tf.train.load_variable(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = array # Build TF to PyTorch weights loading map __lowerCAmelCase = _build_tf_to_pytorch_map(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) for name, pointer in tf_to_pt_map.items(): logger.info(F"""Importing {name}""" ) if name not in tf_weights: logger.info(F"""{name} not in tf pre-trained weights, skipping""" ) continue __lowerCAmelCase = tf_weights[name] if "depthwise_weights" in name: logger.info('Transposing depthwise' ) __lowerCAmelCase = np.transpose(lowerCAmelCase_, (2, 3, 0, 1) ) elif "weights" in name: logger.info('Transposing' ) if len(pointer.shape ) == 2: # copying into linear layer __lowerCAmelCase = array.squeeze().transpose() else: __lowerCAmelCase = np.transpose(lowerCAmelCase_, (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F"""Pointer shape {pointer.shape} and array shape {array.shape} mismatched""" ) logger.info(F"""Initialize PyTorch weight {name} {array.shape}""" ) __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ) tf_weights.pop(lowerCAmelCase_, lowerCAmelCase_ ) tf_weights.pop(name + '/RMSProp', lowerCAmelCase_ ) tf_weights.pop(name + '/RMSProp_1', lowerCAmelCase_ ) tf_weights.pop(name + '/ExponentialMovingAverage', lowerCAmelCase_ ) logger.info(F"""Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}""" ) return model def a_ ( lowerCAmelCase_ : torch.Tensor, lowerCAmelCase_ : nn.Convad ): __lowerCAmelCase , __lowerCAmelCase = features.shape[-2:] __lowerCAmelCase , __lowerCAmelCase = conv_layer.stride __lowerCAmelCase , __lowerCAmelCase = conv_layer.kernel_size if in_height % stride_height == 0: __lowerCAmelCase = max(kernel_height - stride_height, 0 ) else: __lowerCAmelCase = max(kernel_height - (in_height % stride_height), 0 ) if in_width % stride_width == 0: __lowerCAmelCase = max(kernel_width - stride_width, 0 ) else: __lowerCAmelCase = max(kernel_width - (in_width % stride_width), 0 ) __lowerCAmelCase = pad_along_width // 2 __lowerCAmelCase = pad_along_width - pad_left __lowerCAmelCase = pad_along_height // 2 __lowerCAmelCase = pad_along_height - pad_top __lowerCAmelCase = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(lowerCAmelCase_, lowerCAmelCase_, 'constant', 0.0 ) class _UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCAmelCase_ : MobileNetVaConfig , lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : Optional[int] = 1 , lowerCAmelCase_ : Optional[int] = 1 , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : Optional[bool] = True , lowerCAmelCase_ : Optional[bool or str] = True , ) -> None: super().__init__() __lowerCAmelCase = config if in_channels % groups != 0: raise ValueError(f"""Input channels ({in_channels}) are not divisible by {groups} groups.""" ) if out_channels % groups != 0: raise ValueError(f"""Output channels ({out_channels}) are not divisible by {groups} groups.""" ) __lowerCAmelCase = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) __lowerCAmelCase = nn.Convad( in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , kernel_size=lowerCAmelCase_ , stride=lowerCAmelCase_ , padding=lowerCAmelCase_ , groups=lowerCAmelCase_ , bias=lowerCAmelCase_ , padding_mode='zeros' , ) if use_normalization: __lowerCAmelCase = nn.BatchNormad( num_features=lowerCAmelCase_ , eps=config.layer_norm_eps , momentum=0.99_97 , affine=lowerCAmelCase_ , track_running_stats=lowerCAmelCase_ , ) else: __lowerCAmelCase = None if use_activation: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): __lowerCAmelCase = ACTaFN[use_activation] elif isinstance(config.hidden_act , lowerCAmelCase_ ): __lowerCAmelCase = ACTaFN[config.hidden_act] else: __lowerCAmelCase = config.hidden_act else: __lowerCAmelCase = None def lowercase ( self : Dict , lowerCAmelCase_ : torch.Tensor ) -> torch.Tensor: if self.config.tf_padding: __lowerCAmelCase = apply_tf_padding(lowerCAmelCase_ , self.convolution ) __lowerCAmelCase = self.convolution(lowerCAmelCase_ ) if self.normalization is not None: __lowerCAmelCase = self.normalization(lowerCAmelCase_ ) if self.activation is not None: __lowerCAmelCase = self.activation(lowerCAmelCase_ ) return features class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = MobileNetVaConfig a_ = load_tf_weights_in_mobilenet_va a_ = """mobilenet_v1""" a_ = """pixel_values""" a_ = False def lowercase ( self : Optional[int] , lowerCAmelCase_ : Union[nn.Linear, nn.Convad] ) -> None: if isinstance(lowerCAmelCase_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(lowerCAmelCase_ , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) _snake_case : List[Any] = R'\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' _snake_case : Optional[Any] = R'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`MobileNetV1ImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( """The bare MobileNetV1 model outputting raw hidden-states without any specific head on top.""" , _UpperCamelCase , ) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCAmelCase_ : MobileNetVaConfig , lowerCAmelCase_ : bool = True ) -> Optional[Any]: super().__init__(lowerCAmelCase_ ) __lowerCAmelCase = config __lowerCAmelCase = 3_2 __lowerCAmelCase = max(int(depth * config.depth_multiplier ) , config.min_depth ) __lowerCAmelCase = MobileNetVaConvLayer( lowerCAmelCase_ , in_channels=config.num_channels , out_channels=lowerCAmelCase_ , kernel_size=3 , stride=2 , ) __lowerCAmelCase = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] __lowerCAmelCase = nn.ModuleList() for i in range(1_3 ): __lowerCAmelCase = out_channels if strides[i] == 2 or i == 0: depth *= 2 __lowerCAmelCase = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( lowerCAmelCase_ , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , kernel_size=3 , stride=strides[i] , groups=lowerCAmelCase_ , ) ) self.layer.append( MobileNetVaConvLayer( lowerCAmelCase_ , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , kernel_size=1 , ) ) __lowerCAmelCase = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def lowercase ( self : List[str] , lowerCAmelCase_ : Optional[Any] ) -> Dict: raise NotImplementedError @add_start_docstrings_to_model_forward(lowerCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCAmelCase_ , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowercase ( self : Any , lowerCAmelCase_ : Optional[torch.Tensor] = None , lowerCAmelCase_ : Optional[bool] = None , lowerCAmelCase_ : Optional[bool] = None , ) -> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: __lowerCAmelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __lowerCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('You have to specify pixel_values' ) __lowerCAmelCase = self.conv_stem(lowerCAmelCase_ ) __lowerCAmelCase = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): __lowerCAmelCase = layer_module(lowerCAmelCase_ ) if output_hidden_states: __lowerCAmelCase = all_hidden_states + (hidden_states,) __lowerCAmelCase = hidden_states if self.pooler is not None: __lowerCAmelCase = torch.flatten(self.pooler(lowerCAmelCase_ ) , start_dim=1 ) else: __lowerCAmelCase = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCAmelCase_ , pooler_output=lowerCAmelCase_ , hidden_states=lowerCAmelCase_ , ) @add_start_docstrings( """ MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ , _UpperCamelCase , ) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCAmelCase_ : MobileNetVaConfig ) -> None: super().__init__(lowerCAmelCase_ ) __lowerCAmelCase = config.num_labels __lowerCAmelCase = MobileNetVaModel(lowerCAmelCase_ ) __lowerCAmelCase = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head __lowerCAmelCase = nn.Dropout(config.classifier_dropout_prob , inplace=lowerCAmelCase_ ) __lowerCAmelCase = nn.Linear(lowerCAmelCase_ , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowerCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCAmelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowercase ( self : str , lowerCAmelCase_ : Optional[torch.Tensor] = None , lowerCAmelCase_ : Optional[bool] = None , lowerCAmelCase_ : Optional[torch.Tensor] = None , lowerCAmelCase_ : Optional[bool] = None , ) -> Union[tuple, ImageClassifierOutputWithNoAttention]: __lowerCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict __lowerCAmelCase = self.mobilenet_va(lowerCAmelCase_ , output_hidden_states=lowerCAmelCase_ , return_dict=lowerCAmelCase_ ) __lowerCAmelCase = outputs.pooler_output if return_dict else outputs[1] __lowerCAmelCase = self.classifier(self.dropout(lowerCAmelCase_ ) ) __lowerCAmelCase = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: __lowerCAmelCase = 'regression' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): __lowerCAmelCase = 'single_label_classification' else: __lowerCAmelCase = 'multi_label_classification' if self.config.problem_type == "regression": __lowerCAmelCase = MSELoss() if self.num_labels == 1: __lowerCAmelCase = loss_fct(logits.squeeze() , labels.squeeze() ) else: __lowerCAmelCase = loss_fct(lowerCAmelCase_ , lowerCAmelCase_ ) elif self.config.problem_type == "single_label_classification": __lowerCAmelCase = CrossEntropyLoss() __lowerCAmelCase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": __lowerCAmelCase = BCEWithLogitsLoss() __lowerCAmelCase = loss_fct(lowerCAmelCase_ , lowerCAmelCase_ ) if not return_dict: __lowerCAmelCase = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=lowerCAmelCase_ , logits=lowerCAmelCase_ , hidden_states=outputs.hidden_states , )
207
from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run _snake_case : int = True except (ImportError, AttributeError): _snake_case : int = object def a_ ( *lowerCAmelCase_ : List[str], **lowerCAmelCase_ : Optional[Any] ): pass _snake_case : Union[str, Any] = False _snake_case : int = logging.get_logger('transformers-cli/serving') def a_ ( lowerCAmelCase_ : Namespace ): __lowerCAmelCase = pipeline( task=args.task, model=args.model if args.model else None, config=args.config, tokenizer=args.tokenizer, device=args.device, ) return ServeCommand(lowerCAmelCase_, args.host, args.port, args.workers ) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = 42 class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = 42 a_ = 42 class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = 42 class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = 42 class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" @staticmethod def lowercase ( lowerCAmelCase_ : ArgumentParser ) -> Union[str, Any]: __lowerCAmelCase = parser.add_parser( 'serve' , help='CLI tool to run inference requests through REST and GraphQL endpoints.' ) serve_parser.add_argument( '--task' , type=lowerCAmelCase_ , choices=get_supported_tasks() , help='The task to run the pipeline on' , ) serve_parser.add_argument('--host' , type=lowerCAmelCase_ , default='localhost' , help='Interface the server will listen on.' ) serve_parser.add_argument('--port' , type=lowerCAmelCase_ , default=8_8_8_8 , help='Port the serving will listen to.' ) serve_parser.add_argument('--workers' , type=lowerCAmelCase_ , default=1 , help='Number of http workers' ) serve_parser.add_argument('--model' , type=lowerCAmelCase_ , help='Model\'s name or path to stored model.' ) serve_parser.add_argument('--config' , type=lowerCAmelCase_ , help='Model\'s config name or path to stored model.' ) serve_parser.add_argument('--tokenizer' , type=lowerCAmelCase_ , help='Tokenizer name to use.' ) serve_parser.add_argument( '--device' , type=lowerCAmelCase_ , default=-1 , help='Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)' , ) serve_parser.set_defaults(func=lowerCAmelCase_ ) def __init__( self : List[str] , lowerCAmelCase_ : Pipeline , lowerCAmelCase_ : str , lowerCAmelCase_ : int , lowerCAmelCase_ : int ) -> str: __lowerCAmelCase = pipeline __lowerCAmelCase = host __lowerCAmelCase = port __lowerCAmelCase = workers if not _serve_dependencies_installed: raise RuntimeError( 'Using serve command requires FastAPI and uvicorn. ' 'Please install transformers with [serving]: pip install "transformers[serving]".' 'Or install FastAPI and uvicorn separately.' ) else: logger.info(f"""Serving model over {host}:{port}""" ) __lowerCAmelCase = FastAPI( routes=[ APIRoute( '/' , self.model_info , response_model=lowerCAmelCase_ , response_class=lowerCAmelCase_ , methods=['GET'] , ), APIRoute( '/tokenize' , self.tokenize , response_model=lowerCAmelCase_ , response_class=lowerCAmelCase_ , methods=['POST'] , ), APIRoute( '/detokenize' , self.detokenize , response_model=lowerCAmelCase_ , response_class=lowerCAmelCase_ , methods=['POST'] , ), APIRoute( '/forward' , self.forward , response_model=lowerCAmelCase_ , response_class=lowerCAmelCase_ , methods=['POST'] , ), ] , timeout=6_0_0 , ) def lowercase ( self : Tuple ) -> str: run(self._app , host=self.host , port=self.port , workers=self.workers ) def lowercase ( self : Any ) -> List[str]: return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def lowercase ( self : int , lowerCAmelCase_ : str = Body(lowerCAmelCase_ , embed=lowerCAmelCase_ ) , lowerCAmelCase_ : bool = Body(lowerCAmelCase_ , embed=lowerCAmelCase_ ) ) -> Dict: try: __lowerCAmelCase = self._pipeline.tokenizer.tokenize(lowerCAmelCase_ ) if return_ids: __lowerCAmelCase = self._pipeline.tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) return ServeTokenizeResult(tokens=lowerCAmelCase_ , tokens_ids=lowerCAmelCase_ ) else: return ServeTokenizeResult(tokens=lowerCAmelCase_ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'model': '', 'error': str(lowerCAmelCase_ )} ) def lowercase ( self : int , lowerCAmelCase_ : List[int] = Body(lowerCAmelCase_ , embed=lowerCAmelCase_ ) , lowerCAmelCase_ : bool = Body(lowerCAmelCase_ , embed=lowerCAmelCase_ ) , lowerCAmelCase_ : bool = Body(lowerCAmelCase_ , embed=lowerCAmelCase_ ) , ) -> Union[str, Any]: try: __lowerCAmelCase = self._pipeline.tokenizer.decode(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) return ServeDeTokenizeResult(model='' , text=lowerCAmelCase_ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'model': '', 'error': str(lowerCAmelCase_ )} ) async def lowercase ( self : List[Any] , lowerCAmelCase_ : Union[str, Any]=Body(lowerCAmelCase_ , embed=lowerCAmelCase_ ) ) -> int: # Check we don't have empty string if len(lowerCAmelCase_ ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model __lowerCAmelCase = self._pipeline(lowerCAmelCase_ ) return ServeForwardResult(output=lowerCAmelCase_ ) except Exception as e: raise HTTPException(5_0_0 , {'error': str(lowerCAmelCase_ )} )
207
1
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) __UpperCamelCase : str = 'hf-internal-testing/tiny-random-bert' __UpperCamelCase : int = os.path.join(TRANSFORMERS_CACHE, 'models--hf-internal-testing--tiny-random-bert') __UpperCamelCase : Optional[Any] = '9b8c223d42b2188cb49d29af482996f9d0f3e5a6' class lowercase__ ( unittest.TestCase): def __A ( self : Optional[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = cached_file(UpperCamelCase__ , UpperCamelCase__ ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(UpperCamelCase__ ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) ) with open(os.path.join(UpperCamelCase__ , '''refs''' , '''main''' ) ) as f: SCREAMING_SNAKE_CASE : Optional[Any] = f.read() self.assertEqual(UpperCamelCase__ , os.path.join(UpperCamelCase__ , '''snapshots''' , UpperCamelCase__ , UpperCamelCase__ ) ) self.assertTrue(os.path.isfile(UpperCamelCase__ ) ) # File is cached at the same place the second time. SCREAMING_SNAKE_CASE : int = cached_file(UpperCamelCase__ , UpperCamelCase__ ) self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) # Using a specific revision to test the full commit hash. SCREAMING_SNAKE_CASE : Optional[Any] = cached_file(UpperCamelCase__ , UpperCamelCase__ , revision='''9b8c223''' ) self.assertEqual(UpperCamelCase__ , os.path.join(UpperCamelCase__ , '''snapshots''' , UpperCamelCase__ , UpperCamelCase__ ) ) def __A ( self : Any ): '''simple docstring''' with self.assertRaisesRegex(UpperCamelCase__ , '''is not a valid model identifier''' ): SCREAMING_SNAKE_CASE : Dict = cached_file('''tiny-random-bert''' , UpperCamelCase__ ) with self.assertRaisesRegex(UpperCamelCase__ , '''is not a valid git identifier''' ): SCREAMING_SNAKE_CASE : Union[str, Any] = cached_file(UpperCamelCase__ , UpperCamelCase__ , revision='''aaaa''' ) with self.assertRaisesRegex(UpperCamelCase__ , '''does not appear to have a file named''' ): SCREAMING_SNAKE_CASE : Union[str, Any] = cached_file(UpperCamelCase__ , '''conf''' ) def __A ( self : Dict ): '''simple docstring''' with self.assertRaisesRegex(UpperCamelCase__ , '''does not appear to have a file named''' ): SCREAMING_SNAKE_CASE : int = cached_file(UpperCamelCase__ , '''conf''' ) with open(os.path.join(UpperCamelCase__ , '''refs''' , '''main''' ) ) as f: SCREAMING_SNAKE_CASE : str = f.read() self.assertTrue(os.path.isfile(os.path.join(UpperCamelCase__ , '''.no_exist''' , UpperCamelCase__ , '''conf''' ) ) ) SCREAMING_SNAKE_CASE : int = cached_file(UpperCamelCase__ , '''conf''' , _raise_exceptions_for_missing_entries=UpperCamelCase__ ) self.assertIsNone(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : List[Any] = cached_file(UpperCamelCase__ , '''conf''' , local_files_only=UpperCamelCase__ , _raise_exceptions_for_missing_entries=UpperCamelCase__ ) self.assertIsNone(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Tuple = mock.Mock() SCREAMING_SNAKE_CASE : Optional[int] = 500 SCREAMING_SNAKE_CASE : List[str] = {} SCREAMING_SNAKE_CASE : Optional[Any] = HTTPError SCREAMING_SNAKE_CASE : Union[str, Any] = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=UpperCamelCase__ ) as mock_head: SCREAMING_SNAKE_CASE : List[Any] = cached_file(UpperCamelCase__ , '''conf''' , _raise_exceptions_for_connection_errors=UpperCamelCase__ ) self.assertIsNone(UpperCamelCase__ ) # This check we did call the fake head request mock_head.assert_called() def __A ( self : List[Any] ): '''simple docstring''' self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCamelCase__ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCamelCase__ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCamelCase__ ) ) def __A ( self : Tuple ): '''simple docstring''' self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(UpperCamelCase__ , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , UpperCamelCase__ ) # The function raises if the revision does not exist. with self.assertRaisesRegex(UpperCamelCase__ , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , UpperCamelCase__ , revision='''ahaha''' ) SCREAMING_SNAKE_CASE : Optional[int] = get_file_from_repo('''bert-base-cased''' , UpperCamelCase__ ) # The name is the cached name which is not very easy to test, so instead we load the content. SCREAMING_SNAKE_CASE : Tuple = json.loads(open(UpperCamelCase__ , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 768 ) def __A ( self : List[str] ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE : Union[str, Any] = Path(UpperCamelCase__ ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(UpperCamelCase__ , '''a.txt''' ) , str(UpperCamelCase__ ) ) self.assertIsNone(get_file_from_repo(UpperCamelCase__ , '''b.txt''' ) )
182
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class lowercase__ ( unittest.TestCase): def __A ( self : str , UpperCamelCase__ : int ): '''simple docstring''' for model_result in results.values(): for batch_size, sequence_length in zip(model_result['''bs'''] , model_result['''ss'''] ): SCREAMING_SNAKE_CASE : Union[str, Any] = model_result['''result'''][batch_size][sequence_length] self.assertIsNotNone(UpperCamelCase__ ) def __A ( self : Optional[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : List[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : str = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = '''sgugger/tiny-distilbert-classification''' SCREAMING_SNAKE_CASE : Union[str, Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , only_pretrain_model=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : List[str] = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : Any ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : int = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Union[str, Any] = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : Optional[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : int = AutoConfig.from_pretrained(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[int] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : str = TensorFlowBenchmark(UpperCamelCase__ , [config] ) SCREAMING_SNAKE_CASE : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : int ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : Optional[int] = AutoConfig.from_pretrained(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : List[str] = TensorFlowBenchmark(UpperCamelCase__ , [config] ) SCREAMING_SNAKE_CASE : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : List[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : int = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Optional[int] = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self : Optional[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : int = AutoConfig.from_pretrained(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Tuple = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Dict = TensorFlowBenchmark(UpperCamelCase__ , [config] ) SCREAMING_SNAKE_CASE : str = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __A ( self : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = '''patrickvonplaten/t5-tiny-random''' SCREAMING_SNAKE_CASE : Union[str, Any] = AutoConfig.from_pretrained(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Union[str, Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Dict = TensorFlowBenchmark(UpperCamelCase__ , configs=[config] ) SCREAMING_SNAKE_CASE : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices('''GPU''' ) ) == 0 , '''Cannot do xla on CPU.''' ) def __A ( self : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = '''sshleifer/tiny-gpt2''' SCREAMING_SNAKE_CASE : Tuple = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , use_xla=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Any = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __A ( self : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = '''sshleifer/tiny-gpt2''' with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE : Tuple = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCamelCase__ , save_to_csv=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCamelCase__ , '''inf_time.csv''' ) , inference_memory_csv_file=os.path.join(UpperCamelCase__ , '''inf_mem.csv''' ) , env_info_csv_file=os.path.join(UpperCamelCase__ , '''env.csv''' ) , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : Optional[Any] = TensorFlowBenchmark(UpperCamelCase__ ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCamelCase__ , '''inf_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , '''inf_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , '''env.csv''' ) ).exists() ) def __A ( self : int ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = '''sshleifer/tiny-gpt2''' def _check_summary_is_not_empty(UpperCamelCase__ : Dict ): self.assertTrue(hasattr(UpperCamelCase__ , '''sequential''' ) ) self.assertTrue(hasattr(UpperCamelCase__ , '''cumulative''' ) ) self.assertTrue(hasattr(UpperCamelCase__ , '''current''' ) ) self.assertTrue(hasattr(UpperCamelCase__ , '''total''' ) ) with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE : List[str] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCamelCase__ , '''log.txt''' ) , log_print=UpperCamelCase__ , trace_memory_line_by_line=UpperCamelCase__ , eager_mode=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE : List[Any] = TensorFlowBenchmark(UpperCamelCase__ ) SCREAMING_SNAKE_CASE : Dict = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , '''log.txt''' ) ).exists() )
182
1
'''simple docstring''' import inspect import os import torch from transformers import AutoModel from transformers.testing_utils import mockenv_context from transformers.trainer_utils import set_seed import accelerate from accelerate.accelerator import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils.testing import ( AccelerateTestCase, TempDirTestCase, execute_subprocess_async, require_cuda, require_fsdp, require_multi_gpu, slow, ) from accelerate.utils.constants import ( FSDP_AUTO_WRAP_POLICY, FSDP_BACKWARD_PREFETCH, FSDP_SHARDING_STRATEGY, FSDP_STATE_DICT_TYPE, ) from accelerate.utils.dataclasses import FullyShardedDataParallelPlugin from accelerate.utils.other import patch_environment set_seed(4_2) _SCREAMING_SNAKE_CASE = '''bert-base-cased''' _SCREAMING_SNAKE_CASE = '''fp16''' _SCREAMING_SNAKE_CASE = '''bf16''' _SCREAMING_SNAKE_CASE = [FPaa, BFaa] @require_fsdp @require_cuda class __lowercase ( lowerCAmelCase__ ): '''simple docstring''' def _UpperCAmelCase (self ) -> Union[str, Any]: '''simple docstring''' super().setUp() __lowercase = dict( ACCELERATE_USE_FSDP='''true''' ,MASTER_ADDR='''localhost''' ,MASTER_PORT='''10999''' ,RANK='''0''' ,LOCAL_RANK='''0''' ,WORLD_SIZE='''1''' ,) def _UpperCAmelCase (self ) -> Optional[int]: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import ShardingStrategy for i, strategy in enumerate(_lowerCamelCase ): __lowercase = self.dist_env.copy() __lowercase = f"{i + 1}" __lowercase = strategy with mockenv_context(**_lowerCamelCase ): __lowercase = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.sharding_strategy ,ShardingStrategy(i + 1 ) ) def _UpperCAmelCase (self ) -> Dict: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import BackwardPrefetch for i, prefetch_policy in enumerate(_lowerCamelCase ): __lowercase = self.dist_env.copy() __lowercase = prefetch_policy with mockenv_context(**_lowerCamelCase ): __lowercase = FullyShardedDataParallelPlugin() if prefetch_policy == "NO_PREFETCH": self.assertIsNone(fsdp_plugin.backward_prefetch ) else: self.assertEqual(fsdp_plugin.backward_prefetch ,BackwardPrefetch(i + 1 ) ) def _UpperCAmelCase (self ) -> Optional[Any]: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType for i, state_dict_type in enumerate(_lowerCamelCase ): __lowercase = self.dist_env.copy() __lowercase = state_dict_type with mockenv_context(**_lowerCamelCase ): __lowercase = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.state_dict_type ,StateDictType(i + 1 ) ) if state_dict_type == "FULL_STATE_DICT": self.assertTrue(fsdp_plugin.state_dict_config.offload_to_cpu ) self.assertTrue(fsdp_plugin.state_dict_config.ranka_only ) def _UpperCAmelCase (self ) -> int: '''simple docstring''' __lowercase = AutoModel.from_pretrained(_lowerCamelCase ) for policy in FSDP_AUTO_WRAP_POLICY: __lowercase = self.dist_env.copy() __lowercase = policy if policy == "TRANSFORMER_BASED_WRAP": __lowercase = '''BertLayer''' elif policy == "SIZE_BASED_WRAP": __lowercase = '''2000''' with mockenv_context(**_lowerCamelCase ): __lowercase = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_lowerCamelCase ) if policy == "NO_WRAP": self.assertIsNone(fsdp_plugin.auto_wrap_policy ) else: self.assertIsNotNone(fsdp_plugin.auto_wrap_policy ) __lowercase = self.dist_env.copy() __lowercase = '''TRANSFORMER_BASED_WRAP''' __lowercase = '''T5Layer''' with mockenv_context(**_lowerCamelCase ): __lowercase = FullyShardedDataParallelPlugin() with self.assertRaises(_lowerCamelCase ) as cm: fsdp_plugin.set_auto_wrap_policy(_lowerCamelCase ) self.assertTrue('''Could not find the transformer layer class to wrap in the model.''' in str(cm.exception ) ) __lowercase = self.dist_env.copy() __lowercase = '''SIZE_BASED_WRAP''' __lowercase = '''0''' with mockenv_context(**_lowerCamelCase ): __lowercase = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_lowerCamelCase ) self.assertIsNone(fsdp_plugin.auto_wrap_policy ) def _UpperCAmelCase (self ) -> Any: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import MixedPrecision from torch.distributed.fsdp.sharded_grad_scaler import ShardedGradScaler for mp_dtype in dtypes: __lowercase = self.dist_env.copy() __lowercase = mp_dtype with mockenv_context(**_lowerCamelCase ): __lowercase = Accelerator() if mp_dtype == "fp16": __lowercase = torch.floataa elif mp_dtype == "bf16": __lowercase = torch.bfloataa __lowercase = MixedPrecision(param_dtype=_lowerCamelCase ,reduce_dtype=_lowerCamelCase ,buffer_dtype=_lowerCamelCase ) self.assertEqual(accelerator.state.fsdp_plugin.mixed_precision_policy ,_lowerCamelCase ) if mp_dtype == FPaa: self.assertTrue(isinstance(accelerator.scaler ,_lowerCamelCase ) ) elif mp_dtype == BFaa: self.assertIsNone(accelerator.scaler ) AcceleratorState._reset_state(_lowerCamelCase ) def _UpperCAmelCase (self ) -> int: '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import CPUOffload for flag in [True, False]: __lowercase = self.dist_env.copy() __lowercase = str(_lowerCamelCase ).lower() with mockenv_context(**_lowerCamelCase ): __lowercase = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.cpu_offload ,CPUOffload(offload_params=_lowerCamelCase ) ) @require_fsdp @require_multi_gpu @slow class __lowercase ( lowerCAmelCase__ ): '''simple docstring''' def _UpperCAmelCase (self ) -> Tuple: '''simple docstring''' super().setUp() __lowercase = 0.8_2 __lowercase = [ '''fsdp_shard_grad_op_transformer_based_wrap''', '''fsdp_full_shard_transformer_based_wrap''', ] __lowercase = { '''multi_gpu_fp16''': 3200, '''fsdp_shard_grad_op_transformer_based_wrap_fp16''': 2000, '''fsdp_full_shard_transformer_based_wrap_fp16''': 1900, # Disabling below test as it overwhelms the RAM memory usage # on CI self-hosted runner leading to tests getting killed. # "fsdp_full_shard_cpu_offload_transformer_based_wrap_fp32": 1500, # fp16 was leading to indefinite hang } __lowercase = 160 __lowercase = 160 __lowercase = inspect.getfile(accelerate.test_utils ) __lowercase = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''external_deps'''] ) def _UpperCAmelCase (self ) -> List[Any]: '''simple docstring''' __lowercase = os.path.join(self.test_scripts_folder ,'''test_performance.py''' ) __lowercase = ['''accelerate''', '''launch''', '''--num_processes=2''', '''--num_machines=1''', '''--machine_rank=0''', '''--use_fsdp'''] for config in self.performance_configs: __lowercase = cmd.copy() for i, strategy in enumerate(_lowerCamelCase ): if strategy.lower() in config: cmd_config.append(f"--fsdp_sharding_strategy={i+1}" ) break if "fp32" in config: cmd_config.append('''--mixed_precision=no''' ) else: cmd_config.append('''--mixed_precision=fp16''' ) if "cpu_offload" in config: cmd_config.append('''--fsdp_offload_params=True''' ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in config: cmd_config.append(f"--fsdp_auto_wrap_policy={policy}" ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append('''--fsdp_transformer_layer_cls_to_wrap=BertLayer''' ) elif policy == "SIZE_BASED_WRAP": cmd_config.append('''--fsdp_min_num_params=2000''' ) cmd_config.extend( [ self.test_file_path, f"--output_dir={self.tmpdir}", f"--performance_lower_bound={self.performance_lower_bound}", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_lowerCamelCase ,env=os.environ.copy() ) def _UpperCAmelCase (self ) -> List[Any]: '''simple docstring''' __lowercase = os.path.join(self.test_scripts_folder ,'''test_checkpointing.py''' ) __lowercase = [ '''accelerate''', '''launch''', '''--num_processes=2''', '''--num_machines=1''', '''--machine_rank=0''', '''--use_fsdp''', '''--mixed_precision=fp16''', '''--fsdp_transformer_layer_cls_to_wrap=BertLayer''', ] for i, strategy in enumerate(_lowerCamelCase ): __lowercase = cmd.copy() cmd_config.append(f"--fsdp_sharding_strategy={i+1}" ) if strategy != "FULL_SHARD": continue __lowercase = len(_lowerCamelCase ) for state_dict_type in FSDP_STATE_DICT_TYPE: __lowercase = cmd_config[:state_dict_config_index] cmd_config.append(f"--fsdp_state_dict_type={state_dict_type}" ) cmd_config.extend( [ self.test_file_path, f"--output_dir={self.tmpdir}", '''--partial_train_epoch=1''', ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_lowerCamelCase ,env=os.environ.copy() ) __lowercase = cmd_config[:-1] __lowercase = os.path.join(self.tmpdir ,'''epoch_0''' ) cmd_config.extend( [ f"--resume_from_checkpoint={resume_from_checkpoint}", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_lowerCamelCase ,env=os.environ.copy() ) def _UpperCAmelCase (self ) -> Any: '''simple docstring''' __lowercase = os.path.join(self.test_scripts_folder ,'''test_peak_memory_usage.py''' ) __lowercase = [ '''accelerate''', '''launch''', '''--num_processes=2''', '''--num_machines=1''', '''--machine_rank=0''', ] for spec, peak_mem_upper_bound in self.peak_memory_usage_upper_bound.items(): __lowercase = cmd.copy() if "fp16" in spec: cmd_config.extend(['''--mixed_precision=fp16'''] ) else: cmd_config.extend(['''--mixed_precision=no'''] ) if "multi_gpu" in spec: continue else: cmd_config.extend(['''--use_fsdp'''] ) for i, strategy in enumerate(_lowerCamelCase ): if strategy.lower() in spec: cmd_config.append(f"--fsdp_sharding_strategy={i+1}" ) break if "cpu_offload" in spec: cmd_config.append('''--fsdp_offload_params=True''' ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in spec: cmd_config.append(f"--fsdp_auto_wrap_policy={policy}" ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append('''--fsdp_transformer_layer_cls_to_wrap=BertLayer''' ) elif policy == "SIZE_BASED_WRAP": cmd_config.append('''--fsdp_min_num_params=2000''' ) cmd_config.extend( [ self.test_file_path, f"--output_dir={self.tmpdir}", f"--peak_memory_upper_bound={peak_mem_upper_bound}", f"--n_train={self.n_train}", f"--n_val={self.n_val}", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_lowerCamelCase ,env=os.environ.copy() )
350
'''simple docstring''' import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def _lowerCAmelCase ( ): print('''Making key files...''' ) make_key_files('''rsa''' , 1_0_2_4 ) print('''Key files generation successful.''' ) def _lowerCAmelCase ( lowerCamelCase_ : int ): print('''Generating prime p...''' ) __lowercase = rabinMiller.generate_large_prime(lowerCamelCase_ ) print('''Generating prime q...''' ) __lowercase = rabinMiller.generate_large_prime(lowerCamelCase_ ) __lowercase = p * q print('''Generating e that is relatively prime to (p - 1) * (q - 1)...''' ) while True: __lowercase = random.randrange(2 ** (key_size - 1) , 2 ** (key_size) ) if cryptoMath.gcd(lowerCamelCase_ , (p - 1) * (q - 1) ) == 1: break print('''Calculating d that is mod inverse of e...''' ) __lowercase = cryptoMath.find_mod_inverse(lowerCamelCase_ , (p - 1) * (q - 1) ) __lowercase = (n, e) __lowercase = (n, d) return (public_key, private_key) def _lowerCAmelCase ( lowerCamelCase_ : str , lowerCamelCase_ : int ): if os.path.exists(f"{name}_pubkey.txt" ) or os.path.exists(f"{name}_privkey.txt" ): print('''\nWARNING:''' ) print( f"\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n" '''Use a different name or delete these files and re-run this program.''' ) sys.exit() __lowercase , __lowercase = generate_key(lowerCamelCase_ ) print(f"\nWriting public key to file {name}_pubkey.txt..." ) with open(f"{name}_pubkey.txt" , '''w''' ) as out_file: out_file.write(f"{key_size},{public_key[0]},{public_key[1]}" ) print(f"Writing private key to file {name}_privkey.txt..." ) with open(f"{name}_privkey.txt" , '''w''' ) as out_file: out_file.write(f"{key_size},{private_key[0]},{private_key[1]}" ) if __name__ == "__main__": main()
217
0
"""simple docstring""" from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class UpperCAmelCase_ ( __UpperCAmelCase): lowerCamelCase__ : Optional[int] = ['''image_processor''', '''tokenizer'''] lowerCamelCase__ : Dict = '''BlipImageProcessor''' lowerCamelCase__ : Optional[Any] = ('''BertTokenizer''', '''BertTokenizerFast''') def __init__( self , a , a ) -> Any: lowercase__ : List[str] = False super().__init__(a , a ) lowercase__ : List[Any] = self.image_processor def __call__( self , a = None , a = None , a = True , a = False , a = None , a = None , a = 0 , a = None , a = None , a = False , a = False , a = False , a = False , a = False , a = True , a = None , **a , ) -> BatchEncoding: if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None: lowercase__ : Tuple = self.tokenizer lowercase__ : Tuple = self.tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) return text_encoding # add pixel_values lowercase__ : Optional[int] = self.image_processor(a , return_tensors=a ) if text is not None: lowercase__ : List[str] = self.tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) else: lowercase__ : Any = None if text_encoding is not None: encoding_image_processor.update(a ) return encoding_image_processor def _UpperCAmelCase ( self , *a , **a ) -> Union[str, Any]: return self.tokenizer.batch_decode(*a , **a ) def _UpperCAmelCase ( self , *a , **a ) -> List[Any]: return self.tokenizer.decode(*a , **a ) @property def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Tuple = self.tokenizer.model_input_names lowercase__ : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
77
import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class A__ : """simple docstring""" def __init__( self , lowercase , lowercase=13 , lowercase=64 , lowercase=2 , lowercase=3 , lowercase=True , lowercase=True , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=10 , lowercase=0.02 , lowercase=[1, 16, 4, 4] , lowercase=None , ) -> List[Any]: '''simple docstring''' a__ : Optional[int] = parent a__ : Optional[int] = batch_size a__ : Any = image_size a__ : Optional[Any] = patch_size a__ : Optional[Any] = num_channels a__ : int = is_training a__ : List[str] = use_labels a__ : List[str] = hidden_size a__ : Tuple = num_hidden_layers a__ : Optional[Any] = num_attention_heads a__ : Union[str, Any] = intermediate_size a__ : Optional[int] = hidden_act a__ : Optional[Any] = hidden_dropout_prob a__ : Any = attention_probs_dropout_prob a__ : Any = type_sequence_label_size a__ : Tuple = initializer_range a__ : Tuple = scope a__ : int = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size a__ : Any = (self.image_size // 32) ** 2 a__ : List[Any] = num_patches + 1 def __lowercase ( self) -> Any: '''simple docstring''' a__ : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) a__ : int = None if self.use_labels: a__ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size) a__ : List[str] = self.get_config() return config, pixel_values, labels def __lowercase ( self) -> Dict: '''simple docstring''' a__ : List[str] = { 'global_padding': 'same', 'layer_type': 'bottleneck', 'depths': [3, 4, 9], 'out_features': ['stage1', 'stage2', 'stage3'], 'embedding_dynamic_padding': True, 'hidden_sizes': [4, 8, 16, 32], 'num_groups': 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=lowercase , ) def __lowercase ( self , lowercase , lowercase , lowercase) -> List[str]: '''simple docstring''' a__ : List[str] = ViTHybridModel(config=lowercase) model.to(lowercase) model.eval() a__ : Union[str, Any] = model(lowercase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def __lowercase ( self , lowercase , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ : Dict = self.type_sequence_label_size a__ : Union[str, Any] = ViTHybridForImageClassification(lowercase) model.to(lowercase) model.eval() a__ : Tuple = model(lowercase , labels=lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) def __lowercase ( self) -> Any: '''simple docstring''' a__ : List[str] = self.prepare_config_and_inputs() a__ , a__ , a__ : Union[str, Any] = config_and_inputs a__ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): """simple docstring""" __A : Optional[Any] = (ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () __A : List[str] = ( {'''feature-extraction''': ViTHybridModel, '''image-classification''': ViTHybridForImageClassification} if is_torch_available() else {} ) __A : Any = False __A : Optional[int] = False __A : Optional[Any] = False def __lowercase ( self) -> Optional[Any]: '''simple docstring''' a__ : Any = ViTHybridModelTester(self) a__ : Any = ConfigTester(self , config_class=lowercase , has_text_modality=lowercase , hidden_size=37) def __lowercase ( self) -> List[Any]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds') def __lowercase ( self) -> Dict: '''simple docstring''' pass def __lowercase ( self) -> Optional[Any]: '''simple docstring''' a__ , a__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__ : str = model_class(lowercase) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) a__ : str = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear)) def __lowercase ( self) -> int: '''simple docstring''' a__ , a__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__ : Union[str, Any] = model_class(lowercase) a__ : Union[str, Any] = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic a__ : Optional[Any] = [*signature.parameters.keys()] a__ : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , lowercase) def __lowercase ( self) -> Any: '''simple docstring''' a__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase) def __lowercase ( self) -> Optional[Any]: '''simple docstring''' a__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase) def __lowercase ( self) -> Dict: '''simple docstring''' a__ , a__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() a__ : Tuple = _config_zero_init(lowercase) for model_class in self.all_model_classes: a__ : List[Any] = model_class(config=lowercase) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": a__ : Dict = [F'{name}.{key}' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @slow def __lowercase ( self) -> Any: '''simple docstring''' for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__ : Optional[Any] = ViTHybridModel.from_pretrained(lowercase) self.assertIsNotNone(lowercase) def A_ ( ) -> int: a__ : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self) -> Optional[Any]: '''simple docstring''' return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0]) if is_vision_available() else None ) @slow def __lowercase ( self) -> Any: '''simple docstring''' a__ : List[str] = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0]).to( lowercase) a__ : List[str] = self.default_image_processor a__ : List[Any] = prepare_img() a__ : Any = image_processor(images=lowercase , return_tensors='pt').to(lowercase) # forward pass with torch.no_grad(): a__ : Optional[Any] = model(**lowercase) # verify the logits a__ : Optional[Any] = torch.Size((1, 1000)) self.assertEqual(outputs.logits.shape , lowercase) a__ : Any = torch.tensor([-1.90_90, -0.49_93, -0.23_89]).to(lowercase) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase , atol=1e-4)) @slow @require_accelerate def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : List[str] = ViTHybridImageProcessor.from_pretrained('google/vit-hybrid-base-bit-384') a__ : Union[str, Any] = ViTHybridForImageClassification.from_pretrained('google/vit-hybrid-base-bit-384' , device_map='auto') a__ : Any = prepare_img() a__ : str = image_processor(images=lowercase , return_tensors='pt') a__ : List[Any] = model(**lowercase) a__ : int = outputs.logits # model predicts one of the 1000 ImageNet classes a__ : List[str] = logits.argmax(-1).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , 'tabby, tabby cat')
99
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class UpperCAmelCase__ ( A__ ): """simple docstring""" a = ( "This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image." "It takes two arguments named `image` which should be the original image, and `label` which should be a text " "describing the elements what should be identified in the segmentation mask. The tool returns the mask." ) a = "CIDAS/clipseg-rd64-refined" a = "image_segmenter" a = CLIPSegForImageSegmentation a = ["image", "text"] a = ["image"] def __init__( self : Union[str, Any] , *__lowerCamelCase : List[str] , **__lowerCamelCase : Optional[int] ) -> Optional[int]: requires_backends(self , ['''vision'''] ) super().__init__(*__lowerCamelCase , **__lowerCamelCase ) def lowercase_ ( self : List[Any] , __lowerCamelCase : "Image" , __lowerCamelCase : str ) -> List[Any]: return self.pre_processor(text=[label] , images=[image] , padding=__lowerCamelCase , return_tensors='''pt''' ) def lowercase_ ( self : Dict , __lowerCamelCase : Any ) -> Optional[Any]: with torch.no_grad(): SCREAMING_SNAKE_CASE__ = self.model(**__lowerCamelCase ).logits return logits def lowercase_ ( self : List[Any] , __lowerCamelCase : Optional[int] ) -> Tuple: SCREAMING_SNAKE_CASE__ = outputs.cpu().detach().numpy() SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 return Image.fromarray((array * 255).astype(np.uinta ) )
218
from argparse import ArgumentParser from .env import EnvironmentCommand def UpperCAmelCase_ ( ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = ArgumentParser('''Diffusers CLI tool''' , usage='''diffusers-cli <command> [<args>]''' ) SCREAMING_SNAKE_CASE__ = parser.add_subparsers(help='''diffusers-cli command helpers''' ) # Register commands EnvironmentCommand.register_subcommand(_A ) # Let's go SCREAMING_SNAKE_CASE__ = parser.parse_args() if not hasattr(_A , '''func''' ): parser.print_help() exit(1 ) # Run SCREAMING_SNAKE_CASE__ = args.func(_A ) service.run() if __name__ == "__main__": main()
218
1
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__UpperCamelCase ) , 'Tatoeba directory does not exist.' ) class __a ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE__ ( self ) -> List[str]: '''simple docstring''' lowercase__: int = tempfile.mkdtemp() return TatoebaConverter(save_dir=lowerCAmelCase__ ) @slow def SCREAMING_SNAKE_CASE__ ( self ) -> Any: '''simple docstring''' self.resolver.convert_models(['heb-eng'] ) @slow def SCREAMING_SNAKE_CASE__ ( self ) -> Union[str, Any]: '''simple docstring''' lowercase__ , lowercase__: List[str] = self.resolver.write_model_card('opus-mt-he-en' , dry_run=lowerCAmelCase__ ) assert mmeta["long_pair"] == "heb-eng"
196
import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class __a ( __UpperCamelCase , unittest.TestCase ): __lowercase : str = CpmAntTokenizer __lowercase : List[Any] = False def SCREAMING_SNAKE_CASE__ ( self ) -> str: '''simple docstring''' super().setUp() lowercase__: Any = [ '<d>', '</d>', '<s>', '</s>', '</_>', '<unk>', '<pad>', '</n>', '我', '是', 'C', 'P', 'M', 'A', 'n', 't', ] lowercase__: List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) @tooslow def SCREAMING_SNAKE_CASE__ ( self ) -> Tuple: '''simple docstring''' lowercase__: Optional[int] = CpmAntTokenizer.from_pretrained('openbmb/cpm-ant-10b' ) lowercase__: Optional[Any] = '今天天气真好!' lowercase__: str = ['今天', '天气', '真', '好', '!'] lowercase__: Optional[Any] = tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase__: List[str] = '今天天气真好!' lowercase__: List[str] = [tokenizer.bos_token] + tokens lowercase__: Tuple = [6, 9_802, 14_962, 2_082, 831, 244] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , lowerCAmelCase__ ) lowercase__: Any = tokenizer.decode(lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ )
196
1
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase__ ( lowerCamelCase : str ,lowerCamelCase : Any ,lowerCamelCase : List[str] ): # Initialise PyTorch model _A : Optional[Any] = TaConfig.from_json_file(lowerCamelCase ) print(F'Building PyTorch model from configuration: {config}' ) _A : List[Any] = TaForConditionalGeneration(lowerCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_ta(lowerCamelCase ,lowerCamelCase ,lowerCamelCase ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(lowerCamelCase ) if __name__ == "__main__": A : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) A : Optional[Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
227
'''simple docstring''' import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowerCamelCase ( a_ , unittest.TestCase ): """simple docstring""" a = KandinskyVaaPriorPipeline a = ["prompt"] a = ["prompt", "negative_prompt"] a = [ "num_images_per_prompt", "generator", "num_inference_steps", "latents", "negative_prompt", "guidance_scale", "output_type", "return_dict", ] a = False @property def A ( self : List[str]): return 32 @property def A ( self : List[Any]): return 32 @property def A ( self : Dict): return self.time_input_dim @property def A ( self : Tuple): return self.time_input_dim * 4 @property def A ( self : Optional[int]): return 100 @property def A ( self : Dict): _A : Tuple = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') return tokenizer @property def A ( self : Optional[Any]): torch.manual_seed(0) _A : Any = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(SCREAMING_SNAKE_CASE) @property def A ( self : List[Any]): torch.manual_seed(0) _A : Optional[Any] = { 'num_attention_heads': 2, 'attention_head_dim': 12, 'embedding_dim': self.text_embedder_hidden_size, 'num_layers': 1, } _A : Any = PriorTransformer(**SCREAMING_SNAKE_CASE) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 _A : str = nn.Parameter(torch.ones(model.clip_std.shape)) return model @property def A ( self : List[str]): torch.manual_seed(0) _A : List[Any] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) _A : Union[str, Any] = CLIPVisionModelWithProjection(SCREAMING_SNAKE_CASE) return model @property def A ( self : int): _A : Optional[Any] = CLIPImageProcessor( crop_size=224 , do_center_crop=SCREAMING_SNAKE_CASE , do_normalize=SCREAMING_SNAKE_CASE , do_resize=SCREAMING_SNAKE_CASE , image_mean=[0.4814_5466, 0.457_8275, 0.4082_1073] , image_std=[0.2686_2954, 0.2613_0258, 0.2757_7711] , resample=3 , size=224 , ) return image_processor def A ( self : Optional[Any]): _A : Optional[int] = self.dummy_prior _A : Dict = self.dummy_image_encoder _A : Dict = self.dummy_text_encoder _A : str = self.dummy_tokenizer _A : Optional[Any] = self.dummy_image_processor _A : Optional[Any] = UnCLIPScheduler( variance_type='fixed_small_log' , prediction_type='sample' , num_train_timesteps=1000 , clip_sample=SCREAMING_SNAKE_CASE , clip_sample_range=10.0 , ) _A : Dict = { 'prior': prior, 'image_encoder': image_encoder, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'scheduler': scheduler, 'image_processor': image_processor, } return components def A ( self : Tuple , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : List[str]=0): if str(SCREAMING_SNAKE_CASE).startswith('mps'): _A : Dict = torch.manual_seed(SCREAMING_SNAKE_CASE) else: _A : int = torch.Generator(device=SCREAMING_SNAKE_CASE).manual_seed(SCREAMING_SNAKE_CASE) _A : List[Any] = { 'prompt': 'horse', 'generator': generator, 'guidance_scale': 4.0, 'num_inference_steps': 2, 'output_type': 'np', } return inputs def A ( self : List[Any]): _A : str = 'cpu' _A : Tuple = self.get_dummy_components() _A : List[Any] = self.pipeline_class(**SCREAMING_SNAKE_CASE) _A : Any = pipe.to(SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE) _A : Dict = pipe(**self.get_dummy_inputs(SCREAMING_SNAKE_CASE)) _A : str = output.image_embeds _A : Optional[int] = pipe( **self.get_dummy_inputs(SCREAMING_SNAKE_CASE) , return_dict=SCREAMING_SNAKE_CASE , )[0] _A : Optional[int] = image[0, -10:] _A : int = image_from_tuple[0, -10:] assert image.shape == (1, 32) _A : Dict = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 @skip_mps def A ( self : Any): _A : Tuple = torch_device == 'cpu' _A : Optional[int] = True _A : Tuple = False self._test_inference_batch_single_identical( test_max_difference=SCREAMING_SNAKE_CASE , relax_max_difference=SCREAMING_SNAKE_CASE , test_mean_pixel_difference=SCREAMING_SNAKE_CASE , ) @skip_mps def A ( self : int): _A : Tuple = torch_device == 'cpu' _A : Optional[Any] = False self._test_attention_slicing_forward_pass( test_max_difference=SCREAMING_SNAKE_CASE , test_mean_pixel_difference=SCREAMING_SNAKE_CASE , )
227
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : Optional[Any] = {'''tokenization_bertweet''': ['''BertweetTokenizer''']} if TYPE_CHECKING: from .tokenization_bertweet import BertweetTokenizer else: import sys __UpperCamelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
106
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss lowerCAmelCase :Dict = pytest.mark.integration @require_faiss class _lowerCamelCase ( lowercase__ ): '''simple docstring''' def __lowerCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: __magic_name__ : str = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(_A ) for x in np.arange(30 ).tolist()]} ) return dset def __lowerCAmelCase ( self : List[str] ) -> Tuple: import faiss __magic_name__ : Dataset = self._create_dummy_dataset() __magic_name__ : Union[str, Any] = dset.map( lambda _A , _A : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=_A , keep_in_memory=_A ) __magic_name__ : int = dset.add_faiss_index('vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT ) __magic_name__ , __magic_name__ : List[str] = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def __lowerCAmelCase ( self : Any ) -> str: import faiss __magic_name__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT , ) __magic_name__ , __magic_name__ : Any = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def __lowerCAmelCase ( self : Tuple ) -> int: import faiss __magic_name__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=_A ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) __magic_name__ , __magic_name__ : Dict = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def __lowerCAmelCase ( self : Union[str, Any] ) -> Optional[int]: __magic_name__ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(_A , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def __lowerCAmelCase ( self : List[Any] ) -> Tuple: from elasticsearch import Elasticsearch __magic_name__ : Dataset = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __magic_name__ : int = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 30 ) __magic_name__ : List[Any] = {'hits': {'hits': [{'_score': 1, '_id': 29}]}} __magic_name__ : Union[str, Any] = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=_A ) __magic_name__ , __magic_name__ : Tuple = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class _lowerCamelCase ( lowercase__ ): '''simple docstring''' def __lowerCAmelCase ( self : Tuple ) -> List[Any]: import faiss __magic_name__ : int = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 10 ) # single query __magic_name__ : str = np.zeros(5 , dtype=np.floataa ) __magic_name__ : Optional[int] = 1 __magic_name__ , __magic_name__ : str = index.search(_A ) self.assertRaises(_A , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries __magic_name__ : Optional[Any] = np.eye(5 , dtype=np.floataa )[::-1] __magic_name__ , __magic_name__ : str = index.search_batch(_A ) self.assertRaises(_A , index.search_batch , queries[0] ) __magic_name__ : List[Any] = [scores[0] for scores in total_scores] __magic_name__ : List[str] = [indices[0] for indices in total_indices] self.assertGreater(np.min(_A ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , _A ) def __lowerCAmelCase ( self : Dict ) -> Optional[Any]: import faiss __magic_name__ : str = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) __magic_name__ : str = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(_A ): __magic_name__ : Dict = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def __lowerCAmelCase ( self : Union[str, Any] ) -> Dict: import faiss __magic_name__ : Any = faiss.IndexFlat(5 ) __magic_name__ : Optional[Any] = FaissIndex(custom_index=_A ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def __lowerCAmelCase ( self : Dict ) -> Tuple: import faiss __magic_name__ : Optional[int] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=_A ) as tmp_file: index.save(tmp_file.name ) __magic_name__ : Optional[int] = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) __magic_name__ : Dict = np.zeros(5 , dtype=np.floataa ) __magic_name__ : Tuple = 1 __magic_name__ , __magic_name__ : Optional[Any] = index.search(_A ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def lowerCamelCase ( lowerCAmelCase : Tuple ): """simple docstring""" import faiss __magic_name__ : Union[str, Any] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) __magic_name__ : Dict = 'index.faiss' __magic_name__ : Optional[Any] = f'mock://{index_name}' index.save(lowerCAmelCase , storage_options=mockfs.storage_options ) __magic_name__ : Tuple = FaissIndex.load(lowerCAmelCase , storage_options=mockfs.storage_options ) __magic_name__ : Union[str, Any] = np.zeros(5 , dtype=np.floataa ) __magic_name__ : List[str] = 1 __magic_name__ , __magic_name__ : Dict = index.search(lowerCAmelCase ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class _lowerCamelCase ( lowercase__ ): '''simple docstring''' def __lowerCAmelCase ( self : Tuple ) -> Dict: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __magic_name__ : Any = Elasticsearch() __magic_name__ : Union[str, Any] = {'acknowledged': True} __magic_name__ : Tuple = ElasticSearchIndex(es_client=_A ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query __magic_name__ : str = 'foo' __magic_name__ : str = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __magic_name__ , __magic_name__ : Dict = index.search(_A ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout __magic_name__ : str = 'foo' __magic_name__ : Dict = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __magic_name__ , __magic_name__ : Dict = index.search(_A , request_timeout=30 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries __magic_name__ : Optional[Any] = ['foo', 'bar', 'foobar'] __magic_name__ : Optional[Any] = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __magic_name__ , __magic_name__ : Optional[Any] = index.search_batch(_A ) __magic_name__ : Tuple = [scores[0] for scores in total_scores] __magic_name__ : List[str] = [indices[0] for indices in total_indices] self.assertGreater(np.min(_A ) , 0 ) self.assertListEqual([1, 1, 1] , _A ) # batched queries with timeout __magic_name__ : Union[str, Any] = ['foo', 'bar', 'foobar'] __magic_name__ : Tuple = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __magic_name__ , __magic_name__ : Dict = index.search_batch(_A , request_timeout=30 ) __magic_name__ : Optional[int] = [scores[0] for scores in total_scores] __magic_name__ : Union[str, Any] = [indices[0] for indices in total_indices] self.assertGreater(np.min(_A ) , 0 ) self.assertListEqual([1, 1, 1] , _A )
331
0
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : int=7 , __lowerCamelCase : Union[str, Any]=3 , __lowerCamelCase : int=30 , __lowerCamelCase : List[str]=400 , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Tuple=None , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : int=[0.5, 0.5, 0.5] , __lowerCamelCase : Optional[int]=[0.5, 0.5, 0.5] , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Optional[int]=1 / 255 , __lowerCamelCase : Dict=True , ): UpperCamelCase :Optional[int] = size if size is not None else {"""shortest_edge""": 18, """longest_edge""": 1_333} UpperCamelCase :Any = parent UpperCamelCase :Dict = batch_size UpperCamelCase :Optional[int] = num_channels UpperCamelCase :str = min_resolution UpperCamelCase :Optional[Any] = max_resolution UpperCamelCase :Optional[Any] = do_resize UpperCamelCase :Optional[Any] = size UpperCamelCase :Dict = do_normalize UpperCamelCase :int = image_mean UpperCamelCase :Tuple = image_std UpperCamelCase :Tuple = do_rescale UpperCamelCase :Optional[Any] = rescale_factor UpperCamelCase :List[str] = do_pad def _A ( self : int ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _A ( self : Any , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ): if not batched: UpperCamelCase :Tuple = image_inputs[0] if isinstance(_snake_case , Image.Image ): UpperCamelCase , UpperCamelCase :Tuple = image.size else: UpperCamelCase , UpperCamelCase :Union[str, Any] = image.shape[1], image.shape[2] if w < h: UpperCamelCase :str = int(self.size["""shortest_edge"""] * h / w ) UpperCamelCase :Optional[Any] = self.size["""shortest_edge"""] elif w > h: UpperCamelCase :Optional[Any] = self.size["""shortest_edge"""] UpperCamelCase :List[str] = int(self.size["""shortest_edge"""] * w / h ) else: UpperCamelCase :Dict = self.size["""shortest_edge"""] UpperCamelCase :str = self.size["""shortest_edge"""] else: UpperCamelCase :Union[str, Any] = [] for image in image_inputs: UpperCamelCase , UpperCamelCase :Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCamelCase :Dict = max(_snake_case , key=lambda __lowerCamelCase : item[0] )[0] UpperCamelCase :Any = max(_snake_case , key=lambda __lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _SCREAMING_SNAKE_CASE ( _a , unittest.TestCase ): snake_case__ : List[str] = YolosImageProcessor if is_vision_available() else None def _A ( self : List[Any] ): UpperCamelCase :Union[str, Any] = YolosImageProcessingTester(self ) @property def _A ( self : str ): return self.image_processor_tester.prepare_image_processor_dict() def _A ( self : Union[str, Any] ): UpperCamelCase :Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_snake_case , """image_mean""" ) ) self.assertTrue(hasattr(_snake_case , """image_std""" ) ) self.assertTrue(hasattr(_snake_case , """do_normalize""" ) ) self.assertTrue(hasattr(_snake_case , """do_resize""" ) ) self.assertTrue(hasattr(_snake_case , """size""" ) ) def _A ( self : Optional[int] ): UpperCamelCase :List[str] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18, """longest_edge""": 1_333} ) self.assertEqual(image_processor.do_pad , _snake_case ) UpperCamelCase :Optional[int] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_snake_case ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42, """longest_edge""": 84} ) self.assertEqual(image_processor.do_pad , _snake_case ) def _A ( self : int ): pass def _A ( self : List[Any] ): UpperCamelCase :Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCamelCase :str = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case ) for image in image_inputs: self.assertIsInstance(_snake_case , Image.Image ) # Test not batched input UpperCamelCase :Optional[int] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :Tuple = self.image_processor_tester.get_expected_values(_snake_case ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase , UpperCamelCase :Union[str, Any] = self.image_processor_tester.get_expected_values(_snake_case , batched=_snake_case ) UpperCamelCase :Dict = image_processing(_snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _A ( self : Any ): UpperCamelCase :Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase :List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case , numpify=_snake_case ) for image in image_inputs: self.assertIsInstance(_snake_case , np.ndarray ) # Test not batched input UpperCamelCase :Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :str = self.image_processor_tester.get_expected_values(_snake_case ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase :List[Any] = image_processing(_snake_case , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :List[str] = self.image_processor_tester.get_expected_values(_snake_case , batched=_snake_case ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _A ( self : Any ): UpperCamelCase :List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase :Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case , torchify=_snake_case ) for image in image_inputs: self.assertIsInstance(_snake_case , torch.Tensor ) # Test not batched input UpperCamelCase :Optional[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :str = self.image_processor_tester.get_expected_values(_snake_case ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase :Dict = image_processing(_snake_case , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :Optional[Any] = self.image_processor_tester.get_expected_values(_snake_case , batched=_snake_case ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _A ( self : str ): UpperCamelCase :List[str] = self.image_processing_class(**self.image_processor_dict ) UpperCamelCase :Tuple = self.image_processing_class(do_resize=_snake_case , do_normalize=_snake_case , do_rescale=_snake_case ) # create random PyTorch tensors UpperCamelCase :Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case , torchify=_snake_case ) for image in image_inputs: self.assertIsInstance(_snake_case , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors UpperCamelCase :Optional[int] = image_processing_a.pad(_snake_case , return_tensors="""pt""" ) UpperCamelCase :str = image_processing_a(_snake_case , return_tensors="""pt""" ) self.assertTrue( torch.allclose(encoded_images_with_method["""pixel_values"""] , encoded_images["""pixel_values"""] , atol=1E-4 ) ) @slow def _A ( self : Tuple ): UpperCamelCase :str = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_annotations.txt""" , """r""" ) as f: UpperCamelCase :List[str] = json.loads(f.read() ) UpperCamelCase :Any = {"""image_id""": 39_769, """annotations""": target} # encode them UpperCamelCase :List[Any] = YolosImageProcessor.from_pretrained("""hustvl/yolos-small""" ) UpperCamelCase :List[str] = image_processing(images=_snake_case , annotations=_snake_case , return_tensors="""pt""" ) # verify pixel values UpperCamelCase :Optional[Any] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["""pixel_values"""].shape , _snake_case ) UpperCamelCase :int = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , _snake_case , atol=1E-4 ) ) # verify area UpperCamelCase :List[str] = torch.tensor([5887.9600, 11_250.2_061, 489_353.8_438, 837_122.7_500, 147_967.5_156, 165_732.3_438] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , _snake_case ) ) # verify boxes UpperCamelCase :Any = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , _snake_case ) UpperCamelCase :Dict = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , _snake_case , atol=1E-3 ) ) # verify image_id UpperCamelCase :Optional[int] = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , _snake_case ) ) # verify is_crowd UpperCamelCase :Union[str, Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , _snake_case ) ) # verify class_labels UpperCamelCase :int = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , _snake_case ) ) # verify orig_size UpperCamelCase :int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , _snake_case ) ) # verify size UpperCamelCase :int = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , _snake_case ) ) @slow def _A ( self : str ): UpperCamelCase :Union[str, Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt""" , """r""" ) as f: UpperCamelCase :Any = json.loads(f.read() ) UpperCamelCase :Tuple = {"""file_name""": """000000039769.png""", """image_id""": 39_769, """segments_info""": target} UpperCamelCase :int = pathlib.Path("""./tests/fixtures/tests_samples/COCO/coco_panoptic""" ) # encode them UpperCamelCase :Any = YolosImageProcessor(format="""coco_panoptic""" ) UpperCamelCase :Optional[int] = image_processing(images=_snake_case , annotations=_snake_case , masks_path=_snake_case , return_tensors="""pt""" ) # verify pixel values UpperCamelCase :int = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["""pixel_values"""].shape , _snake_case ) UpperCamelCase :Optional[Any] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , _snake_case , atol=1E-4 ) ) # verify area UpperCamelCase :Dict = torch.tensor([147_979.6_875, 165_527.0_469, 484_638.5_938, 11_292.9_375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , _snake_case ) ) # verify boxes UpperCamelCase :int = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , _snake_case ) UpperCamelCase :Tuple = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , _snake_case , atol=1E-3 ) ) # verify image_id UpperCamelCase :Union[str, Any] = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , _snake_case ) ) # verify is_crowd UpperCamelCase :List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , _snake_case ) ) # verify class_labels UpperCamelCase :Optional[Any] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , _snake_case ) ) # verify masks UpperCamelCase :Optional[Any] = 822_873 self.assertEqual(encoding["""labels"""][0]["""masks"""].sum().item() , _snake_case ) # verify orig_size UpperCamelCase :Optional[int] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , _snake_case ) ) # verify size UpperCamelCase :Optional[int] = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , _snake_case ) )
361
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Optional[Any] = logging.get_logger(__name__) UpperCAmelCase_ : int = { '''microsoft/swinv2-tiny-patch4-window8-256''': ( '''https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json''' ), } class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : List[Any] = """swinv2""" snake_case__ : Tuple = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self : Tuple , __lowerCamelCase : List[str]=224 , __lowerCamelCase : List[str]=4 , __lowerCamelCase : Union[str, Any]=3 , __lowerCamelCase : Tuple=96 , __lowerCamelCase : str=[2, 2, 6, 2] , __lowerCamelCase : Union[str, Any]=[3, 6, 12, 24] , __lowerCamelCase : int=7 , __lowerCamelCase : Dict=4.0 , __lowerCamelCase : Any=True , __lowerCamelCase : int=0.0 , __lowerCamelCase : str=0.0 , __lowerCamelCase : Dict=0.1 , __lowerCamelCase : Union[str, Any]="gelu" , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : str=0.02 , __lowerCamelCase : List[Any]=1E-5 , __lowerCamelCase : List[Any]=32 , **__lowerCamelCase : Optional[Any] , ): super().__init__(**__lowerCamelCase ) UpperCamelCase :Optional[Any] = image_size UpperCamelCase :str = patch_size UpperCamelCase :Tuple = num_channels UpperCamelCase :Optional[int] = embed_dim UpperCamelCase :Optional[int] = depths UpperCamelCase :int = len(__lowerCamelCase ) UpperCamelCase :List[Any] = num_heads UpperCamelCase :Union[str, Any] = window_size UpperCamelCase :Any = mlp_ratio UpperCamelCase :Union[str, Any] = qkv_bias UpperCamelCase :List[Any] = hidden_dropout_prob UpperCamelCase :Any = attention_probs_dropout_prob UpperCamelCase :List[Any] = drop_path_rate UpperCamelCase :List[str] = hidden_act UpperCamelCase :Optional[int] = use_absolute_embeddings UpperCamelCase :Optional[int] = layer_norm_eps UpperCamelCase :str = initializer_range UpperCamelCase :List[str] = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCamelCase :List[str] = int(embed_dim * 2 ** (len(__lowerCamelCase ) - 1) ) UpperCamelCase :Dict = (0, 0, 0, 0)
62
0
from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowercase_ ( _lowerCamelCase : str = "laptop"): lowercase__ : Optional[Any] = f'''https://www.amazon.in/laptop/s?k={product}''' lowercase__ : Dict = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } lowercase__ : List[str] = BeautifulSoup(requests.get(_lowerCamelCase , headers=_lowerCamelCase).text) # Initialize a Pandas dataframe with the column titles lowercase__ : Dict = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ]) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" , attrs={"class": "s-result-item", "data-component-type": "s-search-result"} , ) , soup.find_all("div" , attrs={"class": "a-row a-size-base a-color-base"}) , ): try: lowercase__ : List[str] = item.ha.text lowercase__ : List[str] = "https://www.amazon.in/" + item.ha.a["href"] lowercase__ : Optional[Any] = item.find("span" , attrs={"class": "a-offscreen"}).text try: lowercase__ : Tuple = item.find("span" , attrs={"class": "a-icon-alt"}).text except AttributeError: lowercase__ : List[Any] = "Not available" try: lowercase__ : int = ( "₹" + item.find( "span" , attrs={"class": "a-price a-text-price"}).text.split("₹")[1] ) except AttributeError: lowercase__ : Any = "" try: lowercase__ : Optional[Any] = float( ( ( float(product_mrp.strip("₹").replace("," , "")) - float(product_price.strip("₹").replace("," , "")) ) / float(product_mrp.strip("₹").replace("," , "")) ) * 100) except ValueError: lowercase__ : Dict = float("nan") except AttributeError: pass lowercase__ : Tuple = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] lowercase__ : str = " " lowercase__ : Dict = " " data_frame.index += 1 return data_frame if __name__ == "__main__": UpperCamelCase = '''headphones''' get_amazon_product_data(product).to_csv(f"Amazon Product Data for {product}.csv")
87
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowercase_ ( _lowerCamelCase : Any , _lowerCamelCase : List[str]=False): try: lowercase__ : Union[str, Any] = os.environ[key] except KeyError: # KEY isn't set, default to `default`. lowercase__ : int = default else: # KEY is set, convert it to True or False. try: lowercase__ : Optional[int] = strtobool(_lowerCamelCase) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'''If set, {key} must be yes or no.''') return _value UpperCamelCase = parse_flag_from_env('''RUN_SLOW''', default=False) UpperCamelCase = parse_flag_from_env('''RUN_REMOTE''', default=False) UpperCamelCase = parse_flag_from_env('''RUN_LOCAL''', default=True) UpperCamelCase = parse_flag_from_env('''RUN_PACKAGED''', default=True) # Compression UpperCamelCase = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='''test requires lz4''') UpperCamelCase = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='''test requires py7zr''') UpperCamelCase = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='''test requires zstandard''') # Audio UpperCamelCase = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('''soundfile''') is None or version.parse(importlib_metadata.version('''soundfile''')) < version.parse('''0.12.0'''), reason='''test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ''', ) # Beam UpperCamelCase = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('''0.3.2'''), reason='''test requires apache-beam and a compatible dill version''', ) # Dill-cloudpickle compatibility UpperCamelCase = pytest.mark.skipif( config.DILL_VERSION <= version.parse('''0.3.2'''), reason='''test requires dill>0.3.2 for cloudpickle compatibility''', ) # Windows UpperCamelCase = pytest.mark.skipif( sys.platform == '''win32''', reason='''test should not be run on Windows''', ) def lowercase_ ( _lowerCamelCase : int): try: import faiss # noqa except ImportError: lowercase__ : Optional[Any] = unittest.skip("test requires faiss")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : int): try: import regex # noqa except ImportError: lowercase__ : List[Any] = unittest.skip("test requires regex")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : int): try: import elasticsearch # noqa except ImportError: lowercase__ : Optional[int] = unittest.skip("test requires elasticsearch")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : Union[str, Any]): try: import sqlalchemy # noqa except ImportError: lowercase__ : Optional[int] = unittest.skip("test requires sqlalchemy")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : int): if not config.TORCH_AVAILABLE: lowercase__ : Tuple = unittest.skip("test requires PyTorch")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : Tuple): if not config.TF_AVAILABLE: lowercase__ : Any = unittest.skip("test requires TensorFlow")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : Dict): if not config.JAX_AVAILABLE: lowercase__ : List[str] = unittest.skip("test requires JAX")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : int): if not config.PIL_AVAILABLE: lowercase__ : Dict = unittest.skip("test requires Pillow")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : Tuple): try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers")(_lowerCamelCase) else: return test_case def lowercase_ ( _lowerCamelCase : Optional[Any]): try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken")(_lowerCamelCase) else: return test_case def lowercase_ ( _lowerCamelCase : Dict): try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy")(_lowerCamelCase) else: return test_case def lowercase_ ( _lowerCamelCase : Optional[int]): def _require_spacy_model(_lowerCamelCase : Optional[int]): try: import spacy # noqa F401 spacy.load(_lowerCamelCase) except ImportError: return unittest.skip("test requires spacy")(_lowerCamelCase) except OSError: return unittest.skip("test requires spacy model '{}'".format(_lowerCamelCase))(_lowerCamelCase) else: return test_case return _require_spacy_model def lowercase_ ( _lowerCamelCase : Dict): try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark")(_lowerCamelCase) else: return test_case def lowercase_ ( _lowerCamelCase : List[str]): try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark")(_lowerCamelCase) else: return test_case def lowercase_ ( _lowerCamelCase : Dict): if not _run_slow_tests or _run_slow_tests == 0: lowercase__ : Tuple = unittest.skip("test is slow")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : int): if not _run_local_tests or _run_local_tests == 0: lowercase__ : str = unittest.skip("test is local")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : Optional[int]): if not _run_packaged_tests or _run_packaged_tests == 0: lowercase__ : List[Any] = unittest.skip("test is packaged")(_lowerCamelCase) return test_case def lowercase_ ( _lowerCamelCase : Tuple): if not _run_remote_tests or _run_remote_tests == 0: lowercase__ : Union[str, Any] = unittest.skip("test requires remote")(_lowerCamelCase) return test_case def lowercase_ ( *_lowerCamelCase : str): def decorate(cls : str): for name, fn in cls.__dict__.items(): if callable(_lowerCamelCase) and name.startswith("test"): for decorator in decorators: lowercase__ : Optional[int] = decorator(_lowerCamelCase) setattr(cls , _lowerCamelCase , _lowerCamelCase) return cls return decorate class snake_case_ ( __A ): pass class snake_case_ ( __A ): __A : List[Any] = 0 __A : str = 1 __A : int = 2 @contextmanager def lowercase_ ( _lowerCamelCase : List[str]=OfflineSimulationMode.CONNECTION_FAILS , _lowerCamelCase : int=1E-16): lowercase__ : int = requests.Session().request def timeout_request(_lowerCamelCase : str , _lowerCamelCase : Dict , _lowerCamelCase : Dict , **_lowerCamelCase : str): # Change the url to an invalid url so that the connection hangs lowercase__ : Any = "https://10.255.255.1" if kwargs.get("timeout") is None: raise RequestWouldHangIndefinitelyError( f'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''') lowercase__ : Dict = timeout try: return online_request(_lowerCamelCase , _lowerCamelCase , **_lowerCamelCase) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier lowercase__ : Dict = url lowercase__ : Union[str, Any] = e.args[0] lowercase__ : Optional[Any] = (max_retry_error.args[0].replace("10.255.255.1" , f'''OfflineMock[{url}]'''),) lowercase__ : int = (max_retry_error,) raise def raise_connection_error(_lowerCamelCase : Union[str, Any] , _lowerCamelCase : Optional[Any] , **_lowerCamelCase : Tuple): raise requests.ConnectionError("Offline mode is enabled." , request=_lowerCamelCase) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send" , _lowerCamelCase): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request" , _lowerCamelCase): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE" , _lowerCamelCase): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum.") @contextmanager def lowercase_ ( *_lowerCamelCase : str , **_lowerCamelCase : Tuple): lowercase__ : Dict = str(Path().resolve()) with tempfile.TemporaryDirectory(*_lowerCamelCase , **_lowerCamelCase) as tmp_dir: try: os.chdir(_lowerCamelCase) yield finally: os.chdir(_lowerCamelCase) @contextmanager def lowercase_ ( ): import gc gc.collect() lowercase__ : Union[str, Any] = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowercase_ ( ): import gc gc.collect() lowercase__ : int = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowercase_ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[Any]): return deepcopy(_lowerCamelCase).integers(0 , 100 , 10).tolist() == deepcopy(_lowerCamelCase).integers(0 , 100 , 10).tolist() def lowercase_ ( _lowerCamelCase : str): import decorator from requests.exceptions import HTTPError def _wrapper(_lowerCamelCase : str , *_lowerCamelCase : Dict , **_lowerCamelCase : Dict): try: return func(*_lowerCamelCase , **_lowerCamelCase) except HTTPError as err: if str(_lowerCamelCase).startswith("500") or str(_lowerCamelCase).startswith("502"): pytest.xfail(str(_lowerCamelCase)) raise err return decorator.decorator(_wrapper , _lowerCamelCase) class snake_case_ : def __init__( self : int , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : List[str] ) -> List[str]: lowercase__ : Tuple = returncode lowercase__ : int = stdout lowercase__ : Union[str, Any] = stderr async def lowercase_ ( _lowerCamelCase : List[str] , _lowerCamelCase : Dict): while True: lowercase__ : Optional[int] = await stream.readline() if line: callback(_lowerCamelCase) else: break async def lowercase_ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any]=None , _lowerCamelCase : Optional[Any]=None , _lowerCamelCase : int=None , _lowerCamelCase : Optional[Any]=False , _lowerCamelCase : Tuple=False): if echo: print("\nRunning: " , " ".join(_lowerCamelCase)) lowercase__ : Optional[int] = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=_lowerCamelCase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=_lowerCamelCase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) lowercase__ : str = [] lowercase__ : List[str] = [] def tee(_lowerCamelCase : str , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Optional[int]=""): lowercase__ : Optional[int] = line.decode("utf-8").rstrip() sink.append(_lowerCamelCase) if not quiet: print(_lowerCamelCase , _lowerCamelCase , file=_lowerCamelCase) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda _lowerCamelCase: tee(_lowerCamelCase , _lowerCamelCase , sys.stdout , label="stdout:")), _read_stream(p.stderr , lambda _lowerCamelCase: tee(_lowerCamelCase , _lowerCamelCase , sys.stderr , label="stderr:")), ] , timeout=_lowerCamelCase , ) return _RunOutput(await p.wait() , _lowerCamelCase , _lowerCamelCase) def lowercase_ ( _lowerCamelCase : Optional[int] , _lowerCamelCase : List[str]=None , _lowerCamelCase : Dict=None , _lowerCamelCase : int=180 , _lowerCamelCase : Union[str, Any]=False , _lowerCamelCase : Optional[Any]=True): lowercase__ : Any = asyncio.get_event_loop() lowercase__ : Tuple = loop.run_until_complete( _stream_subprocess(_lowerCamelCase , env=_lowerCamelCase , stdin=_lowerCamelCase , timeout=_lowerCamelCase , quiet=_lowerCamelCase , echo=_lowerCamelCase)) lowercase__ : int = " ".join(_lowerCamelCase) if result.returncode > 0: lowercase__ : Any = "\n".join(result.stderr) raise RuntimeError( f'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' f'''The combined stderr from workers follows:\n{stderr}''') # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f'''\'{cmd_str}\' produced no output.''') return result def lowercase_ ( ): lowercase__ : List[str] = os.environ.get("PYTEST_XDIST_WORKER" , "gw0") lowercase__ : str = re.sub(R"^gw" , "" , _lowerCamelCase , 0 , re.M) return int(_lowerCamelCase) def lowercase_ ( ): lowercase__ : Union[str, Any] = 2_9500 lowercase__ : Optional[int] = pytest_xdist_worker_id() return port + uniq_delta
87
1
"""simple docstring""" import comet # From: unbabel-comet import torch import datasets a : Optional[Any] = datasets.logging.get_logger(__name__) a : Union[str, Any] = '''\ @inproceedings{rei-EtAl:2020:WMT, author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon}, title = {Unbabel\'s Participation in the WMT20 Metrics Shared Task}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, month = {November}, year = {2020}, address = {Online}, publisher = {Association for Computational Linguistics}, pages = {909--918}, } @inproceedings{rei-etal-2020-comet, title = "{COMET}: A Neural Framework for {MT} Evaluation", author = "Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.emnlp-main.213", pages = "2685--2702", } ''' a : Tuple = '''\ Crosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA\'s or MQM). With the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition. See the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information. ''' a : Dict = ''' COMET score. Args: `sources` (list of str): Source sentences `predictions` (list of str): candidate translations `references` (list of str): reference translations `cuda` (bool): If set to True, runs COMET using GPU `show_progress` (bool): Shows progress `model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None. Returns: `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`. `scores`: List of scores. Examples: >>> comet_metric = datasets.load_metric(\'comet\') >>> # comet_metric = load_metric(\'comet\', \'wmt20-comet-da\') # you can also choose which model to use >>> source = ["Dem Feuer konnte Einhalt geboten werden", "Schulen und Kindergärten wurden eröffnet."] >>> hypothesis = ["The fire could be stopped", "Schools and kindergartens were open"] >>> reference = ["They were able to control the fire.", "Schools and kindergartens opened"] >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source) >>> print([round(v, 2) for v in results["scores"]]) [0.19, 0.92] ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def __a ( self ) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://unbabel.github.io/COMET/html/index.html" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "sources": datasets.Value("string" , id="sequence" ), "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/Unbabel/COMET"] , reference_urls=[ "https://github.com/Unbabel/COMET", "https://www.aclweb.org/anthology/2020.emnlp-main.213/", "http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6", ] , ) def __a ( self , lowerCAmelCase__ ) -> Optional[int]: if self.config_name == "default": a : List[str] = comet.load_from_checkpoint(comet.download_model("wmt20-comet-da" ) ) else: a : Optional[Any] = comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=False ) -> List[str]: if gpus is None: a : Optional[Any] = 1 if torch.cuda.is_available() else 0 a : Optional[Any] = {"src": sources, "mt": predictions, "ref": references} a : Optional[Any] = [dict(zip(lowerCAmelCase__ , lowerCAmelCase__ ) ) for t in zip(*data.values() )] a, a : str = self.scorer.predict(lowerCAmelCase__ , gpus=lowerCAmelCase__ , progress_bar=lowerCAmelCase__ ) return {"mean_score": mean_score, "scores": scores}
79
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=a__ ) class __UpperCamelCase ( a__ ): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization lowerCamelCase : str =field(default="""summarization""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowerCamelCase : ClassVar[Features] =Features({"""text""": Value("""string""" )} ) lowerCamelCase : ClassVar[Features] =Features({"""summary""": Value("""string""" )} ) lowerCamelCase : str ="text" lowerCamelCase : str ="summary" @property def __a ( self ) -> Dict[str, str]: return {self.text_column: "text", self.summary_column: "summary"}
79
1
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Tuple: lowerCamelCase__ : Tuple = SwinConfig(image_size=192 ) if "base" in model_name: lowerCamelCase__ : Union[str, Any] = 6 lowerCamelCase__ : Dict = 128 lowerCamelCase__ : Any = (2, 2, 18, 2) lowerCamelCase__ : Tuple = (4, 8, 16, 32) elif "large" in model_name: lowerCamelCase__ : Any = 12 lowerCamelCase__ : Tuple = 192 lowerCamelCase__ : List[str] = (2, 2, 18, 2) lowerCamelCase__ : Union[str, Any] = (6, 12, 24, 48) else: raise ValueError("""Model not supported, only supports base and large variants""" ) lowerCamelCase__ : str = window_size lowerCamelCase__ : Dict = embed_dim lowerCamelCase__ : Tuple = depths lowerCamelCase__ : str = num_heads return config def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: if "encoder.mask_token" in name: lowerCamelCase__ : Any = name.replace("""encoder.mask_token""" , """embeddings.mask_token""" ) if "encoder.patch_embed.proj" in name: lowerCamelCase__ : List[Any] = name.replace("""encoder.patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "encoder.patch_embed.norm" in name: lowerCamelCase__ : str = name.replace("""encoder.patch_embed.norm""" , """embeddings.norm""" ) if "attn.proj" in name: lowerCamelCase__ : List[Any] = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: lowerCamelCase__ : List[Any] = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: lowerCamelCase__ : Tuple = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: lowerCamelCase__ : int = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: lowerCamelCase__ : Optional[Any] = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: lowerCamelCase__ : Optional[Any] = name.replace("""mlp.fc2""" , """output.dense""" ) if name == "encoder.norm.weight": lowerCamelCase__ : Optional[int] = """layernorm.weight""" if name == "encoder.norm.bias": lowerCamelCase__ : str = """layernorm.bias""" if "decoder" in name: pass else: lowerCamelCase__ : Tuple = """swin.""" + name return name def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase ) -> Optional[int]: for key in orig_state_dict.copy().keys(): lowerCamelCase__ : Tuple = orig_state_dict.pop(UpperCamelCase ) if "attn_mask" in key: pass elif "qkv" in key: lowerCamelCase__ : Union[str, Any] = key.split(""".""" ) lowerCamelCase__ : Optional[Any] = int(key_split[2] ) lowerCamelCase__ : Any = int(key_split[4] ) lowerCamelCase__ : List[str] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: lowerCamelCase__ : Union[str, Any] = val[:dim, :] lowerCamelCase__ : Tuple = val[ dim : dim * 2, : ] lowerCamelCase__ : Union[str, Any] = val[-dim:, :] else: lowerCamelCase__ : Union[str, Any] = val[ :dim ] lowerCamelCase__ : int = val[ dim : dim * 2 ] lowerCamelCase__ : List[str] = val[ -dim: ] else: lowerCamelCase__ : Union[str, Any] = val return orig_state_dict def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Optional[Any]: lowerCamelCase__ : Optional[int] = torch.load(UpperCamelCase , map_location="""cpu""" )["""model"""] lowerCamelCase__ : List[str] = get_swin_config(UpperCamelCase ) lowerCamelCase__ : str = SwinForMaskedImageModeling(UpperCamelCase ) model.eval() lowerCamelCase__ : List[Any] = convert_state_dict(UpperCamelCase , UpperCamelCase ) model.load_state_dict(UpperCamelCase ) lowerCamelCase__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase__ : Dict = ViTImageProcessor(size={"""height""": 192, """width""": 192} ) lowerCamelCase__ : Dict = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) lowerCamelCase__ : Union[str, Any] = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) with torch.no_grad(): lowerCamelCase__ : List[Any] = model(**UpperCamelCase ).logits print(outputs.keys() ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase ) if push_to_hub: print(f'''Pushing model and image processor for {model_name} to hub''' ) model.push_to_hub(f'''microsoft/{model_name}''' ) image_processor.push_to_hub(f'''microsoft/{model_name}''' ) if __name__ == "__main__": _A : Union[str, Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''swin-base-simmim-window6-192''', type=str, choices=['''swin-base-simmim-window6-192''', '''swin-large-simmim-window12-192'''], help='''Name of the Swin SimMIM model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default='''/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth''', type=str, help='''Path to the original PyTorch checkpoint (.pth file).''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _A : int =parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
41
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _A : Union[str, Any] ={ '''configuration_swiftformer''': [ '''SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwiftFormerConfig''', '''SwiftFormerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[Any] =[ '''SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SwiftFormerForImageClassification''', '''SwiftFormerModel''', '''SwiftFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swiftformer import ( SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, SwiftFormerConfig, SwiftFormerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swiftformer import ( SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, SwiftFormerForImageClassification, SwiftFormerModel, SwiftFormerPreTrainedModel, ) else: import sys _A : Tuple =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
1
def _a ( SCREAMING_SNAKE_CASE_ : List[Any] ): __lowerCAmelCase , __lowerCAmelCase = [], [] while len(SCREAMING_SNAKE_CASE_ ) > 1: __lowerCAmelCase , __lowerCAmelCase = min(SCREAMING_SNAKE_CASE_ ), max(SCREAMING_SNAKE_CASE_ ) start.append(SCREAMING_SNAKE_CASE_ ) end.append(SCREAMING_SNAKE_CASE_ ) collection.remove(SCREAMING_SNAKE_CASE_ ) collection.remove(SCREAMING_SNAKE_CASE_ ) end.reverse() return start + collection + end if __name__ == "__main__": UpperCamelCase__ = input("""Enter numbers separated by a comma:\n""").strip() UpperCamelCase__ = [int(item) for item in user_input.split(""",""")] print(*merge_sort(unsorted), sep=""",""")
102
import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class a__ : def __init__( self , _A , _A=1_4 , _A=7 , _A=True , _A=True , _A=False , _A=True , _A=9_9 , _A=3_2 , _A=4 , _A=4 , _A=4 , _A=3_7 , _A="gelu" , _A=0.1 , _A=0.1 , _A=5_1_2 , _A=0.02 , ): """simple docstring""" __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_input_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = rotary_dim __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = initializer_range __lowerCAmelCase = None __lowerCAmelCase = vocab_size - 1 __lowerCAmelCase = vocab_size - 1 __lowerCAmelCase = vocab_size - 1 def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase = None if self.use_input_mask: __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=_A , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = config_and_inputs __lowerCAmelCase = {"input_ids": input_ids, "attention_mask": attention_mask} return config, inputs_dict def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = 2_0 __lowerCAmelCase = model_class_name(_A ) __lowerCAmelCase = model.init_cache(input_ids.shape[0] , _A ) __lowerCAmelCase = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype="i4" ) __lowerCAmelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __lowerCAmelCase = model( input_ids[:, :-1] , attention_mask=_A , past_key_values=_A , position_ids=_A , ) __lowerCAmelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="i4" ) __lowerCAmelCase = model( input_ids[:, -1:] , attention_mask=_A , past_key_values=outputs_cache.past_key_values , position_ids=_A , ) __lowerCAmelCase = model(_A ) __lowerCAmelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"""Max diff is {diff}""" ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = 2_0 __lowerCAmelCase = model_class_name(_A ) __lowerCAmelCase = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )] , axis=-1 , ) __lowerCAmelCase = model.init_cache(input_ids.shape[0] , _A ) __lowerCAmelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __lowerCAmelCase = model( input_ids[:, :-1] , attention_mask=_A , past_key_values=_A , position_ids=_A , ) __lowerCAmelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="i4" ) __lowerCAmelCase = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=_A , position_ids=_A , ) __lowerCAmelCase = model(_A , attention_mask=_A ) __lowerCAmelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"""Max diff is {diff}""" ) @require_flax class a__ ( snake_case__ , snake_case__ , unittest.TestCase ): _a : str = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () _a : Any = (FlaxGPTJForCausalLM,) if is_flax_available() else () def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = FlaxGPTJModelTester(self ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_class_name in self.all_model_classes: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(_A , _A , _A , _A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_class_name in self.all_model_classes: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( _A , _A , _A , _A ) @tooslow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = GPTaTokenizer.from_pretrained("gpt2" , pad_token="<|endoftext|>" , padding_side="left" ) __lowerCAmelCase = tokenizer(["Hello this is a long string", "Hey"] , return_tensors="np" , padding=_A , truncation=_A ) __lowerCAmelCase = FlaxGPTJForCausalLM.from_pretrained("EleutherAI/gpt-j-6B" ) __lowerCAmelCase = False __lowerCAmelCase = model.config.eos_token_id __lowerCAmelCase = jax.jit(model.generate ) __lowerCAmelCase = jit_generate( inputs["input_ids"] , attention_mask=inputs["attention_mask"] , pad_token_id=tokenizer.pad_token_id ).sequences __lowerCAmelCase = tokenizer.batch_decode(_A , skip_special_tokens=_A ) __lowerCAmelCase = [ "Hello this is a long string of text.\n\nI'm trying to get the text of the", "Hey, I'm a little late to the party. I'm going to", ] self.assertListEqual(_A , _A ) @is_pt_flax_cross_test def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __lowerCAmelCase = self._prepare_for_class(_A , _A ) __lowerCAmelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __lowerCAmelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning __lowerCAmelCase = getattr(_A , _A ) __lowerCAmelCase , __lowerCAmelCase = pt_inputs["input_ids"].shape __lowerCAmelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_A ): __lowerCAmelCase = 0 __lowerCAmelCase = 1 __lowerCAmelCase = 0 __lowerCAmelCase = 1 __lowerCAmelCase = pt_model_class(_A ).eval() __lowerCAmelCase = model_class(_A , dtype=jnp.floataa ) __lowerCAmelCase = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _A ) __lowerCAmelCase = fx_state with torch.no_grad(): __lowerCAmelCase = pt_model(**_A ).to_tuple() __lowerCAmelCase = fx_model(**_A ).to_tuple() self.assertEqual(len(_A ) , len(_A ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(_A , _A ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_A ) __lowerCAmelCase = model_class.from_pretrained(_A , from_pt=_A ) __lowerCAmelCase = fx_model_loaded(**_A ).to_tuple() self.assertEqual( len(_A ) , len(_A ) , "Output lengths differ between Flax and PyTorch" ) for fx_output_loaded, pt_output in zip(_A , _A ): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @is_pt_flax_cross_test def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __lowerCAmelCase = self._prepare_for_class(_A , _A ) __lowerCAmelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __lowerCAmelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning __lowerCAmelCase = getattr(_A , _A ) __lowerCAmelCase = pt_model_class(_A ).eval() __lowerCAmelCase = model_class(_A , dtype=jnp.floataa ) __lowerCAmelCase = load_flax_weights_in_pytorch_model(_A , fx_model.params ) __lowerCAmelCase , __lowerCAmelCase = pt_inputs["input_ids"].shape __lowerCAmelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_A ): __lowerCAmelCase = 0 __lowerCAmelCase = 1 __lowerCAmelCase = 0 __lowerCAmelCase = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): __lowerCAmelCase = pt_model(**_A ).to_tuple() __lowerCAmelCase = fx_model(**_A ).to_tuple() self.assertEqual(len(_A ) , len(_A ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(_A , _A ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_A ) __lowerCAmelCase = pt_model_class.from_pretrained(_A , from_flax=_A ) with torch.no_grad(): __lowerCAmelCase = pt_model_loaded(**_A ).to_tuple() self.assertEqual( len(_A ) , len(_A ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(_A , _A ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @tooslow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_class_name in self.all_model_classes: __lowerCAmelCase = model_class_name.from_pretrained("EleutherAI/gpt-j-6B" ) __lowerCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(_A )
102
1
"""simple docstring""" def _snake_case ( UpperCamelCase : Tuple , UpperCamelCase : Optional[int] ): UpperCAmelCase : Tuple = 0 UpperCAmelCase : str = len(UpperCamelCase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None UpperCAmelCase : List[Any] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(UpperCamelCase ): return None UpperCAmelCase : Tuple = sorted_collection[point] if current_item == item: return point else: if point < left: UpperCAmelCase : List[str] = left UpperCAmelCase : Union[str, Any] = point elif point > right: UpperCAmelCase : Union[str, Any] = right UpperCAmelCase : List[str] = point else: if item < current_item: UpperCAmelCase : Tuple = point - 1 else: UpperCAmelCase : Any = point + 1 return None def _snake_case ( UpperCamelCase : int , UpperCamelCase : Tuple , UpperCamelCase : int , UpperCamelCase : Dict ): # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None UpperCAmelCase : Optional[int] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(UpperCamelCase ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) elif point > right: return interpolation_search_by_recursion(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( UpperCamelCase , UpperCamelCase , UpperCamelCase , point - 1 ) else: return interpolation_search_by_recursion( UpperCamelCase , UpperCamelCase , point + 1 , UpperCamelCase ) def _snake_case ( UpperCamelCase : Optional[Any] ): if collection != sorted(UpperCamelCase ): raise ValueError("""Collection must be ascending sorted""" ) return True if __name__ == "__main__": import sys A: Any = 0 if debug == 1: A: Optional[int] = [1_0, 3_0, 4_0, 4_5, 5_0, 6_6, 7_7, 9_3] try: __assert_sorted(collection) except ValueError: sys.exit("Sequence must be ascending sorted to apply interpolation search") A: str = 6_7 A: List[Any] = interpolation_search(collection, target) if result is not None: print(f"""{target} found at positions: {result}""") else: print("Not found")
109
"""simple docstring""" from collections.abc import Callable import numpy as np def _snake_case ( UpperCamelCase : Callable , UpperCamelCase : float , UpperCamelCase : float , UpperCamelCase : float , UpperCamelCase : float ): UpperCAmelCase : Any = int(np.ceil((x_end - xa) / step_size ) ) UpperCAmelCase : Optional[Any] = np.zeros((n + 1,) ) UpperCAmelCase : Optional[int] = ya UpperCAmelCase : int = xa for k in range(UpperCamelCase ): UpperCAmelCase : Optional[int] = y[k] + step_size * ode_func(UpperCamelCase , y[k] ) UpperCAmelCase : Optional[int] = y[k] + ( (step_size / 2) * (ode_func(UpperCamelCase , y[k] ) + ode_func(x + step_size , UpperCamelCase )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
109
1
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class a ( nn.Module ): _snake_case : int _snake_case : jnp.dtype = jnp.floataa def lowerCAmelCase_ ( self : Dict ): _UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Any , __lowerCAmelCase : Optional[Any] ): _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = hidden_states.shape _UpperCAmelCase = jax.image.resize( __lowerCAmelCase , shape=(batch, height * 2, width * 2, channels) , method="""nearest""" , ) _UpperCAmelCase = self.conv(__lowerCAmelCase ) return hidden_states class a ( nn.Module ): _snake_case : int _snake_case : jnp.dtype = jnp.floataa def lowerCAmelCase_ ( self : Dict ): _UpperCAmelCase = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : int , __lowerCAmelCase : List[Any] ): # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) _UpperCAmelCase = self.conv(__lowerCAmelCase ) return hidden_states class a ( nn.Module ): _snake_case : int _snake_case : int = None _snake_case : float = 0.0 _snake_case : bool = None _snake_case : jnp.dtype = jnp.floataa def lowerCAmelCase_ ( self : Union[str, Any] ): _UpperCAmelCase = self.in_channels if self.out_channels is None else self.out_channels _UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) _UpperCAmelCase = nn.Conv( __lowerCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) _UpperCAmelCase = nn.Dense(__lowerCAmelCase , dtype=self.dtype ) _UpperCAmelCase = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) _UpperCAmelCase = nn.Dropout(self.dropout_prob ) _UpperCAmelCase = nn.Conv( __lowerCAmelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) _UpperCAmelCase = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut _UpperCAmelCase = None if use_nin_shortcut: _UpperCAmelCase = nn.Conv( __lowerCAmelCase , kernel_size=(1, 1) , strides=(1, 1) , padding="""VALID""" , dtype=self.dtype , ) def __call__( self : Optional[int] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Union[str, Any]=True ): _UpperCAmelCase = hidden_states _UpperCAmelCase = self.norma(__lowerCAmelCase ) _UpperCAmelCase = nn.swish(__lowerCAmelCase ) _UpperCAmelCase = self.conva(__lowerCAmelCase ) _UpperCAmelCase = self.time_emb_proj(nn.swish(__lowerCAmelCase ) ) _UpperCAmelCase = jnp.expand_dims(jnp.expand_dims(__lowerCAmelCase , 1 ) , 1 ) _UpperCAmelCase = hidden_states + temb _UpperCAmelCase = self.norma(__lowerCAmelCase ) _UpperCAmelCase = nn.swish(__lowerCAmelCase ) _UpperCAmelCase = self.dropout(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = self.conva(__lowerCAmelCase ) if self.conv_shortcut is not None: _UpperCAmelCase = self.conv_shortcut(__lowerCAmelCase ) return hidden_states + residual
359
"""simple docstring""" import os import pytest from attr import dataclass UpperCAmelCase__ = """us-east-1""" # defaults region @dataclass class a : _snake_case : str _snake_case : Tuple = 'arn:aws:iam::558105141721:role/sagemaker_execution_role' _snake_case : List[Any] = { 'task_name': 'mnli', 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'do_train': True, 'do_eval': True, 'do_predict': True, 'output_dir': '/opt/ml/model', 'overwrite_output_dir': True, 'max_steps': 5_00, 'save_steps': 55_00, } _snake_case : Optional[Any] = {**hyperparameters, 'max_steps': 10_00} @property def lowerCAmelCase_ ( self : Optional[Any] ): if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def lowerCAmelCase_ ( self : Dict ): return f'''{self.framework}-transfromers-test''' @property def lowerCAmelCase_ ( self : Union[str, Any] ): return f'''./tests/sagemaker/scripts/{self.framework}''' @property def lowerCAmelCase_ ( self : Dict ): if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope="""class""" ) def __UpperCAmelCase ( lowercase ): """simple docstring""" _UpperCAmelCase = SageMakerTestEnvironment(framework=request.cls.framework )
30
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __A ={ '''configuration_blenderbot_small''': [ '''BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlenderbotSmallConfig''', '''BlenderbotSmallOnnxConfig''', ], '''tokenization_blenderbot_small''': ['''BlenderbotSmallTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =['''BlenderbotSmallTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =[ '''BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlenderbotSmallForCausalLM''', '''BlenderbotSmallForConditionalGeneration''', '''BlenderbotSmallModel''', '''BlenderbotSmallPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =[ '''TFBlenderbotSmallForConditionalGeneration''', '''TFBlenderbotSmallModel''', '''TFBlenderbotSmallPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =[ '''FlaxBlenderbotSmallForConditionalGeneration''', '''FlaxBlenderbotSmallModel''', '''FlaxBlenderbotSmallPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotSmallConfig, BlenderbotSmallOnnxConfig, ) from .tokenization_blenderbot_small import BlenderbotSmallTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_small_fast import BlenderbotSmallTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotSmallForCausalLM, BlenderbotSmallForConditionalGeneration, BlenderbotSmallModel, BlenderbotSmallPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot_small import ( TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel, TFBlenderbotSmallPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, FlaxBlenderbotSmallPreTrainedModel, ) else: import sys __A =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
19
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing the experiment tracking capability, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __A =1_6 __A =3_2 def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ = 1_6 ): lowerCamelCase_ = AutoTokenizer.from_pretrained("bert-base-cased" ) lowerCamelCase_ = load_dataset("glue" , "mrpc" ) def tokenize_function(lowerCamelCase__ ): # max_length=None => use the model max length (it's actually the default) lowerCamelCase_ = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=lowerCamelCase__ , max_length=lowerCamelCase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowerCamelCase_ = datasets.map( lowerCamelCase__ , batched=lowerCamelCase__ , remove_columns=["idx", "sentence1", "sentence2"] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowerCamelCase_ = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(lowerCamelCase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. lowerCamelCase_ = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowerCamelCase_ = 1_6 elif accelerator.mixed_precision != "no": lowerCamelCase_ = 8 else: lowerCamelCase_ = None return tokenizer.pad( lowerCamelCase__ , padding="longest" , max_length=lowerCamelCase__ , pad_to_multiple_of=lowerCamelCase__ , return_tensors="pt" , ) # Instantiate dataloaders. lowerCamelCase_ = DataLoader( tokenized_datasets["train"] , shuffle=lowerCamelCase__ , collate_fn=lowerCamelCase__ , batch_size=lowerCamelCase__ ) lowerCamelCase_ = DataLoader( tokenized_datasets["validation"] , shuffle=lowerCamelCase__ , collate_fn=lowerCamelCase__ , batch_size=lowerCamelCase__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders __A =mocked_dataloaders # noqa: F811 def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): # For testing only if os.environ.get("TESTING_MOCKED_DATALOADERS" , lowerCamelCase__ ) == "1": lowerCamelCase_ = 2 # Initialize Accelerator # New Code # # We pass in "all" to `log_with` to grab all available trackers in the environment # Note: If using a custom `Tracker` class, should be passed in here such as: # >>> log_with = ["all", MyCustomTrackerClassInstance()] if args.with_tracking: lowerCamelCase_ = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with="all" , project_dir=args.project_dir ) else: lowerCamelCase_ = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowerCamelCase_ = config["lr"] lowerCamelCase_ = int(config["num_epochs"] ) lowerCamelCase_ = int(config["seed"] ) lowerCamelCase_ = int(config["batch_size"] ) set_seed(lowerCamelCase__ ) lowerCamelCase_ , lowerCamelCase_ = get_dataloaders(lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ = evaluate.load("glue" , "mrpc" ) # If the batch size is too big we use gradient accumulation lowerCamelCase_ = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: lowerCamelCase_ = batch_size // MAX_GPU_BATCH_SIZE lowerCamelCase_ = MAX_GPU_BATCH_SIZE # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowerCamelCase_ = AutoModelForSequenceClassification.from_pretrained("bert-base-cased" , return_dict=lowerCamelCase__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowerCamelCase_ = model.to(accelerator.device ) # Instantiate optimizer lowerCamelCase_ = AdamW(params=model.parameters() , lr=lowerCamelCase__ ) # Instantiate scheduler lowerCamelCase_ = get_linear_schedule_with_warmup( optimizer=lowerCamelCase__ , num_warmup_steps=1_0_0 , num_training_steps=(len(lowerCamelCase__ ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = accelerator.prepare( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # New Code # # We need to initialize the trackers we use. Overall configurations can also be stored if args.with_tracking: lowerCamelCase_ = os.path.split(lowerCamelCase__ )[-1].split("." )[0] accelerator.init_trackers(lowerCamelCase__ , lowerCamelCase__ ) # Now we train the model for epoch in range(lowerCamelCase__ ): model.train() # New Code # # For our tracking example, we will log the total loss of each epoch if args.with_tracking: lowerCamelCase_ = 0 for step, batch in enumerate(lowerCamelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowerCamelCase_ = model(**lowerCamelCase__ ) lowerCamelCase_ = outputs.loss # New Code # if args.with_tracking: total_loss += loss.detach().float() lowerCamelCase_ = loss / gradient_accumulation_steps accelerator.backward(lowerCamelCase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(lowerCamelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True` (the default). batch.to(accelerator.device ) with torch.no_grad(): lowerCamelCase_ = model(**lowerCamelCase__ ) lowerCamelCase_ = outputs.logits.argmax(dim=-1 ) lowerCamelCase_ , lowerCamelCase_ = accelerator.gather_for_metrics((predictions, batch["labels"]) ) metric.add_batch( predictions=lowerCamelCase__ , references=lowerCamelCase__ , ) lowerCamelCase_ = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}:' , lowerCamelCase__ ) # New Code # # To actually log, we call `Accelerator.log` # The values passed can be of `str`, `int`, `float` or `dict` of `str` to `float`/`int` if args.with_tracking: accelerator.log( { "accuracy": eval_metric["accuracy"], "f1": eval_metric["f1"], "train_loss": total_loss.item() / len(lowerCamelCase__ ), "epoch": epoch, } , step=lowerCamelCase__ , ) # New Code # # When a run is finished, you should call `accelerator.end_training()` # to close all of the open trackers if args.with_tracking: accelerator.end_training() def lowerCamelCase_ ( ): lowerCamelCase_ = argparse.ArgumentParser(description="Simple example of training script." ) parser.add_argument( "--mixed_precision" , type=lowerCamelCase__ , default=lowerCamelCase__ , choices=["no", "fp16", "bf16", "fp8"] , help="Whether to use mixed precision. Choose" "between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10." "and an Nvidia Ampere GPU." , ) parser.add_argument("--cpu" , action="store_true" , help="If passed, will train on the CPU." ) parser.add_argument( "--with_tracking" , action="store_true" , help="Whether to load in all available experiment trackers from the environment and use them for logging." , ) parser.add_argument( "--project_dir" , type=lowerCamelCase__ , default="logs" , help="Location on where to store experiment tracking logs` and relevent project information" , ) lowerCamelCase_ = parser.parse_args() lowerCamelCase_ = {"lr": 2e-5, "num_epochs": 3, "seed": 4_2, "batch_size": 1_6} training_function(lowerCamelCase__ , lowerCamelCase__ ) if __name__ == "__main__": main()
19
1
import operator def lowerCamelCase_ ( UpperCamelCase__ : list, UpperCamelCase__ : bool = False, UpperCamelCase__ : list | None = None ): '''simple docstring''' UpperCamelCase__ = operator.lt if reverse else operator.gt UpperCamelCase__ = solution or [] if not arr: return solution UpperCamelCase__ = [arr.pop(0 )] for i, item in enumerate(UpperCamelCase__ ): if _operator(UpperCamelCase__, sublist[-1] ): sublist.append(UpperCamelCase__ ) arr.pop(UpperCamelCase__ ) # merging sublist into solution list if not solution: solution.extend(UpperCamelCase__ ) else: while sublist: UpperCamelCase__ = sublist.pop(0 ) for i, xx in enumerate(UpperCamelCase__ ): if not _operator(UpperCamelCase__, UpperCamelCase__ ): solution.insert(UpperCamelCase__, UpperCamelCase__ ) break else: solution.append(UpperCamelCase__ ) strand_sort(UpperCamelCase__, UpperCamelCase__, UpperCamelCase__ ) return solution if __name__ == "__main__": assert strand_sort([4, 3, 5, 1, 2]) == [1, 2, 3, 4, 5] assert strand_sort([4, 3, 5, 1, 2], reverse=True) == [5, 4, 3, 2, 1]
35
from __future__ import annotations from typing import Any def lowerCamelCase_ ( UpperCamelCase__ : list ): '''simple docstring''' if not postfix_notation: return 0 UpperCamelCase__ = {'''+''', '''-''', '''*''', '''/'''} UpperCamelCase__ = [] for token in postfix_notation: if token in operations: UpperCamelCase__ , UpperCamelCase__ = stack.pop(), stack.pop() if token == "+": stack.append(a + b ) elif token == "-": stack.append(a - b ) elif token == "*": stack.append(a * b ) else: if a * b < 0 and a % b != 0: stack.append(a // b + 1 ) else: stack.append(a // b ) else: stack.append(int(UpperCamelCase__ ) ) return stack.pop() if __name__ == "__main__": import doctest doctest.testmod()
35
1
"""simple docstring""" UpperCAmelCase = [ """DownloadConfig""", """DownloadManager""", """DownloadMode""", """StreamingDownloadManager""", ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
256
"""simple docstring""" import argparse import os from pathlib import Path from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import PegasusConfig, PegasusForConditionalGeneration, PegasusTokenizer from transformers.models.pegasus.configuration_pegasus import DEFAULTS, task_specific_params UpperCAmelCase = [ # replace left string with right string to get the relevant state_dict key (identical state dict to bart) ["""memory_attention""", """encoder_attn"""], ["""attention""", """attn"""], ["""/""", """."""], [""".LayerNorm.gamma""", """_layer_norm.weight"""], [""".LayerNorm.beta""", """_layer_norm.bias"""], ["""r.layer_""", """r.layers."""], ["""output_proj""", """out_proj"""], ["""ffn.dense_1.""", """fc2."""], ["""ffn.dense.""", """fc1."""], ["""ffn_layer_norm""", """final_layer_norm"""], ["""kernel""", """weight"""], ["""encoder_layer_norm.""", """encoder.layer_norm."""], ["""decoder_layer_norm.""", """decoder.layer_norm."""], ["""embeddings.weights""", """shared.weight"""], ] def lowercase ( a__ : Optional[Any] ) -> Optional[int]: for pegasus_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(a__ , a__ ) return k def lowercase ( a__ : dict , a__ : dict ) -> PegasusForConditionalGeneration: _UpperCamelCase = DEFAULTS.copy() cfg_kwargs.update(a__ ) _UpperCamelCase = PegasusConfig(**a__ ) _UpperCamelCase = PegasusForConditionalGeneration(a__ ) _UpperCamelCase = torch_model.model.state_dict() _UpperCamelCase = {} for k, v in tf_weights.items(): _UpperCamelCase = rename_state_dict_key(a__ ) if new_k not in sd: raise ValueError(F'''could not find new key {new_k} in state dict. (converted from {k})''' ) if "dense" in k or "proj" in new_k: _UpperCamelCase = v.T _UpperCamelCase = torch.tensor(a__ , dtype=sd[new_k].dtype ) assert v.shape == sd[new_k].shape, F'''{new_k}, {k}, {v.shape}, {sd[new_k].shape}''' # make sure embedding.padding_idx is respected _UpperCamelCase = torch.zeros_like(mapping['''shared.weight'''][cfg.pad_token_id + 1] ) _UpperCamelCase = mapping['''shared.weight'''] _UpperCamelCase = mapping['''shared.weight'''] _UpperCamelCase = {k: torch.zeros_like(a__ ) for k, v in sd.items() if k.endswith('''bias''' ) and k not in mapping} mapping.update(**a__ ) _UpperCamelCase , _UpperCamelCase = torch_model.model.load_state_dict(a__ , strict=a__ ) _UpperCamelCase = [ k for k in missing if k not in ['''encoder.embed_positions.weight''', '''decoder.embed_positions.weight'''] ] assert unexpected_missing == [], F'''no matches found for the following torch keys {unexpected_missing}''' assert extra == [], F'''no matches found for the following tf keys {extra}''' return torch_model def lowercase ( a__ : List[str]="./ckpt/aeslc/model.ckpt-32000" ) -> Dict: _UpperCamelCase = tf.train.list_variables(a__ ) _UpperCamelCase = {} _UpperCamelCase = ['''Adafactor''', '''global_step'''] for name, shape in tqdm(a__ , desc='''converting tf checkpoint to dict''' ): _UpperCamelCase = any(pat in name for pat in ignore_name ) if skip_key: continue _UpperCamelCase = tf.train.load_variable(a__ , a__ ) _UpperCamelCase = array return tf_weights def lowercase ( a__ : str , a__ : str ) -> Optional[Any]: # save tokenizer first _UpperCamelCase = Path(a__ ).parent.name _UpperCamelCase = task_specific_params[F'''summarization_{dataset}''']['''max_position_embeddings'''] _UpperCamelCase = PegasusTokenizer.from_pretrained('''sshleifer/pegasus''' , model_max_length=a__ ) assert tok.model_max_length == desired_max_model_length tok.save_pretrained(a__ ) # convert model _UpperCamelCase = get_tf_weights_as_numpy(a__ ) _UpperCamelCase = task_specific_params[F'''summarization_{dataset}'''] if dataset == "large": _UpperCamelCase = task_specific_params _UpperCamelCase = convert_pegasus(a__ , a__ ) torch_model.save_pretrained(a__ ) _UpperCamelCase = torch_model.state_dict() sd.pop('''model.decoder.embed_positions.weight''' ) sd.pop('''model.encoder.embed_positions.weight''' ) torch.save(a__ , Path(a__ ) / '''pytorch_model.bin''' ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("""tf_ckpt_path""", type=str, help="""passed to tf.train.list_variables""") parser.add_argument("""save_dir""", default=None, type=str, help="""Path to the output PyTorch model.""") UpperCAmelCase = parser.parse_args() if args.save_dir is None: UpperCAmelCase = Path(args.tf_ckpt_path).parent.name UpperCAmelCase = os.path.join("""pegasus""", dataset) convert_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir)
256
1
"""simple docstring""" from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageResampling, get_image_size, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging _UpperCamelCase : str = logging.get_logger(__name__) class snake_case ( UpperCAmelCase ): __magic_name__ = ['''pixel_values'''] def __init__( self : Any , A : bool = True , A : int = 3_2 , A : List[str]=PILImageResampling.BILINEAR , A : bool = True , **A : Union[str, Any] , ): '''simple docstring''' a : str = do_resize a : int = do_rescale a : List[str] = size_divisor a : Optional[Any] = resample super().__init__(**A ) def lowerCamelCase__ ( self : Union[str, Any] , A : np.ndarray , A : int , A : List[Any] , A : Optional[ChannelDimension] = None , **A : Optional[Any] ): '''simple docstring''' a, a : Optional[int] = get_image_size(A ) # Rounds the height and width down to the closest multiple of size_divisor a : List[Any] = height // size_divisor * size_divisor a : Dict = width // size_divisor * size_divisor a : Any = resize(A , (new_h, new_w) , resample=A , data_format=A , **A ) return image def lowerCamelCase__ ( self : Union[str, Any] , A : np.ndarray , A : float , A : Optional[ChannelDimension] = None , **A : Optional[Any] ): '''simple docstring''' return rescale(image=A , scale=A , data_format=A , **A ) def lowerCamelCase__ ( self : Dict , A : Union["PIL.Image.Image", TensorType, List["PIL.Image.Image"], List[TensorType]] , A : Optional[bool] = None , A : Optional[int] = None , A : List[Any]=None , A : Optional[bool] = None , A : Optional[Union[TensorType, str]] = None , A : ChannelDimension = ChannelDimension.FIRST , **A : Optional[Any] , ): '''simple docstring''' a : Tuple = do_resize if do_resize is not None else self.do_resize a : List[str] = do_rescale if do_rescale is not None else self.do_rescale a : int = size_divisor if size_divisor is not None else self.size_divisor a : int = resample if resample is not None else self.resample if do_resize and size_divisor is None: raise ValueError('size_divisor is required for resizing' ) a : int = make_list_of_images(A ) if not valid_images(A ): raise ValueError('Invalid image(s)' ) # All transformations expect numpy arrays. a : List[Any] = [to_numpy_array(A ) for img in images] if do_resize: a : Any = [self.resize(A , size_divisor=A , resample=A ) for image in images] if do_rescale: a : Tuple = [self.rescale(A , scale=1 / 2_5_5 ) for image in images] a : str = [to_channel_dimension_format(A , A ) for image in images] a : Any = {'pixel_values': images} return BatchFeature(data=A , tensor_type=A )
186
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _UpperCamelCase : int = logging.get_logger(__name__) class snake_case ( UpperCAmelCase ): __magic_name__ = ['''input_features''', '''attention_mask'''] def __init__( self : Optional[int] , A : Optional[Any]=8_0 , A : str=1_6_0_0_0 , A : List[str]=0.0 , A : Any=1_0 , A : Union[str, Any]=2_5 , A : str="hamming_window" , A : str=3_27_68.0 , A : Union[str, Any]=0.97 , A : Dict=1.0 , A : Any=True , A : Union[str, Any]=True , A : List[Any]=False , **A : Tuple , ): '''simple docstring''' super().__init__(feature_size=A , sampling_rate=A , padding_value=A , **A ) a : Any = feature_size a : List[Any] = sampling_rate a : Any = padding_value a : str = hop_length a : Any = win_length a : List[Any] = frame_signal_scale a : Tuple = preemphasis_coeff a : Dict = mel_floor a : Optional[int] = normalize_means a : List[str] = normalize_vars a : Dict = win_function a : Union[str, Any] = return_attention_mask a : List[Any] = win_length * sampling_rate // 1_0_0_0 a : Tuple = hop_length * sampling_rate // 1_0_0_0 a : List[Any] = optimal_fft_length(self.sample_size ) a : Any = (self.n_fft // 2) + 1 def lowerCamelCase__ ( self : List[Any] , A : np.array ): '''simple docstring''' if self.win_function == "hamming_window": a : List[str] = window_function(window_length=self.sample_size , name=self.win_function , periodic=A ) else: a : Dict = window_function(window_length=self.sample_size , name=self.win_function ) a : str = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) a : List[Any] = spectrogram( one_waveform * self.frame_signal_scale , window=A , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=A , preemphasis=self.preemphasis_coeff , mel_filters=A , mel_floor=self.mel_floor , log_mel='log' , ) return msfc_features.T def lowerCamelCase__ ( self : int , A : Tuple , A : int , A : Optional[int] ): '''simple docstring''' if self.normalize_means: a : Any = x[:input_length].mean(axis=0 ) a : Dict = np.subtract(A , A ) if self.normalize_vars: a : Dict = x[:input_length].std(axis=0 ) a : Dict = np.divide(A , A ) if input_length < x.shape[0]: a : Dict = padding_value # make sure array is in float32 a : Optional[int] = x.astype(np.floataa ) return x def lowerCamelCase__ ( self : str , A : List[np.ndarray] , A : Optional[np.ndarray] = None ): '''simple docstring''' a : str = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(A , A , self.padding_value ) for x, n in zip(A , A )] def __call__( self : Dict , A : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , A : Union[bool, str, PaddingStrategy] = False , A : Optional[int] = None , A : bool = False , A : Optional[int] = None , A : Optional[bool] = None , A : Optional[Union[str, TensorType]] = None , A : Optional[int] = None , **A : Union[str, Any] , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the ``sampling_rate`` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) a : Optional[int] = isinstance(A , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) a : Dict = is_batched_numpy or ( isinstance(A , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: a : str = [np.asarray(A , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(A , np.ndarray ): a : List[str] = np.asarray(A , dtype=np.floataa ) elif isinstance(A , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): a : Optional[int] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: a : Any = [raw_speech] # extract fbank features a : str = [self._extract_mfsc_features(A ) for one_waveform in raw_speech] # convert into correct format for padding a : int = BatchFeature({'input_features': features} ) a : Union[str, Any] = self.pad( A , padding=A , max_length=A , truncation=A , pad_to_multiple_of=A , return_attention_mask=A , **A , ) # make sure list is in array format a : Optional[Any] = padded_inputs.get('input_features' ) if isinstance(input_features[0] , A ): a : List[str] = [np.asarray(A , dtype=np.floataa ) for feature in input_features] a : List[Any] = padded_inputs.get('attention_mask' ) if attention_mask is not None: a : int = [np.asarray(A , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: a : Any = ( np.array(A , dtype=np.intaa ) if self._get_padding_strategies(A , max_length=A ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) a : List[str] = self.normalize( padded_inputs['input_features'] , attention_mask=A ) if return_tensors is not None: a : Optional[int] = padded_inputs.convert_to_tensors(A ) return padded_inputs
186
1